var/home/core/zuul-output/0000755000175000017500000000000015111342070014517 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015111354250015466 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004730214015111354242017674 0ustar rootrootNov 25 15:02:20 crc systemd[1]: Starting Kubernetes Kubelet... Nov 25 15:02:20 crc restorecon[4676]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 15:02:21 crc restorecon[4676]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 15:02:21 crc restorecon[4676]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Nov 25 15:02:21 crc kubenswrapper[4890]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 25 15:02:21 crc kubenswrapper[4890]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Nov 25 15:02:21 crc kubenswrapper[4890]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 25 15:02:21 crc kubenswrapper[4890]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 25 15:02:21 crc kubenswrapper[4890]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Nov 25 15:02:21 crc kubenswrapper[4890]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.949917 4890 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953650 4890 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953665 4890 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953671 4890 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953675 4890 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953680 4890 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953684 4890 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953690 4890 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953694 4890 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953698 4890 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953701 4890 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953705 4890 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953709 4890 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953714 4890 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953719 4890 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953723 4890 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953728 4890 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953732 4890 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953736 4890 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953742 4890 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953746 4890 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953751 4890 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953754 4890 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953758 4890 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953762 4890 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953766 4890 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953771 4890 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953774 4890 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953778 4890 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953782 4890 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953785 4890 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953792 4890 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953796 4890 feature_gate.go:330] unrecognized feature gate: Example Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953800 4890 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953804 4890 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953807 4890 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953811 4890 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953814 4890 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953818 4890 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953822 4890 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953827 4890 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953831 4890 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953835 4890 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953839 4890 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953844 4890 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953848 4890 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953852 4890 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953855 4890 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953859 4890 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953862 4890 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953866 4890 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953869 4890 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953873 4890 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953876 4890 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953880 4890 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953883 4890 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953890 4890 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953895 4890 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953899 4890 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953904 4890 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953908 4890 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953912 4890 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953916 4890 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953919 4890 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953925 4890 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953929 4890 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953932 4890 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953936 4890 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953941 4890 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.953945 4890 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.954004 4890 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.954008 4890 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.955597 4890 flags.go:64] FLAG: --address="0.0.0.0" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.955650 4890 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.955667 4890 flags.go:64] FLAG: --anonymous-auth="true" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.955675 4890 flags.go:64] FLAG: --application-metrics-count-limit="100" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.955683 4890 flags.go:64] FLAG: --authentication-token-webhook="false" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.955688 4890 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.955697 4890 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.955705 4890 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.955717 4890 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.955743 4890 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.955748 4890 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.955755 4890 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.955760 4890 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.955765 4890 flags.go:64] FLAG: --cgroup-root="" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.955770 4890 flags.go:64] FLAG: --cgroups-per-qos="true" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.955775 4890 flags.go:64] FLAG: --client-ca-file="" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.955779 4890 flags.go:64] FLAG: --cloud-config="" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.955789 4890 flags.go:64] FLAG: --cloud-provider="" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.955794 4890 flags.go:64] FLAG: --cluster-dns="[]" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.955801 4890 flags.go:64] FLAG: --cluster-domain="" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956184 4890 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956197 4890 flags.go:64] FLAG: --config-dir="" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956202 4890 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956208 4890 flags.go:64] FLAG: --container-log-max-files="5" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956217 4890 flags.go:64] FLAG: --container-log-max-size="10Mi" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956222 4890 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956227 4890 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956231 4890 flags.go:64] FLAG: --containerd-namespace="k8s.io" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956236 4890 flags.go:64] FLAG: --contention-profiling="false" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956240 4890 flags.go:64] FLAG: --cpu-cfs-quota="true" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956245 4890 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956250 4890 flags.go:64] FLAG: --cpu-manager-policy="none" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956254 4890 flags.go:64] FLAG: --cpu-manager-policy-options="" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956260 4890 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956265 4890 flags.go:64] FLAG: --enable-controller-attach-detach="true" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956270 4890 flags.go:64] FLAG: --enable-debugging-handlers="true" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956274 4890 flags.go:64] FLAG: --enable-load-reader="false" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956279 4890 flags.go:64] FLAG: --enable-server="true" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956283 4890 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956291 4890 flags.go:64] FLAG: --event-burst="100" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956296 4890 flags.go:64] FLAG: --event-qps="50" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956300 4890 flags.go:64] FLAG: --event-storage-age-limit="default=0" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956304 4890 flags.go:64] FLAG: --event-storage-event-limit="default=0" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956309 4890 flags.go:64] FLAG: --eviction-hard="" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956331 4890 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956336 4890 flags.go:64] FLAG: --eviction-minimum-reclaim="" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956340 4890 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956345 4890 flags.go:64] FLAG: --eviction-soft="" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956349 4890 flags.go:64] FLAG: --eviction-soft-grace-period="" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956354 4890 flags.go:64] FLAG: --exit-on-lock-contention="false" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956358 4890 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956362 4890 flags.go:64] FLAG: --experimental-mounter-path="" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956367 4890 flags.go:64] FLAG: --fail-cgroupv1="false" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956371 4890 flags.go:64] FLAG: --fail-swap-on="true" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956376 4890 flags.go:64] FLAG: --feature-gates="" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956382 4890 flags.go:64] FLAG: --file-check-frequency="20s" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956386 4890 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956391 4890 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956395 4890 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956399 4890 flags.go:64] FLAG: --healthz-port="10248" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956404 4890 flags.go:64] FLAG: --help="false" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956408 4890 flags.go:64] FLAG: --hostname-override="" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956412 4890 flags.go:64] FLAG: --housekeeping-interval="10s" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956417 4890 flags.go:64] FLAG: --http-check-frequency="20s" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956421 4890 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956425 4890 flags.go:64] FLAG: --image-credential-provider-config="" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956429 4890 flags.go:64] FLAG: --image-gc-high-threshold="85" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956433 4890 flags.go:64] FLAG: --image-gc-low-threshold="80" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956438 4890 flags.go:64] FLAG: --image-service-endpoint="" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956442 4890 flags.go:64] FLAG: --kernel-memcg-notification="false" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956446 4890 flags.go:64] FLAG: --kube-api-burst="100" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956450 4890 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956455 4890 flags.go:64] FLAG: --kube-api-qps="50" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956459 4890 flags.go:64] FLAG: --kube-reserved="" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956463 4890 flags.go:64] FLAG: --kube-reserved-cgroup="" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956467 4890 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956471 4890 flags.go:64] FLAG: --kubelet-cgroups="" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956476 4890 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956482 4890 flags.go:64] FLAG: --lock-file="" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956486 4890 flags.go:64] FLAG: --log-cadvisor-usage="false" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956491 4890 flags.go:64] FLAG: --log-flush-frequency="5s" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956495 4890 flags.go:64] FLAG: --log-json-info-buffer-size="0" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956502 4890 flags.go:64] FLAG: --log-json-split-stream="false" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956508 4890 flags.go:64] FLAG: --log-text-info-buffer-size="0" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956512 4890 flags.go:64] FLAG: --log-text-split-stream="false" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956516 4890 flags.go:64] FLAG: --logging-format="text" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956520 4890 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956525 4890 flags.go:64] FLAG: --make-iptables-util-chains="true" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956529 4890 flags.go:64] FLAG: --manifest-url="" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956533 4890 flags.go:64] FLAG: --manifest-url-header="" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956540 4890 flags.go:64] FLAG: --max-housekeeping-interval="15s" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956544 4890 flags.go:64] FLAG: --max-open-files="1000000" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956550 4890 flags.go:64] FLAG: --max-pods="110" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956554 4890 flags.go:64] FLAG: --maximum-dead-containers="-1" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956559 4890 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956563 4890 flags.go:64] FLAG: --memory-manager-policy="None" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956567 4890 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956571 4890 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956576 4890 flags.go:64] FLAG: --node-ip="192.168.126.11" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956580 4890 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956595 4890 flags.go:64] FLAG: --node-status-max-images="50" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956599 4890 flags.go:64] FLAG: --node-status-update-frequency="10s" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956603 4890 flags.go:64] FLAG: --oom-score-adj="-999" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956607 4890 flags.go:64] FLAG: --pod-cidr="" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956611 4890 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956618 4890 flags.go:64] FLAG: --pod-manifest-path="" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956623 4890 flags.go:64] FLAG: --pod-max-pids="-1" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956628 4890 flags.go:64] FLAG: --pods-per-core="0" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956633 4890 flags.go:64] FLAG: --port="10250" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956637 4890 flags.go:64] FLAG: --protect-kernel-defaults="false" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956642 4890 flags.go:64] FLAG: --provider-id="" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956648 4890 flags.go:64] FLAG: --qos-reserved="" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956652 4890 flags.go:64] FLAG: --read-only-port="10255" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956657 4890 flags.go:64] FLAG: --register-node="true" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956661 4890 flags.go:64] FLAG: --register-schedulable="true" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956665 4890 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956673 4890 flags.go:64] FLAG: --registry-burst="10" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956678 4890 flags.go:64] FLAG: --registry-qps="5" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956682 4890 flags.go:64] FLAG: --reserved-cpus="" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956687 4890 flags.go:64] FLAG: --reserved-memory="" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956693 4890 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956697 4890 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956701 4890 flags.go:64] FLAG: --rotate-certificates="false" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956706 4890 flags.go:64] FLAG: --rotate-server-certificates="false" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956710 4890 flags.go:64] FLAG: --runonce="false" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956714 4890 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956718 4890 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956723 4890 flags.go:64] FLAG: --seccomp-default="false" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956727 4890 flags.go:64] FLAG: --serialize-image-pulls="true" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956731 4890 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956735 4890 flags.go:64] FLAG: --storage-driver-db="cadvisor" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956740 4890 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956744 4890 flags.go:64] FLAG: --storage-driver-password="root" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956748 4890 flags.go:64] FLAG: --storage-driver-secure="false" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956753 4890 flags.go:64] FLAG: --storage-driver-table="stats" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956757 4890 flags.go:64] FLAG: --storage-driver-user="root" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956761 4890 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956765 4890 flags.go:64] FLAG: --sync-frequency="1m0s" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956769 4890 flags.go:64] FLAG: --system-cgroups="" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956774 4890 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956780 4890 flags.go:64] FLAG: --system-reserved-cgroup="" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956784 4890 flags.go:64] FLAG: --tls-cert-file="" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956789 4890 flags.go:64] FLAG: --tls-cipher-suites="[]" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956795 4890 flags.go:64] FLAG: --tls-min-version="" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956799 4890 flags.go:64] FLAG: --tls-private-key-file="" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956804 4890 flags.go:64] FLAG: --topology-manager-policy="none" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956808 4890 flags.go:64] FLAG: --topology-manager-policy-options="" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956812 4890 flags.go:64] FLAG: --topology-manager-scope="container" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956817 4890 flags.go:64] FLAG: --v="2" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956824 4890 flags.go:64] FLAG: --version="false" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956830 4890 flags.go:64] FLAG: --vmodule="" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956836 4890 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.956840 4890 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.956990 4890 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.956995 4890 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957002 4890 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957007 4890 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957011 4890 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957015 4890 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957020 4890 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957025 4890 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957029 4890 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957033 4890 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957037 4890 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957040 4890 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957044 4890 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957048 4890 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957051 4890 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957055 4890 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957058 4890 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957062 4890 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957066 4890 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957070 4890 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957073 4890 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957077 4890 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957081 4890 feature_gate.go:330] unrecognized feature gate: Example Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957085 4890 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957089 4890 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957094 4890 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957097 4890 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957101 4890 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957105 4890 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957108 4890 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957112 4890 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957116 4890 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957119 4890 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957123 4890 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957126 4890 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957130 4890 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957133 4890 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957137 4890 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957141 4890 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957145 4890 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957148 4890 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957153 4890 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957174 4890 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957178 4890 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957182 4890 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957185 4890 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957189 4890 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957193 4890 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957197 4890 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957201 4890 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957205 4890 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957208 4890 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957212 4890 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957216 4890 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957219 4890 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957223 4890 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957227 4890 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957230 4890 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957234 4890 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957238 4890 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957241 4890 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957245 4890 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957248 4890 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957252 4890 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957256 4890 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957260 4890 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957263 4890 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957267 4890 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957271 4890 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957276 4890 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.957280 4890 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.957295 4890 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.969737 4890 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.969807 4890 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.969883 4890 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.969894 4890 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.969898 4890 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.969904 4890 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.969908 4890 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.969911 4890 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.969915 4890 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.969919 4890 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.969923 4890 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.969927 4890 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.969931 4890 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.969936 4890 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.969939 4890 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.969943 4890 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.969946 4890 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.969950 4890 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.969954 4890 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.969957 4890 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.969961 4890 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.969966 4890 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.969973 4890 feature_gate.go:330] unrecognized feature gate: Example Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.969978 4890 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.969984 4890 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.969990 4890 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.969997 4890 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970002 4890 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970008 4890 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970015 4890 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970020 4890 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970024 4890 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970028 4890 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970032 4890 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970036 4890 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970040 4890 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970046 4890 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970049 4890 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970053 4890 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970057 4890 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970060 4890 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970065 4890 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970068 4890 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970072 4890 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970076 4890 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970079 4890 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970083 4890 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970088 4890 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970092 4890 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970097 4890 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970101 4890 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970106 4890 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970111 4890 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970116 4890 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970136 4890 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970140 4890 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970145 4890 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970149 4890 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970152 4890 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970173 4890 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970177 4890 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970181 4890 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970185 4890 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970189 4890 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970192 4890 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970195 4890 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970199 4890 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970203 4890 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970206 4890 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970210 4890 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970213 4890 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970216 4890 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970221 4890 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.970228 4890 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970355 4890 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970363 4890 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970368 4890 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970373 4890 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970376 4890 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970380 4890 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970384 4890 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970389 4890 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970393 4890 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970397 4890 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970402 4890 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970406 4890 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970410 4890 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970413 4890 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970417 4890 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970421 4890 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970424 4890 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970428 4890 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970431 4890 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970435 4890 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970438 4890 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970441 4890 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970445 4890 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970448 4890 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970452 4890 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970456 4890 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970460 4890 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970467 4890 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970475 4890 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970480 4890 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970485 4890 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970490 4890 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970494 4890 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970498 4890 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970505 4890 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970510 4890 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970514 4890 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970519 4890 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970523 4890 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970528 4890 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970533 4890 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970537 4890 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970544 4890 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970549 4890 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970554 4890 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970559 4890 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970564 4890 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970569 4890 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970575 4890 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970581 4890 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970587 4890 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970593 4890 feature_gate.go:330] unrecognized feature gate: Example Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970599 4890 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970605 4890 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970610 4890 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970615 4890 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970620 4890 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970626 4890 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970631 4890 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970636 4890 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970641 4890 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970646 4890 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970652 4890 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970657 4890 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970664 4890 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970670 4890 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970676 4890 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970682 4890 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970688 4890 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970695 4890 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 25 15:02:21 crc kubenswrapper[4890]: W1125 15:02:21.970704 4890 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.970714 4890 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.971854 4890 server.go:940] "Client rotation is on, will bootstrap in background" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.978500 4890 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.978658 4890 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.982012 4890 server.go:997] "Starting client certificate rotation" Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.982059 4890 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.982980 4890 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2026-01-03 16:22:01.027150091 +0000 UTC Nov 25 15:02:21 crc kubenswrapper[4890]: I1125 15:02:21.983061 4890 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 937h19m39.044091909s for next certificate rotation Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.016504 4890 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.021275 4890 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.040316 4890 log.go:25] "Validated CRI v1 runtime API" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.074497 4890 log.go:25] "Validated CRI v1 image API" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.077082 4890 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.084335 4890 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-11-25-14-58-01-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.084383 4890 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.102503 4890 manager.go:217] Machine: {Timestamp:2025-11-25 15:02:22.100094574 +0000 UTC m=+0.542557194 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:5350de30-707e-4bde-b990-1fa65ae3a01e BootID:7c7b4592-d206-4ee7-bf93-ada4b8bc755a Filesystems:[{Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:5c:77:47 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:5c:77:47 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:8b:38:f5 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:ac:20:59 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:a1:e2:05 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:8c:c9:b8 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:d2:34:ba:24:4c:70 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:82:0c:43:71:23:1f Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.102805 4890 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.102995 4890 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.103377 4890 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.103534 4890 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.103566 4890 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.103772 4890 topology_manager.go:138] "Creating topology manager with none policy" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.103781 4890 container_manager_linux.go:303] "Creating device plugin manager" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.104248 4890 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.104276 4890 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.104980 4890 state_mem.go:36] "Initialized new in-memory state store" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.105079 4890 server.go:1245] "Using root directory" path="/var/lib/kubelet" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.108107 4890 kubelet.go:418] "Attempting to sync node with API server" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.108135 4890 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.108185 4890 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.108203 4890 kubelet.go:324] "Adding apiserver pod source" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.108216 4890 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.112555 4890 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.113553 4890 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Nov 25 15:02:22 crc kubenswrapper[4890]: W1125 15:02:22.114365 4890 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Nov 25 15:02:22 crc kubenswrapper[4890]: E1125 15:02:22.114478 4890 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.12:6443: connect: connection refused" logger="UnhandledError" Nov 25 15:02:22 crc kubenswrapper[4890]: W1125 15:02:22.114524 4890 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Nov 25 15:02:22 crc kubenswrapper[4890]: E1125 15:02:22.114570 4890 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.12:6443: connect: connection refused" logger="UnhandledError" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.115471 4890 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.117557 4890 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.117586 4890 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.117596 4890 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.117605 4890 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.117621 4890 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.117630 4890 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.117639 4890 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.117660 4890 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.117669 4890 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.117688 4890 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.117708 4890 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.117717 4890 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.118615 4890 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.119233 4890 server.go:1280] "Started kubelet" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.119465 4890 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.119620 4890 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.120360 4890 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 25 15:02:22 crc systemd[1]: Started Kubernetes Kubelet. Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.121626 4890 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.121665 4890 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.121822 4890 volume_manager.go:287] "The desired_state_of_world populator starts" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.121841 4890 volume_manager.go:289] "Starting Kubelet Volume Manager" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.121839 4890 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 08:12:03.881203508 +0000 UTC Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.121884 4890 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 617h9m41.759321725s for next certificate rotation Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.121889 4890 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Nov 25 15:02:22 crc kubenswrapper[4890]: E1125 15:02:22.121903 4890 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.126089 4890 factory.go:55] Registering systemd factory Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.126119 4890 factory.go:221] Registration of the systemd container factory successfully Nov 25 15:02:22 crc kubenswrapper[4890]: W1125 15:02:22.127197 4890 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Nov 25 15:02:22 crc kubenswrapper[4890]: E1125 15:02:22.127308 4890 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.12:6443: connect: connection refused" logger="UnhandledError" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.129007 4890 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.129111 4890 factory.go:153] Registering CRI-O factory Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.129133 4890 factory.go:221] Registration of the crio container factory successfully Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.129253 4890 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.129544 4890 factory.go:103] Registering Raw factory Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.129602 4890 manager.go:1196] Started watching for new ooms in manager Nov 25 15:02:22 crc kubenswrapper[4890]: E1125 15:02:22.130346 4890 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" interval="200ms" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.131914 4890 server.go:460] "Adding debug handlers to kubelet server" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.136045 4890 manager.go:319] Starting recovery of all containers Nov 25 15:02:22 crc kubenswrapper[4890]: E1125 15:02:22.135840 4890 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.12:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187b4816df573b2a default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-25 15:02:22.119197482 +0000 UTC m=+0.561660092,LastTimestamp:2025-11-25 15:02:22.119197482 +0000 UTC m=+0.561660092,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.137841 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.137895 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.137909 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.137920 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.137928 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.137937 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.137946 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.137955 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.137966 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.137978 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.137991 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.138002 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.138012 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.138151 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.138175 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.138188 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.138198 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.138226 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.138235 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.138245 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.138254 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.138263 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.138272 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.138281 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.138289 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.138301 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.138339 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.138349 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.138357 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.138372 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141278 4890 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141312 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141325 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141337 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141350 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141361 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141372 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141383 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141395 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141406 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141417 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141428 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141438 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141449 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141460 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141470 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141479 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141492 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141505 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141515 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141526 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141536 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141547 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141585 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141599 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141610 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141622 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141634 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141645 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141655 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141665 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141678 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141701 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141720 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141733 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141745 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141758 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141769 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141780 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141792 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141804 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141816 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141829 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141842 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141854 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141867 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141879 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141930 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141950 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141964 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141980 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.141991 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142002 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142014 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142025 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142042 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142053 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142066 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142078 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142091 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142102 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142113 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142124 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142136 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142147 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142226 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142265 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142278 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142296 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142309 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142321 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142332 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142344 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142354 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142365 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142393 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142405 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142416 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142427 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142438 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142448 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142460 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142474 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142483 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142494 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142504 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142515 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142525 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142536 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142547 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142556 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142566 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142576 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142585 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142594 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142606 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142617 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142626 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142637 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142646 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142656 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142665 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142675 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142691 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142710 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142724 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142737 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142749 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142760 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142770 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142779 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142789 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142798 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142807 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142820 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142829 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142838 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142848 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142857 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142866 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142876 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142887 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142903 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142921 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142935 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142947 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142959 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142971 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142983 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.142994 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.143005 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.143015 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.143026 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.143036 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.143046 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.143056 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.143069 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.143079 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.143092 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.143109 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.143121 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.143134 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.143146 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.143184 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.143196 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.143208 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.143221 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.143233 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.143244 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.143255 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.143266 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.143276 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.143286 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.143295 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.143305 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.143315 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.143325 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.143336 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.143345 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.143356 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.143367 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.143378 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.143389 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.143399 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.143410 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.143422 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.143432 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.143442 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.143453 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.143464 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.143476 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.143487 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.143497 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.143507 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.143519 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.143529 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.143541 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.143551 4890 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.143563 4890 reconstruct.go:97] "Volume reconstruction finished" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.143570 4890 reconciler.go:26] "Reconciler: start to sync state" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.157961 4890 manager.go:324] Recovery completed Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.170247 4890 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.171659 4890 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.171692 4890 status_manager.go:217] "Starting to sync pod status with apiserver" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.171900 4890 kubelet.go:2335] "Starting kubelet main sync loop" Nov 25 15:02:22 crc kubenswrapper[4890]: E1125 15:02:22.172080 4890 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.172422 4890 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 15:02:22 crc kubenswrapper[4890]: W1125 15:02:22.173663 4890 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Nov 25 15:02:22 crc kubenswrapper[4890]: E1125 15:02:22.173735 4890 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.12:6443: connect: connection refused" logger="UnhandledError" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.174010 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.174047 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.174057 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.174784 4890 cpu_manager.go:225] "Starting CPU manager" policy="none" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.174802 4890 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.174820 4890 state_mem.go:36] "Initialized new in-memory state store" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.201335 4890 policy_none.go:49] "None policy: Start" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.203821 4890 memory_manager.go:170] "Starting memorymanager" policy="None" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.203864 4890 state_mem.go:35] "Initializing new in-memory state store" Nov 25 15:02:22 crc kubenswrapper[4890]: E1125 15:02:22.222476 4890 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.248244 4890 manager.go:334] "Starting Device Plugin manager" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.248314 4890 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.248332 4890 server.go:79] "Starting device plugin registration server" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.248885 4890 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.248905 4890 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.249127 4890 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.249345 4890 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.249357 4890 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 25 15:02:22 crc kubenswrapper[4890]: E1125 15:02:22.257275 4890 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.272967 4890 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc"] Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.273109 4890 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.274918 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.274957 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.274971 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.275171 4890 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.275509 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.275574 4890 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.276191 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.276228 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.276239 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.276398 4890 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.276543 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.276573 4890 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.277084 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.277150 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.277198 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.277240 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.277263 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.277243 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.277277 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.277381 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.277409 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.277461 4890 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.277639 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.277673 4890 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.278313 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.278333 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.278356 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.278338 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.278364 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.278371 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.278507 4890 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.278644 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.278698 4890 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.279311 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.279337 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.279348 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.279502 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.279530 4890 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.280176 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.280200 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.280211 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.280179 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.280273 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.280282 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:22 crc kubenswrapper[4890]: E1125 15:02:22.331222 4890 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" interval="400ms" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.345462 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.345544 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.345581 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.345615 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.345651 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.345683 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.345716 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.345745 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.345774 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.345804 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.345856 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.345901 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.345934 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.346002 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.346040 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.349765 4890 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.351206 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.351237 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.351249 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.351280 4890 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 15:02:22 crc kubenswrapper[4890]: E1125 15:02:22.351673 4890 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.12:6443: connect: connection refused" node="crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.447031 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.447118 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.447193 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.447275 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.447304 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.447323 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.447346 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.447365 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.447402 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.447399 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.447453 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.447458 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.447431 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.447486 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.447547 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.447404 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.447495 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.447419 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.447677 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.447695 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.447714 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.447759 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.447713 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.447785 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.447848 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.447934 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.447954 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.447989 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.448008 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.448102 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.551980 4890 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.553667 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.553714 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.553729 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.553772 4890 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 15:02:22 crc kubenswrapper[4890]: E1125 15:02:22.555423 4890 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.12:6443: connect: connection refused" node="crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.601889 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.624315 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.631171 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: W1125 15:02:22.650456 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-48ecf001ee0cb12e44c96f35f0539cb4179a280ec4a9600d01112d2e6b3072b2 WatchSource:0}: Error finding container 48ecf001ee0cb12e44c96f35f0539cb4179a280ec4a9600d01112d2e6b3072b2: Status 404 returned error can't find the container with id 48ecf001ee0cb12e44c96f35f0539cb4179a280ec4a9600d01112d2e6b3072b2 Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.656018 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.658634 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 25 15:02:22 crc kubenswrapper[4890]: W1125 15:02:22.663196 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-ad57cd55ff8078f2602896a41dad2c076543bc36d98cb5aded0f93a13cac6c3d WatchSource:0}: Error finding container ad57cd55ff8078f2602896a41dad2c076543bc36d98cb5aded0f93a13cac6c3d: Status 404 returned error can't find the container with id ad57cd55ff8078f2602896a41dad2c076543bc36d98cb5aded0f93a13cac6c3d Nov 25 15:02:22 crc kubenswrapper[4890]: W1125 15:02:22.677667 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-d4a174171db5552d180df793bc77fd2ee08bedb33f6d5c17b02173040a0955ab WatchSource:0}: Error finding container d4a174171db5552d180df793bc77fd2ee08bedb33f6d5c17b02173040a0955ab: Status 404 returned error can't find the container with id d4a174171db5552d180df793bc77fd2ee08bedb33f6d5c17b02173040a0955ab Nov 25 15:02:22 crc kubenswrapper[4890]: W1125 15:02:22.679121 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-5d845d7ebce43c43a27e9ee8c5b7b322311b5bc2ed4803d9a1e20648bfd577f4 WatchSource:0}: Error finding container 5d845d7ebce43c43a27e9ee8c5b7b322311b5bc2ed4803d9a1e20648bfd577f4: Status 404 returned error can't find the container with id 5d845d7ebce43c43a27e9ee8c5b7b322311b5bc2ed4803d9a1e20648bfd577f4 Nov 25 15:02:22 crc kubenswrapper[4890]: W1125 15:02:22.684205 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-e62fe3d102e5917228085f435e1971da8b4c7b4fe04948cc245fc232c90e738b WatchSource:0}: Error finding container e62fe3d102e5917228085f435e1971da8b4c7b4fe04948cc245fc232c90e738b: Status 404 returned error can't find the container with id e62fe3d102e5917228085f435e1971da8b4c7b4fe04948cc245fc232c90e738b Nov 25 15:02:22 crc kubenswrapper[4890]: E1125 15:02:22.732814 4890 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" interval="800ms" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.956153 4890 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.957967 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.958020 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.958031 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:22 crc kubenswrapper[4890]: I1125 15:02:22.958068 4890 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 15:02:22 crc kubenswrapper[4890]: E1125 15:02:22.958675 4890 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.12:6443: connect: connection refused" node="crc" Nov 25 15:02:23 crc kubenswrapper[4890]: W1125 15:02:23.051241 4890 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Nov 25 15:02:23 crc kubenswrapper[4890]: E1125 15:02:23.051365 4890 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.12:6443: connect: connection refused" logger="UnhandledError" Nov 25 15:02:23 crc kubenswrapper[4890]: W1125 15:02:23.056322 4890 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Nov 25 15:02:23 crc kubenswrapper[4890]: E1125 15:02:23.056458 4890 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.12:6443: connect: connection refused" logger="UnhandledError" Nov 25 15:02:23 crc kubenswrapper[4890]: I1125 15:02:23.130382 4890 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Nov 25 15:02:23 crc kubenswrapper[4890]: W1125 15:02:23.136818 4890 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Nov 25 15:02:23 crc kubenswrapper[4890]: E1125 15:02:23.136924 4890 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.12:6443: connect: connection refused" logger="UnhandledError" Nov 25 15:02:23 crc kubenswrapper[4890]: I1125 15:02:23.177875 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ad57cd55ff8078f2602896a41dad2c076543bc36d98cb5aded0f93a13cac6c3d"} Nov 25 15:02:23 crc kubenswrapper[4890]: I1125 15:02:23.179334 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"48ecf001ee0cb12e44c96f35f0539cb4179a280ec4a9600d01112d2e6b3072b2"} Nov 25 15:02:23 crc kubenswrapper[4890]: I1125 15:02:23.180910 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d4a174171db5552d180df793bc77fd2ee08bedb33f6d5c17b02173040a0955ab"} Nov 25 15:02:23 crc kubenswrapper[4890]: I1125 15:02:23.182909 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"5d845d7ebce43c43a27e9ee8c5b7b322311b5bc2ed4803d9a1e20648bfd577f4"} Nov 25 15:02:23 crc kubenswrapper[4890]: I1125 15:02:23.183898 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"e62fe3d102e5917228085f435e1971da8b4c7b4fe04948cc245fc232c90e738b"} Nov 25 15:02:23 crc kubenswrapper[4890]: W1125 15:02:23.331766 4890 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Nov 25 15:02:23 crc kubenswrapper[4890]: E1125 15:02:23.331879 4890 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.12:6443: connect: connection refused" logger="UnhandledError" Nov 25 15:02:23 crc kubenswrapper[4890]: E1125 15:02:23.534530 4890 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" interval="1.6s" Nov 25 15:02:23 crc kubenswrapper[4890]: I1125 15:02:23.759444 4890 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 15:02:23 crc kubenswrapper[4890]: I1125 15:02:23.760819 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:23 crc kubenswrapper[4890]: I1125 15:02:23.760857 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:23 crc kubenswrapper[4890]: I1125 15:02:23.760868 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:23 crc kubenswrapper[4890]: I1125 15:02:23.760895 4890 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 15:02:23 crc kubenswrapper[4890]: E1125 15:02:23.761517 4890 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.12:6443: connect: connection refused" node="crc" Nov 25 15:02:23 crc kubenswrapper[4890]: E1125 15:02:23.859128 4890 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.12:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187b4816df573b2a default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-25 15:02:22.119197482 +0000 UTC m=+0.561660092,LastTimestamp:2025-11-25 15:02:22.119197482 +0000 UTC m=+0.561660092,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 25 15:02:24 crc kubenswrapper[4890]: I1125 15:02:24.130008 4890 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Nov 25 15:02:24 crc kubenswrapper[4890]: I1125 15:02:24.189280 4890 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453" exitCode=0 Nov 25 15:02:24 crc kubenswrapper[4890]: I1125 15:02:24.189403 4890 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 15:02:24 crc kubenswrapper[4890]: I1125 15:02:24.189421 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453"} Nov 25 15:02:24 crc kubenswrapper[4890]: I1125 15:02:24.190386 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:24 crc kubenswrapper[4890]: I1125 15:02:24.190426 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:24 crc kubenswrapper[4890]: I1125 15:02:24.190437 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:24 crc kubenswrapper[4890]: I1125 15:02:24.191350 4890 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325" exitCode=0 Nov 25 15:02:24 crc kubenswrapper[4890]: I1125 15:02:24.191406 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325"} Nov 25 15:02:24 crc kubenswrapper[4890]: I1125 15:02:24.191538 4890 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 15:02:24 crc kubenswrapper[4890]: I1125 15:02:24.192489 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:24 crc kubenswrapper[4890]: I1125 15:02:24.192519 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:24 crc kubenswrapper[4890]: I1125 15:02:24.192528 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:24 crc kubenswrapper[4890]: I1125 15:02:24.192707 4890 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 15:02:24 crc kubenswrapper[4890]: I1125 15:02:24.193761 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:24 crc kubenswrapper[4890]: I1125 15:02:24.193798 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:24 crc kubenswrapper[4890]: I1125 15:02:24.193810 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:24 crc kubenswrapper[4890]: I1125 15:02:24.193973 4890 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="2a6ad38c7fe3f9eb0dfc09bbd22aba8c16d4b51b9bd9142f4204f25b914155c9" exitCode=0 Nov 25 15:02:24 crc kubenswrapper[4890]: I1125 15:02:24.194075 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"2a6ad38c7fe3f9eb0dfc09bbd22aba8c16d4b51b9bd9142f4204f25b914155c9"} Nov 25 15:02:24 crc kubenswrapper[4890]: I1125 15:02:24.194116 4890 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 15:02:24 crc kubenswrapper[4890]: I1125 15:02:24.196435 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:24 crc kubenswrapper[4890]: I1125 15:02:24.196469 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:24 crc kubenswrapper[4890]: I1125 15:02:24.196479 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:24 crc kubenswrapper[4890]: I1125 15:02:24.197839 4890 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="f8092652bfe2886bf2ea5e6e0cbd6c84efc82b21182a4509a624b399943eede7" exitCode=0 Nov 25 15:02:24 crc kubenswrapper[4890]: I1125 15:02:24.197945 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"f8092652bfe2886bf2ea5e6e0cbd6c84efc82b21182a4509a624b399943eede7"} Nov 25 15:02:24 crc kubenswrapper[4890]: I1125 15:02:24.198021 4890 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 15:02:24 crc kubenswrapper[4890]: I1125 15:02:24.199662 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:24 crc kubenswrapper[4890]: I1125 15:02:24.199708 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:24 crc kubenswrapper[4890]: I1125 15:02:24.199726 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:24 crc kubenswrapper[4890]: I1125 15:02:24.203936 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"3281314d85a20945d56b75ab7b7aef1e382a73e74dff03773d118d03eb0ee16e"} Nov 25 15:02:24 crc kubenswrapper[4890]: I1125 15:02:24.204170 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"20505e91fc60158eeb3a8cfa0aafc5b4e4436d5def5e735a3a58b0f9f3621058"} Nov 25 15:02:24 crc kubenswrapper[4890]: I1125 15:02:24.204188 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d19bc48acb392051f0f6c5a6820135bf29ebe6d2afd098697553298933cf280d"} Nov 25 15:02:24 crc kubenswrapper[4890]: I1125 15:02:24.204199 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"93325a6407389c4a7db0ba3c1a9c84a6980db3c3277cdf3860583a564b562bbe"} Nov 25 15:02:24 crc kubenswrapper[4890]: I1125 15:02:24.204066 4890 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 15:02:24 crc kubenswrapper[4890]: I1125 15:02:24.206396 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:24 crc kubenswrapper[4890]: I1125 15:02:24.206486 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:24 crc kubenswrapper[4890]: I1125 15:02:24.206501 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:24 crc kubenswrapper[4890]: I1125 15:02:24.343594 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 15:02:25 crc kubenswrapper[4890]: I1125 15:02:25.130188 4890 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Nov 25 15:02:25 crc kubenswrapper[4890]: W1125 15:02:25.134491 4890 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Nov 25 15:02:25 crc kubenswrapper[4890]: E1125 15:02:25.134591 4890 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.12:6443: connect: connection refused" logger="UnhandledError" Nov 25 15:02:25 crc kubenswrapper[4890]: E1125 15:02:25.135909 4890 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" interval="3.2s" Nov 25 15:02:25 crc kubenswrapper[4890]: I1125 15:02:25.219636 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f506ec649ce24f4df6fd33e4aeafd13e42167dd71323524a5a3164517659fcc9"} Nov 25 15:02:25 crc kubenswrapper[4890]: I1125 15:02:25.219706 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"63e3a186b14ea199d1aeba5751290c266ec918cb055120cfc5c8a2ae03b1469e"} Nov 25 15:02:25 crc kubenswrapper[4890]: I1125 15:02:25.219722 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6a860a998c40aee3e94e4cfd748f54391b3096fd3ed75c84c7ea2a5a7cefa33a"} Nov 25 15:02:25 crc kubenswrapper[4890]: I1125 15:02:25.219733 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"76d277b4001db9d63b0af88b98afcc9019321101429c9be47b3022d3575d4e06"} Nov 25 15:02:25 crc kubenswrapper[4890]: I1125 15:02:25.223717 4890 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787" exitCode=0 Nov 25 15:02:25 crc kubenswrapper[4890]: I1125 15:02:25.223821 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787"} Nov 25 15:02:25 crc kubenswrapper[4890]: I1125 15:02:25.224003 4890 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 15:02:25 crc kubenswrapper[4890]: I1125 15:02:25.226371 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"83ee18494b4658ba014b0013673acea38ee70056ab1f6cd6aa4e173fe8531458"} Nov 25 15:02:25 crc kubenswrapper[4890]: I1125 15:02:25.226405 4890 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 15:02:25 crc kubenswrapper[4890]: I1125 15:02:25.227086 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:25 crc kubenswrapper[4890]: I1125 15:02:25.227289 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:25 crc kubenswrapper[4890]: I1125 15:02:25.227499 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:25 crc kubenswrapper[4890]: I1125 15:02:25.227603 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:25 crc kubenswrapper[4890]: I1125 15:02:25.227650 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:25 crc kubenswrapper[4890]: I1125 15:02:25.227666 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:25 crc kubenswrapper[4890]: I1125 15:02:25.229836 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"539bab27d80f8d76cc54a60f508a9985e8a3aa3d2f949ea5a5c2f0d92227da30"} Nov 25 15:02:25 crc kubenswrapper[4890]: I1125 15:02:25.229933 4890 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 15:02:25 crc kubenswrapper[4890]: I1125 15:02:25.229943 4890 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 15:02:25 crc kubenswrapper[4890]: I1125 15:02:25.229933 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"ca4d48ac675ee73e0946cc802994bb407b6770695320ce2380ddfce26f1b5b8f"} Nov 25 15:02:25 crc kubenswrapper[4890]: I1125 15:02:25.230072 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"d1154cddf4cb46b8a9db26e7f982ed9dd382d85b5ceb95e50325a475eed408f3"} Nov 25 15:02:25 crc kubenswrapper[4890]: I1125 15:02:25.232148 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:25 crc kubenswrapper[4890]: I1125 15:02:25.232217 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:25 crc kubenswrapper[4890]: I1125 15:02:25.232230 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:25 crc kubenswrapper[4890]: I1125 15:02:25.232213 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:25 crc kubenswrapper[4890]: I1125 15:02:25.232387 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:25 crc kubenswrapper[4890]: I1125 15:02:25.232404 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:25 crc kubenswrapper[4890]: I1125 15:02:25.236202 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 15:02:25 crc kubenswrapper[4890]: I1125 15:02:25.362297 4890 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 15:02:25 crc kubenswrapper[4890]: I1125 15:02:25.363721 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:25 crc kubenswrapper[4890]: I1125 15:02:25.363781 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:25 crc kubenswrapper[4890]: I1125 15:02:25.363797 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:25 crc kubenswrapper[4890]: I1125 15:02:25.363829 4890 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 15:02:25 crc kubenswrapper[4890]: E1125 15:02:25.364326 4890 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.12:6443: connect: connection refused" node="crc" Nov 25 15:02:25 crc kubenswrapper[4890]: W1125 15:02:25.503278 4890 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Nov 25 15:02:25 crc kubenswrapper[4890]: E1125 15:02:25.503380 4890 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.12:6443: connect: connection refused" logger="UnhandledError" Nov 25 15:02:26 crc kubenswrapper[4890]: I1125 15:02:26.236399 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"081a6d623cd043c0919d04e2ca93314b2b85fe33048634aa2637297d54b52cee"} Nov 25 15:02:26 crc kubenswrapper[4890]: I1125 15:02:26.236437 4890 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 15:02:26 crc kubenswrapper[4890]: I1125 15:02:26.237523 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:26 crc kubenswrapper[4890]: I1125 15:02:26.237574 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:26 crc kubenswrapper[4890]: I1125 15:02:26.237586 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:26 crc kubenswrapper[4890]: I1125 15:02:26.238730 4890 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17" exitCode=0 Nov 25 15:02:26 crc kubenswrapper[4890]: I1125 15:02:26.238773 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17"} Nov 25 15:02:26 crc kubenswrapper[4890]: I1125 15:02:26.238828 4890 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 15:02:26 crc kubenswrapper[4890]: I1125 15:02:26.238836 4890 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 15:02:26 crc kubenswrapper[4890]: I1125 15:02:26.238881 4890 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 15:02:26 crc kubenswrapper[4890]: I1125 15:02:26.239000 4890 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 15:02:26 crc kubenswrapper[4890]: I1125 15:02:26.239145 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 15:02:26 crc kubenswrapper[4890]: I1125 15:02:26.240236 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:26 crc kubenswrapper[4890]: I1125 15:02:26.240261 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:26 crc kubenswrapper[4890]: I1125 15:02:26.240274 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:26 crc kubenswrapper[4890]: I1125 15:02:26.240274 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:26 crc kubenswrapper[4890]: I1125 15:02:26.240302 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:26 crc kubenswrapper[4890]: I1125 15:02:26.240637 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:26 crc kubenswrapper[4890]: I1125 15:02:26.240244 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:26 crc kubenswrapper[4890]: I1125 15:02:26.240731 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:26 crc kubenswrapper[4890]: I1125 15:02:26.240742 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:26 crc kubenswrapper[4890]: I1125 15:02:26.244026 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:26 crc kubenswrapper[4890]: I1125 15:02:26.244095 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:26 crc kubenswrapper[4890]: I1125 15:02:26.244111 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:26 crc kubenswrapper[4890]: I1125 15:02:26.567534 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 15:02:27 crc kubenswrapper[4890]: I1125 15:02:27.246046 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"659004bd47948def6b33498d8d1474f28c2e98986754f61af012267f3984269e"} Nov 25 15:02:27 crc kubenswrapper[4890]: I1125 15:02:27.246105 4890 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 15:02:27 crc kubenswrapper[4890]: I1125 15:02:27.246117 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 15:02:27 crc kubenswrapper[4890]: I1125 15:02:27.246137 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"1caa240816a6e9d2821006ce240eccf27d8f57a7f0c5394314fa7177488da4f6"} Nov 25 15:02:27 crc kubenswrapper[4890]: I1125 15:02:27.246149 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"e5f9285acc7ac3cbb503a1e546728748c42971c8c9019cc8ca837076f3e55558"} Nov 25 15:02:27 crc kubenswrapper[4890]: I1125 15:02:27.246185 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ca4c4921f0973b03f6dade29c2ef32684a481cd1c3e1e23b64ed636943c9bf11"} Nov 25 15:02:27 crc kubenswrapper[4890]: I1125 15:02:27.246198 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b6d2dcd8df64e44df8cdb4c3a022f03fccb39451631ca35b30d4c7c35271ff0b"} Nov 25 15:02:27 crc kubenswrapper[4890]: I1125 15:02:27.246214 4890 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 15:02:27 crc kubenswrapper[4890]: I1125 15:02:27.246076 4890 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 15:02:27 crc kubenswrapper[4890]: I1125 15:02:27.246316 4890 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 15:02:27 crc kubenswrapper[4890]: I1125 15:02:27.247987 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:27 crc kubenswrapper[4890]: I1125 15:02:27.248025 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:27 crc kubenswrapper[4890]: I1125 15:02:27.248030 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:27 crc kubenswrapper[4890]: I1125 15:02:27.247987 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:27 crc kubenswrapper[4890]: I1125 15:02:27.248058 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:27 crc kubenswrapper[4890]: I1125 15:02:27.248068 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:27 crc kubenswrapper[4890]: I1125 15:02:27.248084 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:27 crc kubenswrapper[4890]: I1125 15:02:27.248117 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:27 crc kubenswrapper[4890]: I1125 15:02:27.248047 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:27 crc kubenswrapper[4890]: I1125 15:02:27.248036 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:27 crc kubenswrapper[4890]: I1125 15:02:27.248172 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:27 crc kubenswrapper[4890]: I1125 15:02:27.248131 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:27 crc kubenswrapper[4890]: I1125 15:02:27.250945 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 15:02:27 crc kubenswrapper[4890]: I1125 15:02:27.257135 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 15:02:28 crc kubenswrapper[4890]: I1125 15:02:28.049517 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Nov 25 15:02:28 crc kubenswrapper[4890]: I1125 15:02:28.248576 4890 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 15:02:28 crc kubenswrapper[4890]: I1125 15:02:28.248653 4890 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 15:02:28 crc kubenswrapper[4890]: I1125 15:02:28.248842 4890 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 15:02:28 crc kubenswrapper[4890]: I1125 15:02:28.249808 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:28 crc kubenswrapper[4890]: I1125 15:02:28.249853 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:28 crc kubenswrapper[4890]: I1125 15:02:28.249862 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:28 crc kubenswrapper[4890]: I1125 15:02:28.250094 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:28 crc kubenswrapper[4890]: I1125 15:02:28.250116 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:28 crc kubenswrapper[4890]: I1125 15:02:28.250127 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:28 crc kubenswrapper[4890]: I1125 15:02:28.250200 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:28 crc kubenswrapper[4890]: I1125 15:02:28.250233 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:28 crc kubenswrapper[4890]: I1125 15:02:28.250247 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:28 crc kubenswrapper[4890]: I1125 15:02:28.565053 4890 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 15:02:28 crc kubenswrapper[4890]: I1125 15:02:28.566614 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:28 crc kubenswrapper[4890]: I1125 15:02:28.566657 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:28 crc kubenswrapper[4890]: I1125 15:02:28.566669 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:28 crc kubenswrapper[4890]: I1125 15:02:28.566694 4890 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 15:02:29 crc kubenswrapper[4890]: I1125 15:02:29.251065 4890 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 15:02:29 crc kubenswrapper[4890]: I1125 15:02:29.251370 4890 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 15:02:29 crc kubenswrapper[4890]: I1125 15:02:29.252187 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:29 crc kubenswrapper[4890]: I1125 15:02:29.252220 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:29 crc kubenswrapper[4890]: I1125 15:02:29.252230 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:29 crc kubenswrapper[4890]: I1125 15:02:29.252836 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:29 crc kubenswrapper[4890]: I1125 15:02:29.252902 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:29 crc kubenswrapper[4890]: I1125 15:02:29.252914 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:29 crc kubenswrapper[4890]: I1125 15:02:29.867293 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 15:02:29 crc kubenswrapper[4890]: I1125 15:02:29.867540 4890 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 15:02:29 crc kubenswrapper[4890]: I1125 15:02:29.868826 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:29 crc kubenswrapper[4890]: I1125 15:02:29.868871 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:29 crc kubenswrapper[4890]: I1125 15:02:29.868884 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:30 crc kubenswrapper[4890]: I1125 15:02:30.830470 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 15:02:30 crc kubenswrapper[4890]: I1125 15:02:30.830680 4890 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 15:02:30 crc kubenswrapper[4890]: I1125 15:02:30.831942 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:30 crc kubenswrapper[4890]: I1125 15:02:30.831984 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:30 crc kubenswrapper[4890]: I1125 15:02:30.831996 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:31 crc kubenswrapper[4890]: I1125 15:02:31.100874 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Nov 25 15:02:31 crc kubenswrapper[4890]: I1125 15:02:31.101054 4890 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 15:02:31 crc kubenswrapper[4890]: I1125 15:02:31.104578 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:31 crc kubenswrapper[4890]: I1125 15:02:31.104625 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:31 crc kubenswrapper[4890]: I1125 15:02:31.104636 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:32 crc kubenswrapper[4890]: E1125 15:02:32.257635 4890 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 25 15:02:33 crc kubenswrapper[4890]: I1125 15:02:33.830783 4890 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 25 15:02:33 crc kubenswrapper[4890]: I1125 15:02:33.830891 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 25 15:02:34 crc kubenswrapper[4890]: I1125 15:02:34.348891 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 15:02:34 crc kubenswrapper[4890]: I1125 15:02:34.349039 4890 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 15:02:34 crc kubenswrapper[4890]: I1125 15:02:34.350496 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:34 crc kubenswrapper[4890]: I1125 15:02:34.350538 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:34 crc kubenswrapper[4890]: I1125 15:02:34.350549 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:35 crc kubenswrapper[4890]: W1125 15:02:35.750891 4890 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Nov 25 15:02:35 crc kubenswrapper[4890]: I1125 15:02:35.751110 4890 trace.go:236] Trace[1650310555]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (25-Nov-2025 15:02:25.749) (total time: 10001ms): Nov 25 15:02:35 crc kubenswrapper[4890]: Trace[1650310555]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (15:02:35.750) Nov 25 15:02:35 crc kubenswrapper[4890]: Trace[1650310555]: [10.001843788s] [10.001843788s] END Nov 25 15:02:35 crc kubenswrapper[4890]: E1125 15:02:35.751151 4890 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Nov 25 15:02:35 crc kubenswrapper[4890]: W1125 15:02:35.939760 4890 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Nov 25 15:02:35 crc kubenswrapper[4890]: I1125 15:02:35.939907 4890 trace.go:236] Trace[1317962712]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (25-Nov-2025 15:02:25.938) (total time: 10001ms): Nov 25 15:02:35 crc kubenswrapper[4890]: Trace[1317962712]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (15:02:35.939) Nov 25 15:02:35 crc kubenswrapper[4890]: Trace[1317962712]: [10.001300496s] [10.001300496s] END Nov 25 15:02:35 crc kubenswrapper[4890]: E1125 15:02:35.939940 4890 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Nov 25 15:02:36 crc kubenswrapper[4890]: I1125 15:02:36.132008 4890 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Nov 25 15:02:36 crc kubenswrapper[4890]: I1125 15:02:36.337588 4890 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 25 15:02:36 crc kubenswrapper[4890]: I1125 15:02:36.337652 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 25 15:02:36 crc kubenswrapper[4890]: I1125 15:02:36.348477 4890 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 25 15:02:36 crc kubenswrapper[4890]: I1125 15:02:36.348558 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 25 15:02:39 crc kubenswrapper[4890]: I1125 15:02:39.872054 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 15:02:39 crc kubenswrapper[4890]: I1125 15:02:39.872320 4890 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 15:02:39 crc kubenswrapper[4890]: I1125 15:02:39.873779 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:39 crc kubenswrapper[4890]: I1125 15:02:39.873857 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:39 crc kubenswrapper[4890]: I1125 15:02:39.873873 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:39 crc kubenswrapper[4890]: I1125 15:02:39.878024 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 15:02:40 crc kubenswrapper[4890]: I1125 15:02:40.050948 4890 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 25 15:02:40 crc kubenswrapper[4890]: I1125 15:02:40.120137 4890 apiserver.go:52] "Watching apiserver" Nov 25 15:02:40 crc kubenswrapper[4890]: I1125 15:02:40.127728 4890 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 25 15:02:40 crc kubenswrapper[4890]: I1125 15:02:40.128138 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h"] Nov 25 15:02:40 crc kubenswrapper[4890]: I1125 15:02:40.128633 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 15:02:40 crc kubenswrapper[4890]: I1125 15:02:40.128789 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 15:02:40 crc kubenswrapper[4890]: I1125 15:02:40.128794 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 15:02:40 crc kubenswrapper[4890]: E1125 15:02:40.128922 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 15:02:40 crc kubenswrapper[4890]: E1125 15:02:40.128984 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 15:02:40 crc kubenswrapper[4890]: I1125 15:02:40.129086 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 15:02:40 crc kubenswrapper[4890]: I1125 15:02:40.129114 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:02:40 crc kubenswrapper[4890]: E1125 15:02:40.129189 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 15:02:40 crc kubenswrapper[4890]: I1125 15:02:40.129381 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 15:02:40 crc kubenswrapper[4890]: I1125 15:02:40.131212 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 25 15:02:40 crc kubenswrapper[4890]: I1125 15:02:40.131214 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 25 15:02:40 crc kubenswrapper[4890]: I1125 15:02:40.133067 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 25 15:02:40 crc kubenswrapper[4890]: I1125 15:02:40.133152 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 25 15:02:40 crc kubenswrapper[4890]: I1125 15:02:40.133897 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 25 15:02:40 crc kubenswrapper[4890]: I1125 15:02:40.134532 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 25 15:02:40 crc kubenswrapper[4890]: I1125 15:02:40.134617 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 25 15:02:40 crc kubenswrapper[4890]: I1125 15:02:40.137014 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 25 15:02:40 crc kubenswrapper[4890]: I1125 15:02:40.137022 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 25 15:02:40 crc kubenswrapper[4890]: I1125 15:02:40.161808 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:40 crc kubenswrapper[4890]: I1125 15:02:40.174236 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:40 crc kubenswrapper[4890]: I1125 15:02:40.189152 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:40 crc kubenswrapper[4890]: I1125 15:02:40.201917 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:40 crc kubenswrapper[4890]: I1125 15:02:40.217826 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:40 crc kubenswrapper[4890]: I1125 15:02:40.223629 4890 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Nov 25 15:02:40 crc kubenswrapper[4890]: I1125 15:02:40.230699 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:40 crc kubenswrapper[4890]: I1125 15:02:40.242434 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:40 crc kubenswrapper[4890]: I1125 15:02:40.251235 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:40 crc kubenswrapper[4890]: I1125 15:02:40.292909 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.098952 4890 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.127953 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.140089 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.141000 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.142828 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.153980 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd5e648-6322-404d-bf1b-3b59a8092291\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76d277b4001db9d63b0af88b98afcc9019321101429c9be47b3022d3575d4e06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e3a186b14ea199d1aeba5751290c266ec918cb055120cfc5c8a2ae03b1469e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a860a998c40aee3e94e4cfd748f54391b3096fd3ed75c84c7ea2a5a7cefa33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://081a6d623cd043c0919d04e2ca93314b2b85fe33048634aa2637297d54b52cee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f506ec649ce24f4df6fd33e4aeafd13e42167dd71323524a5a3164517659fcc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.168950 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.181289 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.196388 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.209070 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.220233 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.231010 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.242624 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.254631 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.264662 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.276699 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.280927 4890 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.288780 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.300589 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd5e648-6322-404d-bf1b-3b59a8092291\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76d277b4001db9d63b0af88b98afcc9019321101429c9be47b3022d3575d4e06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e3a186b14ea199d1aeba5751290c266ec918cb055120cfc5c8a2ae03b1469e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a860a998c40aee3e94e4cfd748f54391b3096fd3ed75c84c7ea2a5a7cefa33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://081a6d623cd043c0919d04e2ca93314b2b85fe33048634aa2637297d54b52cee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f506ec649ce24f4df6fd33e4aeafd13e42167dd71323524a5a3164517659fcc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.317625 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86e031a7-b4c8-4030-a078-cdca7ccc9dde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4c4921f0973b03f6dade29c2ef32684a481cd1c3e1e23b64ed636943c9bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f9285acc7ac3cbb503a1e546728748c42971c8c9019cc8ca837076f3e55558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa240816a6e9d2821006ce240eccf27d8f57a7f0c5394314fa7177488da4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://659004bd47948def6b33498d8d1474f28c2e98986754f61af012267f3984269e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6d2dcd8df64e44df8cdb4c3a022f03fccb39451631ca35b30d4c7c35271ff0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:41 crc kubenswrapper[4890]: E1125 15:02:41.337391 4890 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.339204 4890 trace.go:236] Trace[1607312808]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (25-Nov-2025 15:02:29.582) (total time: 11756ms): Nov 25 15:02:41 crc kubenswrapper[4890]: Trace[1607312808]: ---"Objects listed" error: 11756ms (15:02:41.339) Nov 25 15:02:41 crc kubenswrapper[4890]: Trace[1607312808]: [11.756115585s] [11.756115585s] END Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.339248 4890 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.339433 4890 trace.go:236] Trace[458613185]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (25-Nov-2025 15:02:28.970) (total time: 12368ms): Nov 25 15:02:41 crc kubenswrapper[4890]: Trace[458613185]: ---"Objects listed" error: 12368ms (15:02:41.339) Nov 25 15:02:41 crc kubenswrapper[4890]: Trace[458613185]: [12.368405869s] [12.368405869s] END Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.339464 4890 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 25 15:02:41 crc kubenswrapper[4890]: E1125 15:02:41.340009 4890 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.340571 4890 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.383430 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.397634 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.413311 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd5e648-6322-404d-bf1b-3b59a8092291\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76d277b4001db9d63b0af88b98afcc9019321101429c9be47b3022d3575d4e06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e3a186b14ea199d1aeba5751290c266ec918cb055120cfc5c8a2ae03b1469e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a860a998c40aee3e94e4cfd748f54391b3096fd3ed75c84c7ea2a5a7cefa33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://081a6d623cd043c0919d04e2ca93314b2b85fe33048634aa2637297d54b52cee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f506ec649ce24f4df6fd33e4aeafd13e42167dd71323524a5a3164517659fcc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.433426 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86e031a7-b4c8-4030-a078-cdca7ccc9dde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4c4921f0973b03f6dade29c2ef32684a481cd1c3e1e23b64ed636943c9bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f9285acc7ac3cbb503a1e546728748c42971c8c9019cc8ca837076f3e55558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa240816a6e9d2821006ce240eccf27d8f57a7f0c5394314fa7177488da4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://659004bd47948def6b33498d8d1474f28c2e98986754f61af012267f3984269e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6d2dcd8df64e44df8cdb4c3a022f03fccb39451631ca35b30d4c7c35271ff0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.443804 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.443869 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.443890 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.443910 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.443929 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.443947 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.443985 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.444007 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.444026 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.444044 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.444071 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.444098 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.444120 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.444141 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.444183 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.444209 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.444230 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.444307 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.444327 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.444364 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.444381 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.444372 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.444397 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.444485 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.444545 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.444567 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.444549 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.444591 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.444613 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.444635 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.444655 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.444675 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.444696 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.444715 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.444733 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.444750 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.444760 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.444768 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.444820 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.444835 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.444869 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.444888 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.444904 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.444920 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.444938 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.444959 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.444976 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.444992 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.445010 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.445028 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.445044 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.445063 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.445080 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.445097 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.445115 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.445130 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.445147 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.445185 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.445203 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.445220 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.445241 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.445262 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.445345 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.445373 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.445392 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.445412 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.445435 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.445401 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.445455 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.445542 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.445582 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.445617 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.445618 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.445647 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.445676 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.445708 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.445712 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.445738 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.445772 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.445804 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.445793 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.445833 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.445906 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.445795 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.445983 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.446021 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.446046 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.446062 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.446074 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.446103 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.446137 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.446187 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.446193 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.446276 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.446282 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.446344 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.446382 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.446409 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.446433 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.446436 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.446457 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.446484 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.446505 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.446528 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.446554 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.446556 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.446584 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.446564 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.446610 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.446630 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.446636 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.446644 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.446681 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.446767 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.446791 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.446802 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.446817 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.446840 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.446862 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.446865 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.446885 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.446918 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.446944 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.446944 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.446957 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.446969 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.447059 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.447100 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.447123 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.447145 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.447193 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.447217 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.447251 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.447271 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.447293 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.447312 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.447333 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.447357 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.447381 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.447411 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.447438 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.447466 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.447491 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.447516 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.447541 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.447566 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.447591 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.447619 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.447648 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.447688 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.447709 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.447721 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.447734 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.447762 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.447784 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.447806 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.447837 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.447860 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.447882 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.447903 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.447924 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.448000 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.448027 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.448050 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.448319 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.448354 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.448352 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.448382 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.448482 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.448514 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.448547 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.448568 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.448586 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.448606 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.448629 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.448651 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.448687 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.448693 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.448715 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.448811 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.448839 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.448869 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.448893 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.448919 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.448938 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.448957 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.448975 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.448992 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.449020 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.449042 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.449059 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.449106 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.449135 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.449173 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.449196 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.449218 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.449238 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.449261 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.449279 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.449298 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.449316 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.449336 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.449355 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.449348 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.449375 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.449467 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.449503 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.449533 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.449560 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.449579 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.449588 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.449638 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.449662 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.449680 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.449681 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.449700 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.449725 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.449745 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.449768 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.449791 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.449829 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.449857 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.449933 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.449952 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.449983 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.450003 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.450040 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.450066 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.450090 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.450115 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.450138 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.450176 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.450197 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.450215 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.450236 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.450255 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.450273 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.450300 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.450320 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.450400 4890 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.450411 4890 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.450424 4890 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.450433 4890 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.450445 4890 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.450456 4890 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.450467 4890 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.450478 4890 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.450489 4890 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.450500 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.450511 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.450521 4890 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.450531 4890 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.450541 4890 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.450552 4890 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.450562 4890 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.450572 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.450582 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.450594 4890 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.450605 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.450615 4890 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.450625 4890 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.450635 4890 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.450644 4890 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.450656 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.450666 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.450676 4890 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.450685 4890 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.450695 4890 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.450705 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.457512 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.472949 4890 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.479121 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.449990 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.450077 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.450263 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.450316 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.450461 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.450546 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.451262 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.451447 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.451612 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.451858 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.452033 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.452262 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.454244 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.454552 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.455012 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.455227 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.455414 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.455448 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.455668 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.455827 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.455899 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.456070 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.456122 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.456200 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.456390 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.456390 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.456546 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.456666 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.456720 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.456820 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.456854 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.456928 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.457082 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.457484 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.457498 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.457526 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.457762 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.457759 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.457971 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.458065 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.458138 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.458573 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.458592 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.459636 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.460186 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.460244 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.460332 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.460339 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.460739 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.460804 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.460948 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: E1125 15:02:41.461339 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:02:41.961308713 +0000 UTC m=+20.403771323 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.461836 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.461782 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.461985 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.462254 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.462396 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.462940 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.463023 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.463035 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.467343 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.467607 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.467700 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.467930 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.468013 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.468232 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.468557 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.468594 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.468781 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.468905 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.468903 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.468942 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.469085 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.469231 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.469266 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.469312 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.469509 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.469523 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.469622 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.471276 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.472142 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.473344 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.474552 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.474816 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.474570 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.475542 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.475784 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.475253 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.475999 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.475256 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.476214 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.476221 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.476338 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.475905 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.476506 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.477374 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.477607 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.477633 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: E1125 15:02:41.477755 4890 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 15:02:41 crc kubenswrapper[4890]: E1125 15:02:41.480431 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 15:02:41.980404181 +0000 UTC m=+20.422866791 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.477928 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.478234 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.478281 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.478450 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.478572 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.478589 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.478782 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.478797 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: E1125 15:02:41.481182 4890 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 15:02:41 crc kubenswrapper[4890]: E1125 15:02:41.481262 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 15:02:41.981245301 +0000 UTC m=+20.423707911 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.481695 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.481740 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.480932 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.482518 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.483449 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.483456 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.483476 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.483812 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.483966 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.484329 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.484485 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.484490 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.484558 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.484794 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.485016 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.485386 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.485423 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.485764 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.485911 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.486109 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.486137 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.486989 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.487195 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.487352 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.487425 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.488248 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.484670 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.490985 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.491518 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.493253 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.493751 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.493758 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.493923 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.495570 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.496461 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:41 crc kubenswrapper[4890]: E1125 15:02:41.496519 4890 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 15:02:41 crc kubenswrapper[4890]: E1125 15:02:41.496539 4890 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 15:02:41 crc kubenswrapper[4890]: E1125 15:02:41.496553 4890 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 15:02:41 crc kubenswrapper[4890]: E1125 15:02:41.496642 4890 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 15:02:41 crc kubenswrapper[4890]: E1125 15:02:41.496671 4890 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 15:02:41 crc kubenswrapper[4890]: E1125 15:02:41.496684 4890 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.497614 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.497692 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.497543 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.497952 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: E1125 15:02:41.498067 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 15:02:41.99659957 +0000 UTC m=+20.439062180 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 15:02:41 crc kubenswrapper[4890]: E1125 15:02:41.498123 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 15:02:41.998093206 +0000 UTC m=+20.440555816 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.498229 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.498230 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.500004 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.506154 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.507101 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.513601 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.514762 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.515081 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.516346 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.516560 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.516780 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.516832 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.517535 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.517637 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.517744 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.518071 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.518213 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.518223 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.518220 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.520436 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.520885 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.521014 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.521509 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.521767 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.522218 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.524449 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-7g5s8"] Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.525380 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-7g5s8" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.527476 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.534051 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.534307 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.534465 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.534626 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.538575 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.545920 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.551631 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.551705 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4cs2\" (UniqueName: \"kubernetes.io/projected/7a7dc232-835b-491c-afc9-9eb14f257af3-kube-api-access-c4cs2\") pod \"node-resolver-7g5s8\" (UID: \"7a7dc232-835b-491c-afc9-9eb14f257af3\") " pod="openshift-dns/node-resolver-7g5s8" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.551746 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/7a7dc232-835b-491c-afc9-9eb14f257af3-hosts-file\") pod \"node-resolver-7g5s8\" (UID: \"7a7dc232-835b-491c-afc9-9eb14f257af3\") " pod="openshift-dns/node-resolver-7g5s8" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.551787 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.551872 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.551887 4890 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.551900 4890 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.551913 4890 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.551925 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.551937 4890 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.551950 4890 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.551975 4890 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.551994 4890 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552006 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552017 4890 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552029 4890 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552041 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552046 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552054 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552100 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552106 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552116 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552130 4890 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552143 4890 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552177 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552193 4890 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552206 4890 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552219 4890 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552230 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552242 4890 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552252 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552262 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552273 4890 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552282 4890 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552293 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552304 4890 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552316 4890 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552327 4890 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552339 4890 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552350 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552362 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552374 4890 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552391 4890 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552403 4890 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552418 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552430 4890 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552442 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552461 4890 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552477 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552489 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552433 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552506 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552657 4890 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552671 4890 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552682 4890 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552692 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552701 4890 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552711 4890 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552721 4890 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552729 4890 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552738 4890 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552747 4890 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552776 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552785 4890 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552796 4890 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552804 4890 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552813 4890 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552823 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552832 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552842 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552851 4890 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552861 4890 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552869 4890 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552876 4890 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552888 4890 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552897 4890 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552908 4890 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552920 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552932 4890 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552942 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552954 4890 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552965 4890 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552980 4890 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.552992 4890 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553002 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553013 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553025 4890 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553036 4890 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553048 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553060 4890 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553070 4890 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553080 4890 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553100 4890 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553117 4890 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553128 4890 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553139 4890 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553150 4890 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553180 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553188 4890 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553199 4890 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553212 4890 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553223 4890 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553234 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553244 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553256 4890 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553266 4890 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553277 4890 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553287 4890 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553298 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553309 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553320 4890 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553331 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553343 4890 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553357 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553368 4890 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553380 4890 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553391 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553402 4890 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553415 4890 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553425 4890 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553436 4890 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553447 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553458 4890 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553469 4890 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553483 4890 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553494 4890 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553506 4890 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553518 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553529 4890 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553540 4890 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553552 4890 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553564 4890 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553577 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553589 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553601 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553620 4890 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553631 4890 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553642 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553653 4890 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553665 4890 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553675 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553684 4890 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553692 4890 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553700 4890 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553709 4890 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553718 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553726 4890 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553735 4890 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553744 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553755 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553766 4890 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553776 4890 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553787 4890 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553797 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553807 4890 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553863 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553879 4890 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553902 4890 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553914 4890 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553927 4890 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553938 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553950 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553962 4890 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553973 4890 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553984 4890 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.553996 4890 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.554007 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.554019 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.554030 4890 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.563098 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.565708 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.571183 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd5e648-6322-404d-bf1b-3b59a8092291\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76d277b4001db9d63b0af88b98afcc9019321101429c9be47b3022d3575d4e06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e3a186b14ea199d1aeba5751290c266ec918cb055120cfc5c8a2ae03b1469e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a860a998c40aee3e94e4cfd748f54391b3096fd3ed75c84c7ea2a5a7cefa33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://081a6d623cd043c0919d04e2ca93314b2b85fe33048634aa2637297d54b52cee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f506ec649ce24f4df6fd33e4aeafd13e42167dd71323524a5a3164517659fcc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.592337 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86e031a7-b4c8-4030-a078-cdca7ccc9dde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4c4921f0973b03f6dade29c2ef32684a481cd1c3e1e23b64ed636943c9bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f9285acc7ac3cbb503a1e546728748c42971c8c9019cc8ca837076f3e55558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa240816a6e9d2821006ce240eccf27d8f57a7f0c5394314fa7177488da4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://659004bd47948def6b33498d8d1474f28c2e98986754f61af012267f3984269e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6d2dcd8df64e44df8cdb4c3a022f03fccb39451631ca35b30d4c7c35271ff0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.602858 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.614043 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.632086 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7g5s8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a7dc232-835b-491c-afc9-9eb14f257af3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4cs2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7g5s8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.647791 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.650611 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.654868 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4cs2\" (UniqueName: \"kubernetes.io/projected/7a7dc232-835b-491c-afc9-9eb14f257af3-kube-api-access-c4cs2\") pod \"node-resolver-7g5s8\" (UID: \"7a7dc232-835b-491c-afc9-9eb14f257af3\") " pod="openshift-dns/node-resolver-7g5s8" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.654932 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/7a7dc232-835b-491c-afc9-9eb14f257af3-hosts-file\") pod \"node-resolver-7g5s8\" (UID: \"7a7dc232-835b-491c-afc9-9eb14f257af3\") " pod="openshift-dns/node-resolver-7g5s8" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.654989 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.655005 4890 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.655075 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/7a7dc232-835b-491c-afc9-9eb14f257af3-hosts-file\") pod \"node-resolver-7g5s8\" (UID: \"7a7dc232-835b-491c-afc9-9eb14f257af3\") " pod="openshift-dns/node-resolver-7g5s8" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.659208 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 15:02:41 crc kubenswrapper[4890]: W1125 15:02:41.665726 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-929b2bbfc9c38fc070f625dca94ce700089b4763ef63e430978f0ae526c21b58 WatchSource:0}: Error finding container 929b2bbfc9c38fc070f625dca94ce700089b4763ef63e430978f0ae526c21b58: Status 404 returned error can't find the container with id 929b2bbfc9c38fc070f625dca94ce700089b4763ef63e430978f0ae526c21b58 Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.668967 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.673828 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4cs2\" (UniqueName: \"kubernetes.io/projected/7a7dc232-835b-491c-afc9-9eb14f257af3-kube-api-access-c4cs2\") pod \"node-resolver-7g5s8\" (UID: \"7a7dc232-835b-491c-afc9-9eb14f257af3\") " pod="openshift-dns/node-resolver-7g5s8" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.674456 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:41 crc kubenswrapper[4890]: W1125 15:02:41.684935 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-aa6d7429444aec1bc2d696f124e2ce8d7a46c60459cd064cba13cb2aa861159f WatchSource:0}: Error finding container aa6d7429444aec1bc2d696f124e2ce8d7a46c60459cd064cba13cb2aa861159f: Status 404 returned error can't find the container with id aa6d7429444aec1bc2d696f124e2ce8d7a46c60459cd064cba13cb2aa861159f Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.695400 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.821326 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.835777 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.852553 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd5e648-6322-404d-bf1b-3b59a8092291\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76d277b4001db9d63b0af88b98afcc9019321101429c9be47b3022d3575d4e06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e3a186b14ea199d1aeba5751290c266ec918cb055120cfc5c8a2ae03b1469e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a860a998c40aee3e94e4cfd748f54391b3096fd3ed75c84c7ea2a5a7cefa33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://081a6d623cd043c0919d04e2ca93314b2b85fe33048634aa2637297d54b52cee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f506ec649ce24f4df6fd33e4aeafd13e42167dd71323524a5a3164517659fcc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.855728 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-7g5s8" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.886044 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86e031a7-b4c8-4030-a078-cdca7ccc9dde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4c4921f0973b03f6dade29c2ef32684a481cd1c3e1e23b64ed636943c9bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f9285acc7ac3cbb503a1e546728748c42971c8c9019cc8ca837076f3e55558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa240816a6e9d2821006ce240eccf27d8f57a7f0c5394314fa7177488da4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://659004bd47948def6b33498d8d1474f28c2e98986754f61af012267f3984269e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6d2dcd8df64e44df8cdb4c3a022f03fccb39451631ca35b30d4c7c35271ff0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.897375 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.905178 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7g5s8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a7dc232-835b-491c-afc9-9eb14f257af3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4cs2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7g5s8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.909705 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.923665 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.925381 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-zg454"] Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.926209 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-zg454" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.930049 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7zdk8"] Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.934362 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.935491 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.935802 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.935897 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.939494 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.940252 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-xx2tz"] Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.940636 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-4txd2"] Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.941021 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-4txd2" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.941672 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.942125 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.946490 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.946711 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.946870 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.947119 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.947277 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.947439 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.947513 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.947571 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.949340 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.949536 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.949695 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.949879 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.949918 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.950438 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.950470 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.957064 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9a404958-dffa-4a6a-99fc-3e4950614bd4-ovnkube-script-lib\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.957098 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92hnd\" (UniqueName: \"kubernetes.io/projected/9a404958-dffa-4a6a-99fc-3e4950614bd4-kube-api-access-92hnd\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.957120 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4e4f849d-f239-4727-a73e-18327856929a-mcd-auth-proxy-config\") pod \"machine-config-daemon-xx2tz\" (UID: \"4e4f849d-f239-4727-a73e-18327856929a\") " pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.957151 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/81411ec7-f865-48e2-b464-ab6213ef5e52-cnibin\") pod \"multus-additional-cni-plugins-zg454\" (UID: \"81411ec7-f865-48e2-b464-ab6213ef5e52\") " pod="openshift-multus/multus-additional-cni-plugins-zg454" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.957182 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-host-run-multus-certs\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.957198 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/81411ec7-f865-48e2-b464-ab6213ef5e52-os-release\") pod \"multus-additional-cni-plugins-zg454\" (UID: \"81411ec7-f865-48e2-b464-ab6213ef5e52\") " pod="openshift-multus/multus-additional-cni-plugins-zg454" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.957216 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-host-run-netns\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.957233 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-node-log\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.957251 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9a404958-dffa-4a6a-99fc-3e4950614bd4-env-overrides\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.957278 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-cni-binary-copy\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.957365 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-host-slash\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.957428 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-log-socket\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.957460 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/4e4f849d-f239-4727-a73e-18327856929a-rootfs\") pod \"machine-config-daemon-xx2tz\" (UID: \"4e4f849d-f239-4727-a73e-18327856929a\") " pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.957487 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/81411ec7-f865-48e2-b464-ab6213ef5e52-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-zg454\" (UID: \"81411ec7-f865-48e2-b464-ab6213ef5e52\") " pod="openshift-multus/multus-additional-cni-plugins-zg454" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.957517 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-multus-daemon-config\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.957547 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-run-ovn\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.957570 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-host-cni-bin\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.957596 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-cnibin\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.957692 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-var-lib-openvswitch\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.957717 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-systemd-units\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.957747 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-etc-openvswitch\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.957772 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9a404958-dffa-4a6a-99fc-3e4950614bd4-ovnkube-config\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.957799 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-host-var-lib-cni-bin\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.957824 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9a404958-dffa-4a6a-99fc-3e4950614bd4-ovn-node-metrics-cert\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.957845 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-os-release\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.957872 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-host-kubelet\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.957896 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-host-cni-netd\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.957921 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4e4f849d-f239-4727-a73e-18327856929a-proxy-tls\") pod \"machine-config-daemon-xx2tz\" (UID: \"4e4f849d-f239-4727-a73e-18327856929a\") " pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.957951 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-host-run-netns\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.957973 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-etc-kubernetes\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.958003 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/81411ec7-f865-48e2-b464-ab6213ef5e52-tuning-conf-dir\") pod \"multus-additional-cni-plugins-zg454\" (UID: \"81411ec7-f865-48e2-b464-ab6213ef5e52\") " pod="openshift-multus/multus-additional-cni-plugins-zg454" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.958066 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4g7z\" (UniqueName: \"kubernetes.io/projected/81411ec7-f865-48e2-b464-ab6213ef5e52-kube-api-access-l4g7z\") pod \"multus-additional-cni-plugins-zg454\" (UID: \"81411ec7-f865-48e2-b464-ab6213ef5e52\") " pod="openshift-multus/multus-additional-cni-plugins-zg454" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.958093 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfcq7\" (UniqueName: \"kubernetes.io/projected/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-kube-api-access-jfcq7\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.958122 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-run-systemd\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.958147 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-run-openvswitch\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.958202 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-host-run-ovn-kubernetes\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.958278 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-host-run-k8s-cni-cncf-io\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.958305 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-multus-conf-dir\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.958331 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/81411ec7-f865-48e2-b464-ab6213ef5e52-cni-binary-copy\") pod \"multus-additional-cni-plugins-zg454\" (UID: \"81411ec7-f865-48e2-b464-ab6213ef5e52\") " pod="openshift-multus/multus-additional-cni-plugins-zg454" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.958357 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-multus-cni-dir\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.958398 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x828g\" (UniqueName: \"kubernetes.io/projected/4e4f849d-f239-4727-a73e-18327856929a-kube-api-access-x828g\") pod \"machine-config-daemon-xx2tz\" (UID: \"4e4f849d-f239-4727-a73e-18327856929a\") " pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.958420 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/81411ec7-f865-48e2-b464-ab6213ef5e52-system-cni-dir\") pod \"multus-additional-cni-plugins-zg454\" (UID: \"81411ec7-f865-48e2-b464-ab6213ef5e52\") " pod="openshift-multus/multus-additional-cni-plugins-zg454" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.958466 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-host-var-lib-kubelet\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.958491 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-hostroot\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.958516 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.958544 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-system-cni-dir\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.958565 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-multus-socket-dir-parent\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.958587 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-host-var-lib-cni-multus\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.965450 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.978459 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:41 crc kubenswrapper[4890]: I1125 15:02:41.993373 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.018391 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86e031a7-b4c8-4030-a078-cdca7ccc9dde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4c4921f0973b03f6dade29c2ef32684a481cd1c3e1e23b64ed636943c9bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f9285acc7ac3cbb503a1e546728748c42971c8c9019cc8ca837076f3e55558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa240816a6e9d2821006ce240eccf27d8f57a7f0c5394314fa7177488da4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://659004bd47948def6b33498d8d1474f28c2e98986754f61af012267f3984269e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6d2dcd8df64e44df8cdb4c3a022f03fccb39451631ca35b30d4c7c35271ff0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.038560 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg454" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81411ec7-f865-48e2-b464-ab6213ef5e52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg454\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.051892 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4txd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfcq7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4txd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.059879 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.060025 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.060078 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-host-run-k8s-cni-cncf-io\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.060123 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-multus-conf-dir\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.060146 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.060204 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x828g\" (UniqueName: \"kubernetes.io/projected/4e4f849d-f239-4727-a73e-18327856929a-kube-api-access-x828g\") pod \"machine-config-daemon-xx2tz\" (UID: \"4e4f849d-f239-4727-a73e-18327856929a\") " pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.060221 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/81411ec7-f865-48e2-b464-ab6213ef5e52-system-cni-dir\") pod \"multus-additional-cni-plugins-zg454\" (UID: \"81411ec7-f865-48e2-b464-ab6213ef5e52\") " pod="openshift-multus/multus-additional-cni-plugins-zg454" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.060239 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/81411ec7-f865-48e2-b464-ab6213ef5e52-cni-binary-copy\") pod \"multus-additional-cni-plugins-zg454\" (UID: \"81411ec7-f865-48e2-b464-ab6213ef5e52\") " pod="openshift-multus/multus-additional-cni-plugins-zg454" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.060278 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-multus-cni-dir\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.060296 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.060325 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-system-cni-dir\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.060368 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-multus-socket-dir-parent\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.060384 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-host-var-lib-cni-multus\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.060400 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-host-var-lib-kubelet\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.060431 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-hostroot\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.060452 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9a404958-dffa-4a6a-99fc-3e4950614bd4-ovnkube-script-lib\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.060468 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92hnd\" (UniqueName: \"kubernetes.io/projected/9a404958-dffa-4a6a-99fc-3e4950614bd4-kube-api-access-92hnd\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.060486 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4e4f849d-f239-4727-a73e-18327856929a-mcd-auth-proxy-config\") pod \"machine-config-daemon-xx2tz\" (UID: \"4e4f849d-f239-4727-a73e-18327856929a\") " pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.060518 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/81411ec7-f865-48e2-b464-ab6213ef5e52-cnibin\") pod \"multus-additional-cni-plugins-zg454\" (UID: \"81411ec7-f865-48e2-b464-ab6213ef5e52\") " pod="openshift-multus/multus-additional-cni-plugins-zg454" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.060534 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-host-run-multus-certs\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.060534 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/81411ec7-f865-48e2-b464-ab6213ef5e52-system-cni-dir\") pod \"multus-additional-cni-plugins-zg454\" (UID: \"81411ec7-f865-48e2-b464-ab6213ef5e52\") " pod="openshift-multus/multus-additional-cni-plugins-zg454" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.060549 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/81411ec7-f865-48e2-b464-ab6213ef5e52-os-release\") pod \"multus-additional-cni-plugins-zg454\" (UID: \"81411ec7-f865-48e2-b464-ab6213ef5e52\") " pod="openshift-multus/multus-additional-cni-plugins-zg454" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.060622 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-node-log\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.060641 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9a404958-dffa-4a6a-99fc-3e4950614bd4-env-overrides\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.060659 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-host-run-netns\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.060688 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.060706 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-cni-binary-copy\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.060723 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-host-slash\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.060749 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-log-socket\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.060766 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/4e4f849d-f239-4727-a73e-18327856929a-rootfs\") pod \"machine-config-daemon-xx2tz\" (UID: \"4e4f849d-f239-4727-a73e-18327856929a\") " pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.060791 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/81411ec7-f865-48e2-b464-ab6213ef5e52-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-zg454\" (UID: \"81411ec7-f865-48e2-b464-ab6213ef5e52\") " pod="openshift-multus/multus-additional-cni-plugins-zg454" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.060807 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-multus-daemon-config\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.060825 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-run-ovn\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.060842 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-host-cni-bin\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.060907 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.060923 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-var-lib-openvswitch\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.060944 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-cnibin\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.060960 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-etc-openvswitch\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.060973 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9a404958-dffa-4a6a-99fc-3e4950614bd4-ovnkube-config\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.060976 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/81411ec7-f865-48e2-b464-ab6213ef5e52-os-release\") pod \"multus-additional-cni-plugins-zg454\" (UID: \"81411ec7-f865-48e2-b464-ab6213ef5e52\") " pod="openshift-multus/multus-additional-cni-plugins-zg454" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.060991 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-host-var-lib-cni-bin\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.061007 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-systemd-units\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.061029 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-host-kubelet\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.061044 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-host-cni-netd\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.061047 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-host-var-lib-kubelet\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.061059 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9a404958-dffa-4a6a-99fc-3e4950614bd4-ovn-node-metrics-cert\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.061076 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-os-release\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.061101 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4e4f849d-f239-4727-a73e-18327856929a-proxy-tls\") pod \"machine-config-daemon-xx2tz\" (UID: \"4e4f849d-f239-4727-a73e-18327856929a\") " pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.061118 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-host-run-netns\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.061132 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-etc-kubernetes\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.061150 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-run-systemd\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.061178 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-run-openvswitch\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.061195 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-host-run-ovn-kubernetes\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.061210 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/81411ec7-f865-48e2-b464-ab6213ef5e52-tuning-conf-dir\") pod \"multus-additional-cni-plugins-zg454\" (UID: \"81411ec7-f865-48e2-b464-ab6213ef5e52\") " pod="openshift-multus/multus-additional-cni-plugins-zg454" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.061225 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4g7z\" (UniqueName: \"kubernetes.io/projected/81411ec7-f865-48e2-b464-ab6213ef5e52-kube-api-access-l4g7z\") pod \"multus-additional-cni-plugins-zg454\" (UID: \"81411ec7-f865-48e2-b464-ab6213ef5e52\") " pod="openshift-multus/multus-additional-cni-plugins-zg454" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.061258 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfcq7\" (UniqueName: \"kubernetes.io/projected/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-kube-api-access-jfcq7\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:42 crc kubenswrapper[4890]: E1125 15:02:42.061616 4890 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 15:02:42 crc kubenswrapper[4890]: E1125 15:02:42.061650 4890 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 15:02:42 crc kubenswrapper[4890]: E1125 15:02:42.061664 4890 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.061676 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-cnibin\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.061705 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-etc-openvswitch\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:42 crc kubenswrapper[4890]: E1125 15:02:42.061719 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 15:02:43.061703418 +0000 UTC m=+21.504166028 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 15:02:42 crc kubenswrapper[4890]: E1125 15:02:42.061784 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:02:43.061777579 +0000 UTC m=+21.504240189 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:02:42 crc kubenswrapper[4890]: E1125 15:02:42.061834 4890 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 15:02:42 crc kubenswrapper[4890]: E1125 15:02:42.061843 4890 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 15:02:42 crc kubenswrapper[4890]: E1125 15:02:42.061851 4890 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 15:02:42 crc kubenswrapper[4890]: E1125 15:02:42.061872 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 15:02:43.061865891 +0000 UTC m=+21.504328491 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.061869 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/81411ec7-f865-48e2-b464-ab6213ef5e52-cni-binary-copy\") pod \"multus-additional-cni-plugins-zg454\" (UID: \"81411ec7-f865-48e2-b464-ab6213ef5e52\") " pod="openshift-multus/multus-additional-cni-plugins-zg454" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.061931 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-host-run-k8s-cni-cncf-io\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.061962 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-multus-conf-dir\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:42 crc kubenswrapper[4890]: E1125 15:02:42.062000 4890 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 15:02:42 crc kubenswrapper[4890]: E1125 15:02:42.062022 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 15:02:43.062015585 +0000 UTC m=+21.504478195 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.062057 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-multus-cni-dir\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.062098 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.062151 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-system-cni-dir\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.062218 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-multus-socket-dir-parent\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.062267 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-host-var-lib-cni-multus\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.062292 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-node-log\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.062295 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/4e4f849d-f239-4727-a73e-18327856929a-rootfs\") pod \"machine-config-daemon-xx2tz\" (UID: \"4e4f849d-f239-4727-a73e-18327856929a\") " pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.062343 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9a404958-dffa-4a6a-99fc-3e4950614bd4-ovnkube-config\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.062382 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-host-var-lib-cni-bin\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.062408 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-systemd-units\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.062429 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-host-kubelet\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.062455 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-host-cni-netd\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.062752 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9a404958-dffa-4a6a-99fc-3e4950614bd4-env-overrides\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.062750 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-hostroot\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.062808 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-host-run-netns\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:42 crc kubenswrapper[4890]: E1125 15:02:42.062872 4890 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 15:02:42 crc kubenswrapper[4890]: E1125 15:02:42.062915 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 15:02:43.062903456 +0000 UTC m=+21.505366066 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.063118 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-multus-daemon-config\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.063181 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-run-ovn\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.063219 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-host-cni-bin\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.061634 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-var-lib-openvswitch\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.063250 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9a404958-dffa-4a6a-99fc-3e4950614bd4-ovnkube-script-lib\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.063386 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-cni-binary-copy\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.063427 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-host-slash\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.063451 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-log-socket\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.063549 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-run-systemd\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.063589 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-os-release\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.063636 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.063780 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-host-run-ovn-kubernetes\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.063822 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-run-openvswitch\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.063900 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4e4f849d-f239-4727-a73e-18327856929a-mcd-auth-proxy-config\") pod \"machine-config-daemon-xx2tz\" (UID: \"4e4f849d-f239-4727-a73e-18327856929a\") " pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.063952 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/81411ec7-f865-48e2-b464-ab6213ef5e52-cnibin\") pod \"multus-additional-cni-plugins-zg454\" (UID: \"81411ec7-f865-48e2-b464-ab6213ef5e52\") " pod="openshift-multus/multus-additional-cni-plugins-zg454" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.064010 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-host-run-multus-certs\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.064184 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-host-run-netns\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.064213 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-etc-kubernetes\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.066559 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/81411ec7-f865-48e2-b464-ab6213ef5e52-tuning-conf-dir\") pod \"multus-additional-cni-plugins-zg454\" (UID: \"81411ec7-f865-48e2-b464-ab6213ef5e52\") " pod="openshift-multus/multus-additional-cni-plugins-zg454" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.066946 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/81411ec7-f865-48e2-b464-ab6213ef5e52-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-zg454\" (UID: \"81411ec7-f865-48e2-b464-ab6213ef5e52\") " pod="openshift-multus/multus-additional-cni-plugins-zg454" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.069235 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4e4f849d-f239-4727-a73e-18327856929a-proxy-tls\") pod \"machine-config-daemon-xx2tz\" (UID: \"4e4f849d-f239-4727-a73e-18327856929a\") " pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.069666 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9a404958-dffa-4a6a-99fc-3e4950614bd4-ovn-node-metrics-cert\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.076247 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.079183 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4g7z\" (UniqueName: \"kubernetes.io/projected/81411ec7-f865-48e2-b464-ab6213ef5e52-kube-api-access-l4g7z\") pod \"multus-additional-cni-plugins-zg454\" (UID: \"81411ec7-f865-48e2-b464-ab6213ef5e52\") " pod="openshift-multus/multus-additional-cni-plugins-zg454" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.083240 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfcq7\" (UniqueName: \"kubernetes.io/projected/0fcdeaed-8701-42ad-8534-a77b7f3f93f6-kube-api-access-jfcq7\") pod \"multus-4txd2\" (UID: \"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\") " pod="openshift-multus/multus-4txd2" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.084085 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92hnd\" (UniqueName: \"kubernetes.io/projected/9a404958-dffa-4a6a-99fc-3e4950614bd4-kube-api-access-92hnd\") pod \"ovnkube-node-7zdk8\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.088114 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x828g\" (UniqueName: \"kubernetes.io/projected/4e4f849d-f239-4727-a73e-18327856929a-kube-api-access-x828g\") pod \"machine-config-daemon-xx2tz\" (UID: \"4e4f849d-f239-4727-a73e-18327856929a\") " pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.089420 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7g5s8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a7dc232-835b-491c-afc9-9eb14f257af3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4cs2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7g5s8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.110236 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a404958-dffa-4a6a-99fc-3e4950614bd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zdk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.126627 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e4f849d-f239-4727-a73e-18327856929a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xx2tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.139133 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd5e648-6322-404d-bf1b-3b59a8092291\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76d277b4001db9d63b0af88b98afcc9019321101429c9be47b3022d3575d4e06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e3a186b14ea199d1aeba5751290c266ec918cb055120cfc5c8a2ae03b1469e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a860a998c40aee3e94e4cfd748f54391b3096fd3ed75c84c7ea2a5a7cefa33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://081a6d623cd043c0919d04e2ca93314b2b85fe33048634aa2637297d54b52cee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f506ec649ce24f4df6fd33e4aeafd13e42167dd71323524a5a3164517659fcc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.194443 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 15:02:42 crc kubenswrapper[4890]: E1125 15:02:42.194614 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.195120 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 15:02:42 crc kubenswrapper[4890]: E1125 15:02:42.195206 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.197400 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:02:42 crc kubenswrapper[4890]: E1125 15:02:42.197573 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.200029 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.200892 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.202953 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.203818 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.205719 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.206662 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.208001 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d5a9ccb-13ba-4dd5-9904-42aac5eca374\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d19bc48acb392051f0f6c5a6820135bf29ebe6d2afd098697553298933cf280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93325a6407389c4a7db0ba3c1a9c84a6980db3c3277cdf3860583a564b562bbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20505e91fc60158eeb3a8cfa0aafc5b4e4436d5def5e735a3a58b0f9f3621058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3281314d85a20945d56b75ab7b7aef1e382a73e74dff03773d118d03eb0ee16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.208893 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.213253 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.218900 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.220285 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.220898 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.222110 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.222858 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.223887 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.224477 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.225692 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.229394 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.230962 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.231632 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.232735 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.233384 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.234013 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.235057 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.235532 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.236883 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.237455 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.238651 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.239308 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.239529 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.239773 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.242995 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.243600 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.244110 4890 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.244246 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.245713 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.246330 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.248734 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.250440 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.251189 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.252146 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.252908 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.254210 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.254696 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.255796 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.256778 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.257965 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.258535 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.258966 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-zg454" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.259677 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.260454 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.261767 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.262401 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.263312 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.263776 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.264352 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.265438 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.265944 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.266625 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.270943 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-4txd2" Nov 25 15:02:42 crc kubenswrapper[4890]: W1125 15:02:42.272714 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod81411ec7_f865_48e2_b464_ab6213ef5e52.slice/crio-17917529dae1b421d07c892a21737e67441d14b53402d4108598606d7614d611 WatchSource:0}: Error finding container 17917529dae1b421d07c892a21737e67441d14b53402d4108598606d7614d611: Status 404 returned error can't find the container with id 17917529dae1b421d07c892a21737e67441d14b53402d4108598606d7614d611 Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.281243 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.288410 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.289757 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"583964bf462fa0005d471f82f6827552a60659ab97984eb1faf074c7b82c2189"} Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.289811 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"929b2bbfc9c38fc070f625dca94ce700089b4763ef63e430978f0ae526c21b58"} Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.290037 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.291286 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zg454" event={"ID":"81411ec7-f865-48e2-b464-ab6213ef5e52","Type":"ContainerStarted","Data":"17917529dae1b421d07c892a21737e67441d14b53402d4108598606d7614d611"} Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.299205 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.303537 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"3358a356ae5b422d6bba72bc516374e14ccdace80726aa71b10e25102e0f1100"} Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.303613 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"ba3acdae961f49f797f66bbebe350e34a4df5882bcf2a4fe8d37f2ca95865c11"} Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.303629 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"743f554d07f750571115c2d265ff718968ca2d691ebffc759559acffdac6ca9e"} Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.317548 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.329460 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-7g5s8" event={"ID":"7a7dc232-835b-491c-afc9-9eb14f257af3","Type":"ContainerStarted","Data":"e9a7ac317fbb9c589573e80babe484c21bef4da3d9ab79dfb7ebf5b8300c472c"} Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.329541 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-7g5s8" event={"ID":"7a7dc232-835b-491c-afc9-9eb14f257af3","Type":"ContainerStarted","Data":"c5c7d7b2f814dcc67b1ffc136d0c01a43ca48a94fe2693b215b83610f9ec78d4"} Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.334336 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7g5s8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a7dc232-835b-491c-afc9-9eb14f257af3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4cs2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7g5s8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.337109 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"aa6d7429444aec1bc2d696f124e2ce8d7a46c60459cd064cba13cb2aa861159f"} Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.363618 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a404958-dffa-4a6a-99fc-3e4950614bd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zdk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.377216 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e4f849d-f239-4727-a73e-18327856929a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xx2tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.442004 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd5e648-6322-404d-bf1b-3b59a8092291\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76d277b4001db9d63b0af88b98afcc9019321101429c9be47b3022d3575d4e06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e3a186b14ea199d1aeba5751290c266ec918cb055120cfc5c8a2ae03b1469e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a860a998c40aee3e94e4cfd748f54391b3096fd3ed75c84c7ea2a5a7cefa33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://081a6d623cd043c0919d04e2ca93314b2b85fe33048634aa2637297d54b52cee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f506ec649ce24f4df6fd33e4aeafd13e42167dd71323524a5a3164517659fcc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.460297 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d5a9ccb-13ba-4dd5-9904-42aac5eca374\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d19bc48acb392051f0f6c5a6820135bf29ebe6d2afd098697553298933cf280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93325a6407389c4a7db0ba3c1a9c84a6980db3c3277cdf3860583a564b562bbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20505e91fc60158eeb3a8cfa0aafc5b4e4436d5def5e735a3a58b0f9f3621058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3281314d85a20945d56b75ab7b7aef1e382a73e74dff03773d118d03eb0ee16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.474053 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.512541 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.539625 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.584630 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.627098 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86e031a7-b4c8-4030-a078-cdca7ccc9dde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4c4921f0973b03f6dade29c2ef32684a481cd1c3e1e23b64ed636943c9bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f9285acc7ac3cbb503a1e546728748c42971c8c9019cc8ca837076f3e55558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa240816a6e9d2821006ce240eccf27d8f57a7f0c5394314fa7177488da4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://659004bd47948def6b33498d8d1474f28c2e98986754f61af012267f3984269e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6d2dcd8df64e44df8cdb4c3a022f03fccb39451631ca35b30d4c7c35271ff0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.662524 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg454" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81411ec7-f865-48e2-b464-ab6213ef5e52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg454\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.700064 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4txd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfcq7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4txd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.737021 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.778870 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd5e648-6322-404d-bf1b-3b59a8092291\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76d277b4001db9d63b0af88b98afcc9019321101429c9be47b3022d3575d4e06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e3a186b14ea199d1aeba5751290c266ec918cb055120cfc5c8a2ae03b1469e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a860a998c40aee3e94e4cfd748f54391b3096fd3ed75c84c7ea2a5a7cefa33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://081a6d623cd043c0919d04e2ca93314b2b85fe33048634aa2637297d54b52cee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f506ec649ce24f4df6fd33e4aeafd13e42167dd71323524a5a3164517659fcc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.816763 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d5a9ccb-13ba-4dd5-9904-42aac5eca374\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d19bc48acb392051f0f6c5a6820135bf29ebe6d2afd098697553298933cf280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93325a6407389c4a7db0ba3c1a9c84a6980db3c3277cdf3860583a564b562bbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20505e91fc60158eeb3a8cfa0aafc5b4e4436d5def5e735a3a58b0f9f3621058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3281314d85a20945d56b75ab7b7aef1e382a73e74dff03773d118d03eb0ee16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.856828 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3358a356ae5b422d6bba72bc516374e14ccdace80726aa71b10e25102e0f1100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba3acdae961f49f797f66bbebe350e34a4df5882bcf2a4fe8d37f2ca95865c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.896266 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.936669 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:42 crc kubenswrapper[4890]: I1125 15:02:42.982900 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86e031a7-b4c8-4030-a078-cdca7ccc9dde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4c4921f0973b03f6dade29c2ef32684a481cd1c3e1e23b64ed636943c9bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f9285acc7ac3cbb503a1e546728748c42971c8c9019cc8ca837076f3e55558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa240816a6e9d2821006ce240eccf27d8f57a7f0c5394314fa7177488da4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://659004bd47948def6b33498d8d1474f28c2e98986754f61af012267f3984269e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6d2dcd8df64e44df8cdb4c3a022f03fccb39451631ca35b30d4c7c35271ff0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:43 crc kubenswrapper[4890]: I1125 15:02:43.020304 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg454" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81411ec7-f865-48e2-b464-ab6213ef5e52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg454\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:43 crc kubenswrapper[4890]: I1125 15:02:43.060542 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4txd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfcq7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4txd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:43 crc kubenswrapper[4890]: I1125 15:02:43.102096 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a404958-dffa-4a6a-99fc-3e4950614bd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zdk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:43 crc kubenswrapper[4890]: I1125 15:02:43.103850 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:02:43 crc kubenswrapper[4890]: E1125 15:02:43.104035 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:02:45.104009634 +0000 UTC m=+23.546472254 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:02:43 crc kubenswrapper[4890]: I1125 15:02:43.104099 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:02:43 crc kubenswrapper[4890]: I1125 15:02:43.104196 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:02:43 crc kubenswrapper[4890]: E1125 15:02:43.104242 4890 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 15:02:43 crc kubenswrapper[4890]: E1125 15:02:43.104370 4890 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 15:02:43 crc kubenswrapper[4890]: I1125 15:02:43.104449 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 15:02:43 crc kubenswrapper[4890]: I1125 15:02:43.104519 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 15:02:43 crc kubenswrapper[4890]: E1125 15:02:43.104567 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 15:02:45.104551777 +0000 UTC m=+23.547014607 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 15:02:43 crc kubenswrapper[4890]: E1125 15:02:43.104706 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 15:02:45.104697111 +0000 UTC m=+23.547159721 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 15:02:43 crc kubenswrapper[4890]: E1125 15:02:43.104600 4890 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 15:02:43 crc kubenswrapper[4890]: E1125 15:02:43.104762 4890 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 15:02:43 crc kubenswrapper[4890]: E1125 15:02:43.104777 4890 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 15:02:43 crc kubenswrapper[4890]: E1125 15:02:43.104644 4890 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 15:02:43 crc kubenswrapper[4890]: E1125 15:02:43.104824 4890 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 15:02:43 crc kubenswrapper[4890]: E1125 15:02:43.104834 4890 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 15:02:43 crc kubenswrapper[4890]: E1125 15:02:43.104872 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 15:02:45.104862205 +0000 UTC m=+23.547324815 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 15:02:43 crc kubenswrapper[4890]: E1125 15:02:43.104907 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 15:02:45.104899176 +0000 UTC m=+23.547361786 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 15:02:43 crc kubenswrapper[4890]: I1125 15:02:43.136707 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e4f849d-f239-4727-a73e-18327856929a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xx2tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 15:02:43 crc kubenswrapper[4890]: I1125 15:02:43.182213 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583964bf462fa0005d471f82f6827552a60659ab97984eb1faf074c7b82c2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:43Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:43 crc kubenswrapper[4890]: I1125 15:02:43.221102 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:43Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:43 crc kubenswrapper[4890]: I1125 15:02:43.261129 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7g5s8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a7dc232-835b-491c-afc9-9eb14f257af3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9a7ac317fbb9c589573e80babe484c21bef4da3d9ab79dfb7ebf5b8300c472c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4cs2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7g5s8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:43Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:43 crc kubenswrapper[4890]: I1125 15:02:43.341451 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" event={"ID":"4e4f849d-f239-4727-a73e-18327856929a","Type":"ContainerStarted","Data":"490f1595bc77030d2ee789197fa3748324e1b7a0c9d9180ba08383bbac43adc6"} Nov 25 15:02:43 crc kubenswrapper[4890]: I1125 15:02:43.341506 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" event={"ID":"4e4f849d-f239-4727-a73e-18327856929a","Type":"ContainerStarted","Data":"0902d6f8e56aecc5a2a13a7148f94a8a00ad139027b3c036aba0215ba0a67f59"} Nov 25 15:02:43 crc kubenswrapper[4890]: I1125 15:02:43.341518 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" event={"ID":"4e4f849d-f239-4727-a73e-18327856929a","Type":"ContainerStarted","Data":"f96f57fbbfd52fb4143768b34ade00170ac4c7319399905b22b3b005e3b1d2a8"} Nov 25 15:02:43 crc kubenswrapper[4890]: I1125 15:02:43.343668 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4txd2" event={"ID":"0fcdeaed-8701-42ad-8534-a77b7f3f93f6","Type":"ContainerStarted","Data":"b4322dbd78237bf443a985ce3c79bca4cba2103862d81177f5474a76f4905eed"} Nov 25 15:02:43 crc kubenswrapper[4890]: I1125 15:02:43.343705 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4txd2" event={"ID":"0fcdeaed-8701-42ad-8534-a77b7f3f93f6","Type":"ContainerStarted","Data":"3bfc03f520776429ac1069abc3f1892530df504d72f779162c4b254392bfb90d"} Nov 25 15:02:43 crc kubenswrapper[4890]: I1125 15:02:43.345695 4890 generic.go:334] "Generic (PLEG): container finished" podID="81411ec7-f865-48e2-b464-ab6213ef5e52" containerID="6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17" exitCode=0 Nov 25 15:02:43 crc kubenswrapper[4890]: I1125 15:02:43.345778 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zg454" event={"ID":"81411ec7-f865-48e2-b464-ab6213ef5e52","Type":"ContainerDied","Data":"6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17"} Nov 25 15:02:43 crc kubenswrapper[4890]: I1125 15:02:43.347446 4890 generic.go:334] "Generic (PLEG): container finished" podID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerID="50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0" exitCode=0 Nov 25 15:02:43 crc kubenswrapper[4890]: I1125 15:02:43.347494 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" event={"ID":"9a404958-dffa-4a6a-99fc-3e4950614bd4","Type":"ContainerDied","Data":"50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0"} Nov 25 15:02:43 crc kubenswrapper[4890]: I1125 15:02:43.347540 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" event={"ID":"9a404958-dffa-4a6a-99fc-3e4950614bd4","Type":"ContainerStarted","Data":"006be7dbdea159ba7aed943d4fdf527a2dfec84e73e9e193e2f96a4465006a9b"} Nov 25 15:02:43 crc kubenswrapper[4890]: I1125 15:02:43.364390 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd5e648-6322-404d-bf1b-3b59a8092291\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76d277b4001db9d63b0af88b98afcc9019321101429c9be47b3022d3575d4e06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e3a186b14ea199d1aeba5751290c266ec918cb055120cfc5c8a2ae03b1469e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a860a998c40aee3e94e4cfd748f54391b3096fd3ed75c84c7ea2a5a7cefa33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://081a6d623cd043c0919d04e2ca93314b2b85fe33048634aa2637297d54b52cee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f506ec649ce24f4df6fd33e4aeafd13e42167dd71323524a5a3164517659fcc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:43Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:43 crc kubenswrapper[4890]: I1125 15:02:43.387971 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d5a9ccb-13ba-4dd5-9904-42aac5eca374\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d19bc48acb392051f0f6c5a6820135bf29ebe6d2afd098697553298933cf280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93325a6407389c4a7db0ba3c1a9c84a6980db3c3277cdf3860583a564b562bbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20505e91fc60158eeb3a8cfa0aafc5b4e4436d5def5e735a3a58b0f9f3621058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3281314d85a20945d56b75ab7b7aef1e382a73e74dff03773d118d03eb0ee16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:43Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:43 crc kubenswrapper[4890]: I1125 15:02:43.402193 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:43Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:43 crc kubenswrapper[4890]: I1125 15:02:43.423363 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:43Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:43 crc kubenswrapper[4890]: I1125 15:02:43.461413 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:43Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:43 crc kubenswrapper[4890]: I1125 15:02:43.500153 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3358a356ae5b422d6bba72bc516374e14ccdace80726aa71b10e25102e0f1100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba3acdae961f49f797f66bbebe350e34a4df5882bcf2a4fe8d37f2ca95865c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:43Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:43 crc kubenswrapper[4890]: I1125 15:02:43.544248 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86e031a7-b4c8-4030-a078-cdca7ccc9dde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4c4921f0973b03f6dade29c2ef32684a481cd1c3e1e23b64ed636943c9bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f9285acc7ac3cbb503a1e546728748c42971c8c9019cc8ca837076f3e55558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa240816a6e9d2821006ce240eccf27d8f57a7f0c5394314fa7177488da4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://659004bd47948def6b33498d8d1474f28c2e98986754f61af012267f3984269e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6d2dcd8df64e44df8cdb4c3a022f03fccb39451631ca35b30d4c7c35271ff0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:43Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:43 crc kubenswrapper[4890]: I1125 15:02:43.581637 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg454" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81411ec7-f865-48e2-b464-ab6213ef5e52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg454\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:43Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:43 crc kubenswrapper[4890]: I1125 15:02:43.622124 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4txd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfcq7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4txd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:43Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:43 crc kubenswrapper[4890]: I1125 15:02:43.661890 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e4f849d-f239-4727-a73e-18327856929a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://490f1595bc77030d2ee789197fa3748324e1b7a0c9d9180ba08383bbac43adc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0902d6f8e56aecc5a2a13a7148f94a8a00ad139027b3c036aba0215ba0a67f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xx2tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:43Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:43 crc kubenswrapper[4890]: I1125 15:02:43.700382 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583964bf462fa0005d471f82f6827552a60659ab97984eb1faf074c7b82c2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:43Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:43 crc kubenswrapper[4890]: I1125 15:02:43.739297 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:43Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:43 crc kubenswrapper[4890]: I1125 15:02:43.778733 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7g5s8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a7dc232-835b-491c-afc9-9eb14f257af3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9a7ac317fbb9c589573e80babe484c21bef4da3d9ab79dfb7ebf5b8300c472c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4cs2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7g5s8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:43Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:43 crc kubenswrapper[4890]: I1125 15:02:43.815316 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-8tdrh"] Nov 25 15:02:43 crc kubenswrapper[4890]: I1125 15:02:43.815957 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-8tdrh" Nov 25 15:02:43 crc kubenswrapper[4890]: I1125 15:02:43.823817 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a404958-dffa-4a6a-99fc-3e4950614bd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zdk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:43Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:43 crc kubenswrapper[4890]: I1125 15:02:43.830689 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 25 15:02:43 crc kubenswrapper[4890]: I1125 15:02:43.851123 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 25 15:02:43 crc kubenswrapper[4890]: I1125 15:02:43.870476 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 25 15:02:43 crc kubenswrapper[4890]: I1125 15:02:43.890965 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 25 15:02:43 crc kubenswrapper[4890]: I1125 15:02:43.912526 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbz6r\" (UniqueName: \"kubernetes.io/projected/d1097de4-74e4-4178-b9b4-14a22539da9a-kube-api-access-qbz6r\") pod \"node-ca-8tdrh\" (UID: \"d1097de4-74e4-4178-b9b4-14a22539da9a\") " pod="openshift-image-registry/node-ca-8tdrh" Nov 25 15:02:43 crc kubenswrapper[4890]: I1125 15:02:43.912582 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d1097de4-74e4-4178-b9b4-14a22539da9a-host\") pod \"node-ca-8tdrh\" (UID: \"d1097de4-74e4-4178-b9b4-14a22539da9a\") " pod="openshift-image-registry/node-ca-8tdrh" Nov 25 15:02:43 crc kubenswrapper[4890]: I1125 15:02:43.912661 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/d1097de4-74e4-4178-b9b4-14a22539da9a-serviceca\") pod \"node-ca-8tdrh\" (UID: \"d1097de4-74e4-4178-b9b4-14a22539da9a\") " pod="openshift-image-registry/node-ca-8tdrh" Nov 25 15:02:43 crc kubenswrapper[4890]: I1125 15:02:43.939069 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd5e648-6322-404d-bf1b-3b59a8092291\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76d277b4001db9d63b0af88b98afcc9019321101429c9be47b3022d3575d4e06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e3a186b14ea199d1aeba5751290c266ec918cb055120cfc5c8a2ae03b1469e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a860a998c40aee3e94e4cfd748f54391b3096fd3ed75c84c7ea2a5a7cefa33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://081a6d623cd043c0919d04e2ca93314b2b85fe33048634aa2637297d54b52cee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f506ec649ce24f4df6fd33e4aeafd13e42167dd71323524a5a3164517659fcc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:43Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:43 crc kubenswrapper[4890]: I1125 15:02:43.978679 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d5a9ccb-13ba-4dd5-9904-42aac5eca374\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d19bc48acb392051f0f6c5a6820135bf29ebe6d2afd098697553298933cf280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93325a6407389c4a7db0ba3c1a9c84a6980db3c3277cdf3860583a564b562bbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20505e91fc60158eeb3a8cfa0aafc5b4e4436d5def5e735a3a58b0f9f3621058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3281314d85a20945d56b75ab7b7aef1e382a73e74dff03773d118d03eb0ee16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:43Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:44 crc kubenswrapper[4890]: I1125 15:02:44.013944 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d1097de4-74e4-4178-b9b4-14a22539da9a-host\") pod \"node-ca-8tdrh\" (UID: \"d1097de4-74e4-4178-b9b4-14a22539da9a\") " pod="openshift-image-registry/node-ca-8tdrh" Nov 25 15:02:44 crc kubenswrapper[4890]: I1125 15:02:44.013997 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/d1097de4-74e4-4178-b9b4-14a22539da9a-serviceca\") pod \"node-ca-8tdrh\" (UID: \"d1097de4-74e4-4178-b9b4-14a22539da9a\") " pod="openshift-image-registry/node-ca-8tdrh" Nov 25 15:02:44 crc kubenswrapper[4890]: I1125 15:02:44.014052 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbz6r\" (UniqueName: \"kubernetes.io/projected/d1097de4-74e4-4178-b9b4-14a22539da9a-kube-api-access-qbz6r\") pod \"node-ca-8tdrh\" (UID: \"d1097de4-74e4-4178-b9b4-14a22539da9a\") " pod="openshift-image-registry/node-ca-8tdrh" Nov 25 15:02:44 crc kubenswrapper[4890]: I1125 15:02:44.014067 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d1097de4-74e4-4178-b9b4-14a22539da9a-host\") pod \"node-ca-8tdrh\" (UID: \"d1097de4-74e4-4178-b9b4-14a22539da9a\") " pod="openshift-image-registry/node-ca-8tdrh" Nov 25 15:02:44 crc kubenswrapper[4890]: I1125 15:02:44.015597 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/d1097de4-74e4-4178-b9b4-14a22539da9a-serviceca\") pod \"node-ca-8tdrh\" (UID: \"d1097de4-74e4-4178-b9b4-14a22539da9a\") " pod="openshift-image-registry/node-ca-8tdrh" Nov 25 15:02:44 crc kubenswrapper[4890]: I1125 15:02:44.020174 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:44Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:44 crc kubenswrapper[4890]: I1125 15:02:44.051364 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbz6r\" (UniqueName: \"kubernetes.io/projected/d1097de4-74e4-4178-b9b4-14a22539da9a-kube-api-access-qbz6r\") pod \"node-ca-8tdrh\" (UID: \"d1097de4-74e4-4178-b9b4-14a22539da9a\") " pod="openshift-image-registry/node-ca-8tdrh" Nov 25 15:02:44 crc kubenswrapper[4890]: I1125 15:02:44.079539 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:44Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:44 crc kubenswrapper[4890]: I1125 15:02:44.118952 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:44Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:44 crc kubenswrapper[4890]: I1125 15:02:44.130573 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-8tdrh" Nov 25 15:02:44 crc kubenswrapper[4890]: W1125 15:02:44.144025 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1097de4_74e4_4178_b9b4_14a22539da9a.slice/crio-195733ee0e3bd52b4d17b65451b508e63d46fdc6f8545bcb055ef2edf1eca7df WatchSource:0}: Error finding container 195733ee0e3bd52b4d17b65451b508e63d46fdc6f8545bcb055ef2edf1eca7df: Status 404 returned error can't find the container with id 195733ee0e3bd52b4d17b65451b508e63d46fdc6f8545bcb055ef2edf1eca7df Nov 25 15:02:44 crc kubenswrapper[4890]: I1125 15:02:44.156836 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3358a356ae5b422d6bba72bc516374e14ccdace80726aa71b10e25102e0f1100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba3acdae961f49f797f66bbebe350e34a4df5882bcf2a4fe8d37f2ca95865c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:44Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:44 crc kubenswrapper[4890]: I1125 15:02:44.172589 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 15:02:44 crc kubenswrapper[4890]: I1125 15:02:44.172691 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:02:44 crc kubenswrapper[4890]: E1125 15:02:44.172752 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 15:02:44 crc kubenswrapper[4890]: E1125 15:02:44.172844 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 15:02:44 crc kubenswrapper[4890]: I1125 15:02:44.172710 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 15:02:44 crc kubenswrapper[4890]: E1125 15:02:44.172929 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 15:02:44 crc kubenswrapper[4890]: I1125 15:02:44.212060 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8tdrh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1097de4-74e4-4178-b9b4-14a22539da9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8tdrh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:44Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:44 crc kubenswrapper[4890]: I1125 15:02:44.261385 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86e031a7-b4c8-4030-a078-cdca7ccc9dde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4c4921f0973b03f6dade29c2ef32684a481cd1c3e1e23b64ed636943c9bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f9285acc7ac3cbb503a1e546728748c42971c8c9019cc8ca837076f3e55558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa240816a6e9d2821006ce240eccf27d8f57a7f0c5394314fa7177488da4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://659004bd47948def6b33498d8d1474f28c2e98986754f61af012267f3984269e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6d2dcd8df64e44df8cdb4c3a022f03fccb39451631ca35b30d4c7c35271ff0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:44Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:44 crc kubenswrapper[4890]: I1125 15:02:44.282821 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg454" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81411ec7-f865-48e2-b464-ab6213ef5e52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg454\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:44Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:44 crc kubenswrapper[4890]: I1125 15:02:44.320457 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4txd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4322dbd78237bf443a985ce3c79bca4cba2103862d81177f5474a76f4905eed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfcq7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4txd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:44Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:44 crc kubenswrapper[4890]: I1125 15:02:44.359700 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-8tdrh" event={"ID":"d1097de4-74e4-4178-b9b4-14a22539da9a","Type":"ContainerStarted","Data":"195733ee0e3bd52b4d17b65451b508e63d46fdc6f8545bcb055ef2edf1eca7df"} Nov 25 15:02:44 crc kubenswrapper[4890]: I1125 15:02:44.363264 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" event={"ID":"9a404958-dffa-4a6a-99fc-3e4950614bd4","Type":"ContainerStarted","Data":"cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9"} Nov 25 15:02:44 crc kubenswrapper[4890]: I1125 15:02:44.363518 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" event={"ID":"9a404958-dffa-4a6a-99fc-3e4950614bd4","Type":"ContainerStarted","Data":"f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48"} Nov 25 15:02:44 crc kubenswrapper[4890]: I1125 15:02:44.363603 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" event={"ID":"9a404958-dffa-4a6a-99fc-3e4950614bd4","Type":"ContainerStarted","Data":"bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a"} Nov 25 15:02:44 crc kubenswrapper[4890]: I1125 15:02:44.363679 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" event={"ID":"9a404958-dffa-4a6a-99fc-3e4950614bd4","Type":"ContainerStarted","Data":"9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100"} Nov 25 15:02:44 crc kubenswrapper[4890]: I1125 15:02:44.364949 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583964bf462fa0005d471f82f6827552a60659ab97984eb1faf074c7b82c2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:44Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:44 crc kubenswrapper[4890]: I1125 15:02:44.366047 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zg454" event={"ID":"81411ec7-f865-48e2-b464-ab6213ef5e52","Type":"ContainerStarted","Data":"2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b"} Nov 25 15:02:44 crc kubenswrapper[4890]: I1125 15:02:44.402542 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:44Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:44 crc kubenswrapper[4890]: I1125 15:02:44.436871 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7g5s8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a7dc232-835b-491c-afc9-9eb14f257af3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9a7ac317fbb9c589573e80babe484c21bef4da3d9ab79dfb7ebf5b8300c472c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4cs2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7g5s8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:44Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:44 crc kubenswrapper[4890]: I1125 15:02:44.491192 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a404958-dffa-4a6a-99fc-3e4950614bd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zdk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:44Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:44 crc kubenswrapper[4890]: I1125 15:02:44.518233 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e4f849d-f239-4727-a73e-18327856929a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://490f1595bc77030d2ee789197fa3748324e1b7a0c9d9180ba08383bbac43adc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0902d6f8e56aecc5a2a13a7148f94a8a00ad139027b3c036aba0215ba0a67f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xx2tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:44Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:44 crc kubenswrapper[4890]: I1125 15:02:44.575724 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86e031a7-b4c8-4030-a078-cdca7ccc9dde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4c4921f0973b03f6dade29c2ef32684a481cd1c3e1e23b64ed636943c9bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f9285acc7ac3cbb503a1e546728748c42971c8c9019cc8ca837076f3e55558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa240816a6e9d2821006ce240eccf27d8f57a7f0c5394314fa7177488da4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://659004bd47948def6b33498d8d1474f28c2e98986754f61af012267f3984269e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6d2dcd8df64e44df8cdb4c3a022f03fccb39451631ca35b30d4c7c35271ff0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:44Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:44 crc kubenswrapper[4890]: I1125 15:02:44.599735 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg454" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81411ec7-f865-48e2-b464-ab6213ef5e52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg454\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:44Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:44 crc kubenswrapper[4890]: I1125 15:02:44.642148 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4txd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4322dbd78237bf443a985ce3c79bca4cba2103862d81177f5474a76f4905eed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfcq7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4txd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:44Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:44 crc kubenswrapper[4890]: I1125 15:02:44.679796 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583964bf462fa0005d471f82f6827552a60659ab97984eb1faf074c7b82c2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:44Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:44 crc kubenswrapper[4890]: I1125 15:02:44.720275 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:44Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:44 crc kubenswrapper[4890]: I1125 15:02:44.759861 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7g5s8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a7dc232-835b-491c-afc9-9eb14f257af3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9a7ac317fbb9c589573e80babe484c21bef4da3d9ab79dfb7ebf5b8300c472c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4cs2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7g5s8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:44Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:44 crc kubenswrapper[4890]: I1125 15:02:44.804486 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a404958-dffa-4a6a-99fc-3e4950614bd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zdk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:44Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:44 crc kubenswrapper[4890]: I1125 15:02:44.837412 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e4f849d-f239-4727-a73e-18327856929a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://490f1595bc77030d2ee789197fa3748324e1b7a0c9d9180ba08383bbac43adc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0902d6f8e56aecc5a2a13a7148f94a8a00ad139027b3c036aba0215ba0a67f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xx2tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:44Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:44 crc kubenswrapper[4890]: I1125 15:02:44.882992 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd5e648-6322-404d-bf1b-3b59a8092291\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76d277b4001db9d63b0af88b98afcc9019321101429c9be47b3022d3575d4e06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e3a186b14ea199d1aeba5751290c266ec918cb055120cfc5c8a2ae03b1469e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a860a998c40aee3e94e4cfd748f54391b3096fd3ed75c84c7ea2a5a7cefa33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://081a6d623cd043c0919d04e2ca93314b2b85fe33048634aa2637297d54b52cee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f506ec649ce24f4df6fd33e4aeafd13e42167dd71323524a5a3164517659fcc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:44Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:44 crc kubenswrapper[4890]: I1125 15:02:44.919132 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d5a9ccb-13ba-4dd5-9904-42aac5eca374\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d19bc48acb392051f0f6c5a6820135bf29ebe6d2afd098697553298933cf280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93325a6407389c4a7db0ba3c1a9c84a6980db3c3277cdf3860583a564b562bbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20505e91fc60158eeb3a8cfa0aafc5b4e4436d5def5e735a3a58b0f9f3621058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3281314d85a20945d56b75ab7b7aef1e382a73e74dff03773d118d03eb0ee16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:44Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:44 crc kubenswrapper[4890]: I1125 15:02:44.960973 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:44Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:44 crc kubenswrapper[4890]: I1125 15:02:44.998603 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:44Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:45 crc kubenswrapper[4890]: I1125 15:02:45.041859 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:45Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:45 crc kubenswrapper[4890]: I1125 15:02:45.081831 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3358a356ae5b422d6bba72bc516374e14ccdace80726aa71b10e25102e0f1100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba3acdae961f49f797f66bbebe350e34a4df5882bcf2a4fe8d37f2ca95865c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:45Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:45 crc kubenswrapper[4890]: I1125 15:02:45.117374 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8tdrh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1097de4-74e4-4178-b9b4-14a22539da9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8tdrh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:45Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:45 crc kubenswrapper[4890]: I1125 15:02:45.125627 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:02:45 crc kubenswrapper[4890]: I1125 15:02:45.125719 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 15:02:45 crc kubenswrapper[4890]: I1125 15:02:45.125742 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:02:45 crc kubenswrapper[4890]: I1125 15:02:45.125768 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:02:45 crc kubenswrapper[4890]: I1125 15:02:45.125791 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 15:02:45 crc kubenswrapper[4890]: E1125 15:02:45.125927 4890 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 15:02:45 crc kubenswrapper[4890]: E1125 15:02:45.125946 4890 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 15:02:45 crc kubenswrapper[4890]: E1125 15:02:45.125958 4890 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 15:02:45 crc kubenswrapper[4890]: E1125 15:02:45.125926 4890 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 15:02:45 crc kubenswrapper[4890]: E1125 15:02:45.126028 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 15:02:49.126005364 +0000 UTC m=+27.568467974 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 15:02:45 crc kubenswrapper[4890]: E1125 15:02:45.126057 4890 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 15:02:45 crc kubenswrapper[4890]: E1125 15:02:45.126071 4890 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 15:02:45 crc kubenswrapper[4890]: E1125 15:02:45.126079 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 15:02:49.126054856 +0000 UTC m=+27.568517616 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 15:02:45 crc kubenswrapper[4890]: E1125 15:02:45.126095 4890 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 15:02:45 crc kubenswrapper[4890]: E1125 15:02:45.126244 4890 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 15:02:45 crc kubenswrapper[4890]: E1125 15:02:45.126213 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 15:02:49.126191099 +0000 UTC m=+27.568653879 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 15:02:45 crc kubenswrapper[4890]: E1125 15:02:45.126316 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:02:49.126299751 +0000 UTC m=+27.568762551 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:02:45 crc kubenswrapper[4890]: E1125 15:02:45.126339 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 15:02:49.126331782 +0000 UTC m=+27.568794612 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 15:02:45 crc kubenswrapper[4890]: I1125 15:02:45.381906 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" event={"ID":"9a404958-dffa-4a6a-99fc-3e4950614bd4","Type":"ContainerStarted","Data":"7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a"} Nov 25 15:02:45 crc kubenswrapper[4890]: I1125 15:02:45.382341 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" event={"ID":"9a404958-dffa-4a6a-99fc-3e4950614bd4","Type":"ContainerStarted","Data":"89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91"} Nov 25 15:02:45 crc kubenswrapper[4890]: I1125 15:02:45.384817 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"4b33b805b76abbcace899e527d5ee3eaeb9e480fb03dc8eacd3f6dd90d2dd6fb"} Nov 25 15:02:45 crc kubenswrapper[4890]: I1125 15:02:45.387527 4890 generic.go:334] "Generic (PLEG): container finished" podID="81411ec7-f865-48e2-b464-ab6213ef5e52" containerID="2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b" exitCode=0 Nov 25 15:02:45 crc kubenswrapper[4890]: I1125 15:02:45.387651 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zg454" event={"ID":"81411ec7-f865-48e2-b464-ab6213ef5e52","Type":"ContainerDied","Data":"2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b"} Nov 25 15:02:45 crc kubenswrapper[4890]: I1125 15:02:45.389198 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-8tdrh" event={"ID":"d1097de4-74e4-4178-b9b4-14a22539da9a","Type":"ContainerStarted","Data":"cfa0291da1598e930aeb4a76f9f4e8b0728001a770acd8f10215465539f8d8b2"} Nov 25 15:02:45 crc kubenswrapper[4890]: I1125 15:02:45.404295 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b33b805b76abbcace899e527d5ee3eaeb9e480fb03dc8eacd3f6dd90d2dd6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:45Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:45 crc kubenswrapper[4890]: I1125 15:02:45.441140 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd5e648-6322-404d-bf1b-3b59a8092291\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76d277b4001db9d63b0af88b98afcc9019321101429c9be47b3022d3575d4e06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e3a186b14ea199d1aeba5751290c266ec918cb055120cfc5c8a2ae03b1469e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a860a998c40aee3e94e4cfd748f54391b3096fd3ed75c84c7ea2a5a7cefa33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://081a6d623cd043c0919d04e2ca93314b2b85fe33048634aa2637297d54b52cee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f506ec649ce24f4df6fd33e4aeafd13e42167dd71323524a5a3164517659fcc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:45Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:45 crc kubenswrapper[4890]: I1125 15:02:45.454216 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d5a9ccb-13ba-4dd5-9904-42aac5eca374\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d19bc48acb392051f0f6c5a6820135bf29ebe6d2afd098697553298933cf280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93325a6407389c4a7db0ba3c1a9c84a6980db3c3277cdf3860583a564b562bbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20505e91fc60158eeb3a8cfa0aafc5b4e4436d5def5e735a3a58b0f9f3621058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3281314d85a20945d56b75ab7b7aef1e382a73e74dff03773d118d03eb0ee16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:45Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:45 crc kubenswrapper[4890]: I1125 15:02:45.468034 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:45Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:45 crc kubenswrapper[4890]: I1125 15:02:45.480643 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3358a356ae5b422d6bba72bc516374e14ccdace80726aa71b10e25102e0f1100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba3acdae961f49f797f66bbebe350e34a4df5882bcf2a4fe8d37f2ca95865c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:45Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:45 crc kubenswrapper[4890]: I1125 15:02:45.490627 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8tdrh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1097de4-74e4-4178-b9b4-14a22539da9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8tdrh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:45Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:45 crc kubenswrapper[4890]: I1125 15:02:45.505270 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:45Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:45 crc kubenswrapper[4890]: I1125 15:02:45.522883 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4txd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4322dbd78237bf443a985ce3c79bca4cba2103862d81177f5474a76f4905eed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfcq7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4txd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:45Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:45 crc kubenswrapper[4890]: I1125 15:02:45.544094 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86e031a7-b4c8-4030-a078-cdca7ccc9dde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4c4921f0973b03f6dade29c2ef32684a481cd1c3e1e23b64ed636943c9bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f9285acc7ac3cbb503a1e546728748c42971c8c9019cc8ca837076f3e55558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa240816a6e9d2821006ce240eccf27d8f57a7f0c5394314fa7177488da4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://659004bd47948def6b33498d8d1474f28c2e98986754f61af012267f3984269e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6d2dcd8df64e44df8cdb4c3a022f03fccb39451631ca35b30d4c7c35271ff0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:45Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:45 crc kubenswrapper[4890]: I1125 15:02:45.565521 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg454" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81411ec7-f865-48e2-b464-ab6213ef5e52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg454\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:45Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:45 crc kubenswrapper[4890]: I1125 15:02:45.580302 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7g5s8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a7dc232-835b-491c-afc9-9eb14f257af3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9a7ac317fbb9c589573e80babe484c21bef4da3d9ab79dfb7ebf5b8300c472c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4cs2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7g5s8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:45Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:45 crc kubenswrapper[4890]: I1125 15:02:45.609674 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a404958-dffa-4a6a-99fc-3e4950614bd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zdk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:45Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:45 crc kubenswrapper[4890]: I1125 15:02:45.640713 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e4f849d-f239-4727-a73e-18327856929a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://490f1595bc77030d2ee789197fa3748324e1b7a0c9d9180ba08383bbac43adc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0902d6f8e56aecc5a2a13a7148f94a8a00ad139027b3c036aba0215ba0a67f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xx2tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:45Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:45 crc kubenswrapper[4890]: I1125 15:02:45.680470 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583964bf462fa0005d471f82f6827552a60659ab97984eb1faf074c7b82c2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:45Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:45 crc kubenswrapper[4890]: I1125 15:02:45.719806 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:45Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:45 crc kubenswrapper[4890]: I1125 15:02:45.759266 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd5e648-6322-404d-bf1b-3b59a8092291\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76d277b4001db9d63b0af88b98afcc9019321101429c9be47b3022d3575d4e06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e3a186b14ea199d1aeba5751290c266ec918cb055120cfc5c8a2ae03b1469e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a860a998c40aee3e94e4cfd748f54391b3096fd3ed75c84c7ea2a5a7cefa33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://081a6d623cd043c0919d04e2ca93314b2b85fe33048634aa2637297d54b52cee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f506ec649ce24f4df6fd33e4aeafd13e42167dd71323524a5a3164517659fcc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:45Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:45 crc kubenswrapper[4890]: I1125 15:02:45.806695 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d5a9ccb-13ba-4dd5-9904-42aac5eca374\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d19bc48acb392051f0f6c5a6820135bf29ebe6d2afd098697553298933cf280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93325a6407389c4a7db0ba3c1a9c84a6980db3c3277cdf3860583a564b562bbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20505e91fc60158eeb3a8cfa0aafc5b4e4436d5def5e735a3a58b0f9f3621058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3281314d85a20945d56b75ab7b7aef1e382a73e74dff03773d118d03eb0ee16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:45Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:45 crc kubenswrapper[4890]: I1125 15:02:45.840184 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b33b805b76abbcace899e527d5ee3eaeb9e480fb03dc8eacd3f6dd90d2dd6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:45Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:45 crc kubenswrapper[4890]: I1125 15:02:45.879779 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:45Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:45 crc kubenswrapper[4890]: I1125 15:02:45.923207 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:45Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:45 crc kubenswrapper[4890]: I1125 15:02:45.959401 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3358a356ae5b422d6bba72bc516374e14ccdace80726aa71b10e25102e0f1100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba3acdae961f49f797f66bbebe350e34a4df5882bcf2a4fe8d37f2ca95865c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:45Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:45 crc kubenswrapper[4890]: I1125 15:02:45.996782 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8tdrh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1097de4-74e4-4178-b9b4-14a22539da9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfa0291da1598e930aeb4a76f9f4e8b0728001a770acd8f10215465539f8d8b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8tdrh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:45Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:46 crc kubenswrapper[4890]: I1125 15:02:46.045925 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86e031a7-b4c8-4030-a078-cdca7ccc9dde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4c4921f0973b03f6dade29c2ef32684a481cd1c3e1e23b64ed636943c9bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f9285acc7ac3cbb503a1e546728748c42971c8c9019cc8ca837076f3e55558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa240816a6e9d2821006ce240eccf27d8f57a7f0c5394314fa7177488da4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://659004bd47948def6b33498d8d1474f28c2e98986754f61af012267f3984269e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6d2dcd8df64e44df8cdb4c3a022f03fccb39451631ca35b30d4c7c35271ff0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:46Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:46 crc kubenswrapper[4890]: I1125 15:02:46.080460 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg454" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81411ec7-f865-48e2-b464-ab6213ef5e52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg454\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:46Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:46 crc kubenswrapper[4890]: I1125 15:02:46.120222 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4txd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4322dbd78237bf443a985ce3c79bca4cba2103862d81177f5474a76f4905eed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfcq7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4txd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:46Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:46 crc kubenswrapper[4890]: I1125 15:02:46.160694 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583964bf462fa0005d471f82f6827552a60659ab97984eb1faf074c7b82c2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:46Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:46 crc kubenswrapper[4890]: I1125 15:02:46.172520 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 15:02:46 crc kubenswrapper[4890]: I1125 15:02:46.172552 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 15:02:46 crc kubenswrapper[4890]: I1125 15:02:46.172604 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:02:46 crc kubenswrapper[4890]: E1125 15:02:46.172632 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 15:02:46 crc kubenswrapper[4890]: E1125 15:02:46.172733 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 15:02:46 crc kubenswrapper[4890]: E1125 15:02:46.172841 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 15:02:46 crc kubenswrapper[4890]: I1125 15:02:46.198121 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:46Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:46 crc kubenswrapper[4890]: I1125 15:02:46.237398 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7g5s8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a7dc232-835b-491c-afc9-9eb14f257af3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9a7ac317fbb9c589573e80babe484c21bef4da3d9ab79dfb7ebf5b8300c472c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4cs2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7g5s8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:46Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:46 crc kubenswrapper[4890]: I1125 15:02:46.282666 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a404958-dffa-4a6a-99fc-3e4950614bd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zdk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:46Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:46 crc kubenswrapper[4890]: I1125 15:02:46.318683 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e4f849d-f239-4727-a73e-18327856929a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://490f1595bc77030d2ee789197fa3748324e1b7a0c9d9180ba08383bbac43adc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0902d6f8e56aecc5a2a13a7148f94a8a00ad139027b3c036aba0215ba0a67f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xx2tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:46Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:46 crc kubenswrapper[4890]: I1125 15:02:46.393972 4890 generic.go:334] "Generic (PLEG): container finished" podID="81411ec7-f865-48e2-b464-ab6213ef5e52" containerID="4a366b314408eb8453ab7f970a243c511a2716b5472d92b03c81061135e4e59c" exitCode=0 Nov 25 15:02:46 crc kubenswrapper[4890]: I1125 15:02:46.394034 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zg454" event={"ID":"81411ec7-f865-48e2-b464-ab6213ef5e52","Type":"ContainerDied","Data":"4a366b314408eb8453ab7f970a243c511a2716b5472d92b03c81061135e4e59c"} Nov 25 15:02:46 crc kubenswrapper[4890]: I1125 15:02:46.425124 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86e031a7-b4c8-4030-a078-cdca7ccc9dde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4c4921f0973b03f6dade29c2ef32684a481cd1c3e1e23b64ed636943c9bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f9285acc7ac3cbb503a1e546728748c42971c8c9019cc8ca837076f3e55558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa240816a6e9d2821006ce240eccf27d8f57a7f0c5394314fa7177488da4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://659004bd47948def6b33498d8d1474f28c2e98986754f61af012267f3984269e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6d2dcd8df64e44df8cdb4c3a022f03fccb39451631ca35b30d4c7c35271ff0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:46Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:46 crc kubenswrapper[4890]: I1125 15:02:46.442283 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg454" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81411ec7-f865-48e2-b464-ab6213ef5e52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a366b314408eb8453ab7f970a243c511a2716b5472d92b03c81061135e4e59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a366b314408eb8453ab7f970a243c511a2716b5472d92b03c81061135e4e59c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg454\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:46Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:46 crc kubenswrapper[4890]: I1125 15:02:46.456626 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4txd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4322dbd78237bf443a985ce3c79bca4cba2103862d81177f5474a76f4905eed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfcq7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4txd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:46Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:46 crc kubenswrapper[4890]: I1125 15:02:46.479997 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583964bf462fa0005d471f82f6827552a60659ab97984eb1faf074c7b82c2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:46Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:46 crc kubenswrapper[4890]: I1125 15:02:46.519991 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:46Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:46 crc kubenswrapper[4890]: I1125 15:02:46.558426 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7g5s8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a7dc232-835b-491c-afc9-9eb14f257af3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9a7ac317fbb9c589573e80babe484c21bef4da3d9ab79dfb7ebf5b8300c472c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4cs2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7g5s8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:46Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:46 crc kubenswrapper[4890]: I1125 15:02:46.605045 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a404958-dffa-4a6a-99fc-3e4950614bd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zdk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:46Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:46 crc kubenswrapper[4890]: I1125 15:02:46.639146 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e4f849d-f239-4727-a73e-18327856929a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://490f1595bc77030d2ee789197fa3748324e1b7a0c9d9180ba08383bbac43adc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0902d6f8e56aecc5a2a13a7148f94a8a00ad139027b3c036aba0215ba0a67f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xx2tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:46Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:46 crc kubenswrapper[4890]: I1125 15:02:46.679293 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd5e648-6322-404d-bf1b-3b59a8092291\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76d277b4001db9d63b0af88b98afcc9019321101429c9be47b3022d3575d4e06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e3a186b14ea199d1aeba5751290c266ec918cb055120cfc5c8a2ae03b1469e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a860a998c40aee3e94e4cfd748f54391b3096fd3ed75c84c7ea2a5a7cefa33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://081a6d623cd043c0919d04e2ca93314b2b85fe33048634aa2637297d54b52cee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f506ec649ce24f4df6fd33e4aeafd13e42167dd71323524a5a3164517659fcc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:46Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:46 crc kubenswrapper[4890]: I1125 15:02:46.719748 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d5a9ccb-13ba-4dd5-9904-42aac5eca374\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d19bc48acb392051f0f6c5a6820135bf29ebe6d2afd098697553298933cf280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93325a6407389c4a7db0ba3c1a9c84a6980db3c3277cdf3860583a564b562bbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20505e91fc60158eeb3a8cfa0aafc5b4e4436d5def5e735a3a58b0f9f3621058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3281314d85a20945d56b75ab7b7aef1e382a73e74dff03773d118d03eb0ee16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:46Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:46 crc kubenswrapper[4890]: I1125 15:02:46.758117 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b33b805b76abbcace899e527d5ee3eaeb9e480fb03dc8eacd3f6dd90d2dd6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:46Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:46 crc kubenswrapper[4890]: I1125 15:02:46.797175 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:46Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:46 crc kubenswrapper[4890]: I1125 15:02:46.841512 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:46Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:46 crc kubenswrapper[4890]: I1125 15:02:46.879410 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3358a356ae5b422d6bba72bc516374e14ccdace80726aa71b10e25102e0f1100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba3acdae961f49f797f66bbebe350e34a4df5882bcf2a4fe8d37f2ca95865c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:46Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:46 crc kubenswrapper[4890]: I1125 15:02:46.917661 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8tdrh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1097de4-74e4-4178-b9b4-14a22539da9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfa0291da1598e930aeb4a76f9f4e8b0728001a770acd8f10215465539f8d8b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8tdrh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:46Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.408773 4890 generic.go:334] "Generic (PLEG): container finished" podID="81411ec7-f865-48e2-b464-ab6213ef5e52" containerID="6f841cf4a886c5e17d9919787305ce297d7910b2692be53169324a56bd545be5" exitCode=0 Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.408825 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zg454" event={"ID":"81411ec7-f865-48e2-b464-ab6213ef5e52","Type":"ContainerDied","Data":"6f841cf4a886c5e17d9919787305ce297d7910b2692be53169324a56bd545be5"} Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.422449 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd5e648-6322-404d-bf1b-3b59a8092291\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76d277b4001db9d63b0af88b98afcc9019321101429c9be47b3022d3575d4e06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e3a186b14ea199d1aeba5751290c266ec918cb055120cfc5c8a2ae03b1469e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a860a998c40aee3e94e4cfd748f54391b3096fd3ed75c84c7ea2a5a7cefa33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://081a6d623cd043c0919d04e2ca93314b2b85fe33048634aa2637297d54b52cee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f506ec649ce24f4df6fd33e4aeafd13e42167dd71323524a5a3164517659fcc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:47Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.435234 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d5a9ccb-13ba-4dd5-9904-42aac5eca374\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d19bc48acb392051f0f6c5a6820135bf29ebe6d2afd098697553298933cf280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93325a6407389c4a7db0ba3c1a9c84a6980db3c3277cdf3860583a564b562bbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20505e91fc60158eeb3a8cfa0aafc5b4e4436d5def5e735a3a58b0f9f3621058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3281314d85a20945d56b75ab7b7aef1e382a73e74dff03773d118d03eb0ee16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:47Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.452324 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b33b805b76abbcace899e527d5ee3eaeb9e480fb03dc8eacd3f6dd90d2dd6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:47Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.464929 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8tdrh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1097de4-74e4-4178-b9b4-14a22539da9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfa0291da1598e930aeb4a76f9f4e8b0728001a770acd8f10215465539f8d8b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8tdrh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:47Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.477986 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:47Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.491059 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:47Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.505031 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3358a356ae5b422d6bba72bc516374e14ccdace80726aa71b10e25102e0f1100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba3acdae961f49f797f66bbebe350e34a4df5882bcf2a4fe8d37f2ca95865c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:47Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.523452 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86e031a7-b4c8-4030-a078-cdca7ccc9dde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4c4921f0973b03f6dade29c2ef32684a481cd1c3e1e23b64ed636943c9bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f9285acc7ac3cbb503a1e546728748c42971c8c9019cc8ca837076f3e55558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa240816a6e9d2821006ce240eccf27d8f57a7f0c5394314fa7177488da4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://659004bd47948def6b33498d8d1474f28c2e98986754f61af012267f3984269e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6d2dcd8df64e44df8cdb4c3a022f03fccb39451631ca35b30d4c7c35271ff0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:47Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.541743 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg454" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81411ec7-f865-48e2-b464-ab6213ef5e52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a366b314408eb8453ab7f970a243c511a2716b5472d92b03c81061135e4e59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a366b314408eb8453ab7f970a243c511a2716b5472d92b03c81061135e4e59c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f841cf4a886c5e17d9919787305ce297d7910b2692be53169324a56bd545be5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f841cf4a886c5e17d9919787305ce297d7910b2692be53169324a56bd545be5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg454\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:47Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.585577 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4txd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4322dbd78237bf443a985ce3c79bca4cba2103862d81177f5474a76f4905eed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfcq7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4txd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:47Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.616121 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e4f849d-f239-4727-a73e-18327856929a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://490f1595bc77030d2ee789197fa3748324e1b7a0c9d9180ba08383bbac43adc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0902d6f8e56aecc5a2a13a7148f94a8a00ad139027b3c036aba0215ba0a67f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xx2tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:47Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.639018 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583964bf462fa0005d471f82f6827552a60659ab97984eb1faf074c7b82c2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:47Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.657712 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:47Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.669889 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7g5s8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a7dc232-835b-491c-afc9-9eb14f257af3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9a7ac317fbb9c589573e80babe484c21bef4da3d9ab79dfb7ebf5b8300c472c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4cs2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7g5s8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:47Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.689728 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a404958-dffa-4a6a-99fc-3e4950614bd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zdk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:47Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.740629 4890 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.742633 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.742695 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.742703 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.742810 4890 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.749827 4890 kubelet_node_status.go:115] "Node was previously registered" node="crc" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.750339 4890 kubelet_node_status.go:79] "Successfully registered node" node="crc" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.751850 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.751917 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.751932 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.751954 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.751969 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:47Z","lastTransitionTime":"2025-11-25T15:02:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:47 crc kubenswrapper[4890]: E1125 15:02:47.766870 4890 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:02:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:02:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:02:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:02:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7c7b4592-d206-4ee7-bf93-ada4b8bc755a\\\",\\\"systemUUID\\\":\\\"5350de30-707e-4bde-b990-1fa65ae3a01e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:47Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.771805 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.771837 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.771846 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.771859 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.771871 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:47Z","lastTransitionTime":"2025-11-25T15:02:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:47 crc kubenswrapper[4890]: E1125 15:02:47.786831 4890 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:02:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:02:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:02:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:02:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7c7b4592-d206-4ee7-bf93-ada4b8bc755a\\\",\\\"systemUUID\\\":\\\"5350de30-707e-4bde-b990-1fa65ae3a01e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:47Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.790102 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.790143 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.790171 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.790187 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.790199 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:47Z","lastTransitionTime":"2025-11-25T15:02:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:47 crc kubenswrapper[4890]: E1125 15:02:47.803664 4890 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:02:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:02:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:02:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:02:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7c7b4592-d206-4ee7-bf93-ada4b8bc755a\\\",\\\"systemUUID\\\":\\\"5350de30-707e-4bde-b990-1fa65ae3a01e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:47Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.807592 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.807633 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.807644 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.807661 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.807678 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:47Z","lastTransitionTime":"2025-11-25T15:02:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:47 crc kubenswrapper[4890]: E1125 15:02:47.819146 4890 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:02:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:02:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:02:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:02:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7c7b4592-d206-4ee7-bf93-ada4b8bc755a\\\",\\\"systemUUID\\\":\\\"5350de30-707e-4bde-b990-1fa65ae3a01e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:47Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.824042 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.824081 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.824095 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.824113 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.824125 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:47Z","lastTransitionTime":"2025-11-25T15:02:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:47 crc kubenswrapper[4890]: E1125 15:02:47.841358 4890 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:02:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:02:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:02:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:02:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7c7b4592-d206-4ee7-bf93-ada4b8bc755a\\\",\\\"systemUUID\\\":\\\"5350de30-707e-4bde-b990-1fa65ae3a01e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:47Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:47 crc kubenswrapper[4890]: E1125 15:02:47.841557 4890 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.843732 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.843799 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.843812 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.843836 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.843850 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:47Z","lastTransitionTime":"2025-11-25T15:02:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.946360 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.946406 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.946418 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.946434 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:47 crc kubenswrapper[4890]: I1125 15:02:47.946445 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:47Z","lastTransitionTime":"2025-11-25T15:02:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.049886 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.049934 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.049944 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.049963 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.050981 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:48Z","lastTransitionTime":"2025-11-25T15:02:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.155606 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.155664 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.155685 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.155713 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.155733 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:48Z","lastTransitionTime":"2025-11-25T15:02:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.173544 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.173716 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 15:02:48 crc kubenswrapper[4890]: E1125 15:02:48.173906 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.173949 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 15:02:48 crc kubenswrapper[4890]: E1125 15:02:48.174202 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 15:02:48 crc kubenswrapper[4890]: E1125 15:02:48.174291 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.258723 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.258773 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.258783 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.258801 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.258813 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:48Z","lastTransitionTime":"2025-11-25T15:02:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.361218 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.361260 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.361270 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.361285 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.361294 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:48Z","lastTransitionTime":"2025-11-25T15:02:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.416023 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" event={"ID":"9a404958-dffa-4a6a-99fc-3e4950614bd4","Type":"ContainerStarted","Data":"c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611"} Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.418813 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zg454" event={"ID":"81411ec7-f865-48e2-b464-ab6213ef5e52","Type":"ContainerStarted","Data":"21e7825cb1d51e854b5218b09bca98a38c75b356ac9d3d6c206147e98b050dd3"} Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.440594 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86e031a7-b4c8-4030-a078-cdca7ccc9dde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4c4921f0973b03f6dade29c2ef32684a481cd1c3e1e23b64ed636943c9bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f9285acc7ac3cbb503a1e546728748c42971c8c9019cc8ca837076f3e55558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa240816a6e9d2821006ce240eccf27d8f57a7f0c5394314fa7177488da4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://659004bd47948def6b33498d8d1474f28c2e98986754f61af012267f3984269e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6d2dcd8df64e44df8cdb4c3a022f03fccb39451631ca35b30d4c7c35271ff0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:48Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.459930 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg454" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81411ec7-f865-48e2-b464-ab6213ef5e52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a366b314408eb8453ab7f970a243c511a2716b5472d92b03c81061135e4e59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a366b314408eb8453ab7f970a243c511a2716b5472d92b03c81061135e4e59c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f841cf4a886c5e17d9919787305ce297d7910b2692be53169324a56bd545be5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f841cf4a886c5e17d9919787305ce297d7910b2692be53169324a56bd545be5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e7825cb1d51e854b5218b09bca98a38c75b356ac9d3d6c206147e98b050dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg454\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:48Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.463690 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.463726 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.463736 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.463752 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.463762 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:48Z","lastTransitionTime":"2025-11-25T15:02:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.474191 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4txd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4322dbd78237bf443a985ce3c79bca4cba2103862d81177f5474a76f4905eed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfcq7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4txd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:48Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.487317 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583964bf462fa0005d471f82f6827552a60659ab97984eb1faf074c7b82c2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:48Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.500506 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:48Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.510517 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7g5s8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a7dc232-835b-491c-afc9-9eb14f257af3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9a7ac317fbb9c589573e80babe484c21bef4da3d9ab79dfb7ebf5b8300c472c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4cs2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7g5s8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:48Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.529062 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a404958-dffa-4a6a-99fc-3e4950614bd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zdk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:48Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.539011 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e4f849d-f239-4727-a73e-18327856929a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://490f1595bc77030d2ee789197fa3748324e1b7a0c9d9180ba08383bbac43adc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0902d6f8e56aecc5a2a13a7148f94a8a00ad139027b3c036aba0215ba0a67f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xx2tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:48Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.552743 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd5e648-6322-404d-bf1b-3b59a8092291\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76d277b4001db9d63b0af88b98afcc9019321101429c9be47b3022d3575d4e06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e3a186b14ea199d1aeba5751290c266ec918cb055120cfc5c8a2ae03b1469e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a860a998c40aee3e94e4cfd748f54391b3096fd3ed75c84c7ea2a5a7cefa33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://081a6d623cd043c0919d04e2ca93314b2b85fe33048634aa2637297d54b52cee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f506ec649ce24f4df6fd33e4aeafd13e42167dd71323524a5a3164517659fcc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:48Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.564431 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d5a9ccb-13ba-4dd5-9904-42aac5eca374\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d19bc48acb392051f0f6c5a6820135bf29ebe6d2afd098697553298933cf280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93325a6407389c4a7db0ba3c1a9c84a6980db3c3277cdf3860583a564b562bbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20505e91fc60158eeb3a8cfa0aafc5b4e4436d5def5e735a3a58b0f9f3621058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3281314d85a20945d56b75ab7b7aef1e382a73e74dff03773d118d03eb0ee16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:48Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.565771 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.565827 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.565837 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.565853 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.565872 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:48Z","lastTransitionTime":"2025-11-25T15:02:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.575073 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b33b805b76abbcace899e527d5ee3eaeb9e480fb03dc8eacd3f6dd90d2dd6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:48Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.586312 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:48Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.598849 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:48Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.610578 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3358a356ae5b422d6bba72bc516374e14ccdace80726aa71b10e25102e0f1100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba3acdae961f49f797f66bbebe350e34a4df5882bcf2a4fe8d37f2ca95865c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:48Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.619919 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8tdrh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1097de4-74e4-4178-b9b4-14a22539da9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfa0291da1598e930aeb4a76f9f4e8b0728001a770acd8f10215465539f8d8b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8tdrh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:48Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.668135 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.668189 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.668199 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.668215 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.668233 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:48Z","lastTransitionTime":"2025-11-25T15:02:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.771284 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.771334 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.771345 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.771363 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.771380 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:48Z","lastTransitionTime":"2025-11-25T15:02:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.873429 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.873474 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.873485 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.873503 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.873515 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:48Z","lastTransitionTime":"2025-11-25T15:02:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.976110 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.976145 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.976171 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.976188 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:48 crc kubenswrapper[4890]: I1125 15:02:48.976198 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:48Z","lastTransitionTime":"2025-11-25T15:02:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.078518 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.078559 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.078573 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.078590 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.078602 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:49Z","lastTransitionTime":"2025-11-25T15:02:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.163655 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.163754 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.163782 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.163811 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:02:49 crc kubenswrapper[4890]: E1125 15:02:49.163874 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:02:57.163849118 +0000 UTC m=+35.606311728 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:02:49 crc kubenswrapper[4890]: E1125 15:02:49.163907 4890 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.163940 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 15:02:49 crc kubenswrapper[4890]: E1125 15:02:49.163962 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 15:02:57.163946911 +0000 UTC m=+35.606409561 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 15:02:49 crc kubenswrapper[4890]: E1125 15:02:49.163985 4890 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 15:02:49 crc kubenswrapper[4890]: E1125 15:02:49.164008 4890 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 15:02:49 crc kubenswrapper[4890]: E1125 15:02:49.164008 4890 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 15:02:49 crc kubenswrapper[4890]: E1125 15:02:49.164098 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 15:02:57.164076954 +0000 UTC m=+35.606539614 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 15:02:49 crc kubenswrapper[4890]: E1125 15:02:49.164021 4890 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 15:02:49 crc kubenswrapper[4890]: E1125 15:02:49.164153 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 15:02:57.164142665 +0000 UTC m=+35.606605375 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 15:02:49 crc kubenswrapper[4890]: E1125 15:02:49.164026 4890 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 15:02:49 crc kubenswrapper[4890]: E1125 15:02:49.164205 4890 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 15:02:49 crc kubenswrapper[4890]: E1125 15:02:49.164218 4890 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 15:02:49 crc kubenswrapper[4890]: E1125 15:02:49.164255 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 15:02:57.164240798 +0000 UTC m=+35.606703508 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.180364 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.180395 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.180403 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.180416 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.180424 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:49Z","lastTransitionTime":"2025-11-25T15:02:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.283092 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.283185 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.283205 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.283226 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.283240 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:49Z","lastTransitionTime":"2025-11-25T15:02:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.386082 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.386125 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.386142 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.386179 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.386193 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:49Z","lastTransitionTime":"2025-11-25T15:02:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.424882 4890 generic.go:334] "Generic (PLEG): container finished" podID="81411ec7-f865-48e2-b464-ab6213ef5e52" containerID="21e7825cb1d51e854b5218b09bca98a38c75b356ac9d3d6c206147e98b050dd3" exitCode=0 Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.424943 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zg454" event={"ID":"81411ec7-f865-48e2-b464-ab6213ef5e52","Type":"ContainerDied","Data":"21e7825cb1d51e854b5218b09bca98a38c75b356ac9d3d6c206147e98b050dd3"} Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.450525 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86e031a7-b4c8-4030-a078-cdca7ccc9dde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4c4921f0973b03f6dade29c2ef32684a481cd1c3e1e23b64ed636943c9bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f9285acc7ac3cbb503a1e546728748c42971c8c9019cc8ca837076f3e55558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa240816a6e9d2821006ce240eccf27d8f57a7f0c5394314fa7177488da4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://659004bd47948def6b33498d8d1474f28c2e98986754f61af012267f3984269e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6d2dcd8df64e44df8cdb4c3a022f03fccb39451631ca35b30d4c7c35271ff0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:49Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.465220 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg454" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81411ec7-f865-48e2-b464-ab6213ef5e52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a366b314408eb8453ab7f970a243c511a2716b5472d92b03c81061135e4e59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a366b314408eb8453ab7f970a243c511a2716b5472d92b03c81061135e4e59c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f841cf4a886c5e17d9919787305ce297d7910b2692be53169324a56bd545be5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f841cf4a886c5e17d9919787305ce297d7910b2692be53169324a56bd545be5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e7825cb1d51e854b5218b09bca98a38c75b356ac9d3d6c206147e98b050dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21e7825cb1d51e854b5218b09bca98a38c75b356ac9d3d6c206147e98b050dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg454\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:49Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.483297 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4txd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4322dbd78237bf443a985ce3c79bca4cba2103862d81177f5474a76f4905eed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfcq7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4txd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:49Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.491152 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.491199 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.491209 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.491224 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.491233 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:49Z","lastTransitionTime":"2025-11-25T15:02:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.501545 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583964bf462fa0005d471f82f6827552a60659ab97984eb1faf074c7b82c2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:49Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.515655 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:49Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.526319 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7g5s8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a7dc232-835b-491c-afc9-9eb14f257af3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9a7ac317fbb9c589573e80babe484c21bef4da3d9ab79dfb7ebf5b8300c472c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4cs2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7g5s8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:49Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.543228 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a404958-dffa-4a6a-99fc-3e4950614bd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zdk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:49Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.554430 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e4f849d-f239-4727-a73e-18327856929a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://490f1595bc77030d2ee789197fa3748324e1b7a0c9d9180ba08383bbac43adc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0902d6f8e56aecc5a2a13a7148f94a8a00ad139027b3c036aba0215ba0a67f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xx2tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:49Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.566236 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd5e648-6322-404d-bf1b-3b59a8092291\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76d277b4001db9d63b0af88b98afcc9019321101429c9be47b3022d3575d4e06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e3a186b14ea199d1aeba5751290c266ec918cb055120cfc5c8a2ae03b1469e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a860a998c40aee3e94e4cfd748f54391b3096fd3ed75c84c7ea2a5a7cefa33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://081a6d623cd043c0919d04e2ca93314b2b85fe33048634aa2637297d54b52cee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f506ec649ce24f4df6fd33e4aeafd13e42167dd71323524a5a3164517659fcc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:49Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.578595 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d5a9ccb-13ba-4dd5-9904-42aac5eca374\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d19bc48acb392051f0f6c5a6820135bf29ebe6d2afd098697553298933cf280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93325a6407389c4a7db0ba3c1a9c84a6980db3c3277cdf3860583a564b562bbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20505e91fc60158eeb3a8cfa0aafc5b4e4436d5def5e735a3a58b0f9f3621058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3281314d85a20945d56b75ab7b7aef1e382a73e74dff03773d118d03eb0ee16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:49Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.589238 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b33b805b76abbcace899e527d5ee3eaeb9e480fb03dc8eacd3f6dd90d2dd6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:49Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.593449 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.593483 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.593494 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.593508 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.593517 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:49Z","lastTransitionTime":"2025-11-25T15:02:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.602096 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:49Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.613630 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:49Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.625764 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3358a356ae5b422d6bba72bc516374e14ccdace80726aa71b10e25102e0f1100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba3acdae961f49f797f66bbebe350e34a4df5882bcf2a4fe8d37f2ca95865c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:49Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.636906 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8tdrh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1097de4-74e4-4178-b9b4-14a22539da9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfa0291da1598e930aeb4a76f9f4e8b0728001a770acd8f10215465539f8d8b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8tdrh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:49Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.695503 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.695561 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.695573 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.695595 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.695608 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:49Z","lastTransitionTime":"2025-11-25T15:02:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.798265 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.798323 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.798343 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.798365 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.798380 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:49Z","lastTransitionTime":"2025-11-25T15:02:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.900475 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.900509 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.900522 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.900538 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:49 crc kubenswrapper[4890]: I1125 15:02:49.900549 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:49Z","lastTransitionTime":"2025-11-25T15:02:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.002382 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.002422 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.002441 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.002456 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.002466 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:50Z","lastTransitionTime":"2025-11-25T15:02:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.104236 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.104291 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.104302 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.104318 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.104328 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:50Z","lastTransitionTime":"2025-11-25T15:02:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.174453 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 15:02:50 crc kubenswrapper[4890]: E1125 15:02:50.174618 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.174678 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 15:02:50 crc kubenswrapper[4890]: E1125 15:02:50.174734 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.174863 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:02:50 crc kubenswrapper[4890]: E1125 15:02:50.174970 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.207091 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.207127 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.207135 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.207148 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.207174 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:50Z","lastTransitionTime":"2025-11-25T15:02:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.310286 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.310337 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.310347 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.310362 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.310372 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:50Z","lastTransitionTime":"2025-11-25T15:02:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.413723 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.413801 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.413823 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.413884 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.413904 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:50Z","lastTransitionTime":"2025-11-25T15:02:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.431622 4890 generic.go:334] "Generic (PLEG): container finished" podID="81411ec7-f865-48e2-b464-ab6213ef5e52" containerID="5881047b9fa64b7504f1d62b697a6f352b05d14eedd61e6b64d5ff99d91edda2" exitCode=0 Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.431708 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zg454" event={"ID":"81411ec7-f865-48e2-b464-ab6213ef5e52","Type":"ContainerDied","Data":"5881047b9fa64b7504f1d62b697a6f352b05d14eedd61e6b64d5ff99d91edda2"} Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.445116 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" event={"ID":"9a404958-dffa-4a6a-99fc-3e4950614bd4","Type":"ContainerStarted","Data":"84364aff9c933d3c3e94ba252568db3ce9296b02f8290b087631173b7500c202"} Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.446086 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.446152 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.453805 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd5e648-6322-404d-bf1b-3b59a8092291\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76d277b4001db9d63b0af88b98afcc9019321101429c9be47b3022d3575d4e06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e3a186b14ea199d1aeba5751290c266ec918cb055120cfc5c8a2ae03b1469e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a860a998c40aee3e94e4cfd748f54391b3096fd3ed75c84c7ea2a5a7cefa33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://081a6d623cd043c0919d04e2ca93314b2b85fe33048634aa2637297d54b52cee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f506ec649ce24f4df6fd33e4aeafd13e42167dd71323524a5a3164517659fcc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:50Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.482349 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d5a9ccb-13ba-4dd5-9904-42aac5eca374\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d19bc48acb392051f0f6c5a6820135bf29ebe6d2afd098697553298933cf280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93325a6407389c4a7db0ba3c1a9c84a6980db3c3277cdf3860583a564b562bbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20505e91fc60158eeb3a8cfa0aafc5b4e4436d5def5e735a3a58b0f9f3621058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3281314d85a20945d56b75ab7b7aef1e382a73e74dff03773d118d03eb0ee16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:50Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.483580 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.486711 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.497726 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b33b805b76abbcace899e527d5ee3eaeb9e480fb03dc8eacd3f6dd90d2dd6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:50Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.512792 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8tdrh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1097de4-74e4-4178-b9b4-14a22539da9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfa0291da1598e930aeb4a76f9f4e8b0728001a770acd8f10215465539f8d8b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8tdrh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:50Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.516690 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.516758 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.516770 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.516787 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.516800 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:50Z","lastTransitionTime":"2025-11-25T15:02:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.529865 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:50Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.549040 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:50Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.564052 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3358a356ae5b422d6bba72bc516374e14ccdace80726aa71b10e25102e0f1100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba3acdae961f49f797f66bbebe350e34a4df5882bcf2a4fe8d37f2ca95865c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:50Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.588795 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86e031a7-b4c8-4030-a078-cdca7ccc9dde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4c4921f0973b03f6dade29c2ef32684a481cd1c3e1e23b64ed636943c9bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f9285acc7ac3cbb503a1e546728748c42971c8c9019cc8ca837076f3e55558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa240816a6e9d2821006ce240eccf27d8f57a7f0c5394314fa7177488da4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://659004bd47948def6b33498d8d1474f28c2e98986754f61af012267f3984269e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6d2dcd8df64e44df8cdb4c3a022f03fccb39451631ca35b30d4c7c35271ff0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:50Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.604831 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg454" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81411ec7-f865-48e2-b464-ab6213ef5e52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a366b314408eb8453ab7f970a243c511a2716b5472d92b03c81061135e4e59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a366b314408eb8453ab7f970a243c511a2716b5472d92b03c81061135e4e59c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f841cf4a886c5e17d9919787305ce297d7910b2692be53169324a56bd545be5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f841cf4a886c5e17d9919787305ce297d7910b2692be53169324a56bd545be5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e7825cb1d51e854b5218b09bca98a38c75b356ac9d3d6c206147e98b050dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21e7825cb1d51e854b5218b09bca98a38c75b356ac9d3d6c206147e98b050dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5881047b9fa64b7504f1d62b697a6f352b05d14eedd61e6b64d5ff99d91edda2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5881047b9fa64b7504f1d62b697a6f352b05d14eedd61e6b64d5ff99d91edda2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg454\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:50Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.619912 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.619952 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.619965 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.619980 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.619992 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:50Z","lastTransitionTime":"2025-11-25T15:02:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.623961 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4txd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4322dbd78237bf443a985ce3c79bca4cba2103862d81177f5474a76f4905eed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfcq7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4txd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:50Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.644148 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e4f849d-f239-4727-a73e-18327856929a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://490f1595bc77030d2ee789197fa3748324e1b7a0c9d9180ba08383bbac43adc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0902d6f8e56aecc5a2a13a7148f94a8a00ad139027b3c036aba0215ba0a67f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xx2tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:50Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.662367 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583964bf462fa0005d471f82f6827552a60659ab97984eb1faf074c7b82c2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:50Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.678240 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:50Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.693552 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7g5s8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a7dc232-835b-491c-afc9-9eb14f257af3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9a7ac317fbb9c589573e80babe484c21bef4da3d9ab79dfb7ebf5b8300c472c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4cs2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7g5s8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:50Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.722708 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.722760 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.722771 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.722788 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.722801 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:50Z","lastTransitionTime":"2025-11-25T15:02:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.726137 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a404958-dffa-4a6a-99fc-3e4950614bd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zdk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:50Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.741620 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e4f849d-f239-4727-a73e-18327856929a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://490f1595bc77030d2ee789197fa3748324e1b7a0c9d9180ba08383bbac43adc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0902d6f8e56aecc5a2a13a7148f94a8a00ad139027b3c036aba0215ba0a67f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xx2tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:50Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.757751 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583964bf462fa0005d471f82f6827552a60659ab97984eb1faf074c7b82c2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:50Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.773140 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:50Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.783725 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7g5s8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a7dc232-835b-491c-afc9-9eb14f257af3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9a7ac317fbb9c589573e80babe484c21bef4da3d9ab79dfb7ebf5b8300c472c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4cs2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7g5s8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:50Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.801462 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a404958-dffa-4a6a-99fc-3e4950614bd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84364aff9c933d3c3e94ba252568db3ce9296b02f8290b087631173b7500c202\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zdk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:50Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.815033 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd5e648-6322-404d-bf1b-3b59a8092291\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76d277b4001db9d63b0af88b98afcc9019321101429c9be47b3022d3575d4e06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e3a186b14ea199d1aeba5751290c266ec918cb055120cfc5c8a2ae03b1469e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a860a998c40aee3e94e4cfd748f54391b3096fd3ed75c84c7ea2a5a7cefa33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://081a6d623cd043c0919d04e2ca93314b2b85fe33048634aa2637297d54b52cee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f506ec649ce24f4df6fd33e4aeafd13e42167dd71323524a5a3164517659fcc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:50Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.825378 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.825411 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.825419 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.825433 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.825452 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:50Z","lastTransitionTime":"2025-11-25T15:02:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.827820 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d5a9ccb-13ba-4dd5-9904-42aac5eca374\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d19bc48acb392051f0f6c5a6820135bf29ebe6d2afd098697553298933cf280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93325a6407389c4a7db0ba3c1a9c84a6980db3c3277cdf3860583a564b562bbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20505e91fc60158eeb3a8cfa0aafc5b4e4436d5def5e735a3a58b0f9f3621058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3281314d85a20945d56b75ab7b7aef1e382a73e74dff03773d118d03eb0ee16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:50Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.840251 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b33b805b76abbcace899e527d5ee3eaeb9e480fb03dc8eacd3f6dd90d2dd6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:50Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.849587 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8tdrh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1097de4-74e4-4178-b9b4-14a22539da9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfa0291da1598e930aeb4a76f9f4e8b0728001a770acd8f10215465539f8d8b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8tdrh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:50Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.860327 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:50Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.871776 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:50Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.883504 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3358a356ae5b422d6bba72bc516374e14ccdace80726aa71b10e25102e0f1100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba3acdae961f49f797f66bbebe350e34a4df5882bcf2a4fe8d37f2ca95865c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:50Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.903331 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86e031a7-b4c8-4030-a078-cdca7ccc9dde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4c4921f0973b03f6dade29c2ef32684a481cd1c3e1e23b64ed636943c9bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f9285acc7ac3cbb503a1e546728748c42971c8c9019cc8ca837076f3e55558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa240816a6e9d2821006ce240eccf27d8f57a7f0c5394314fa7177488da4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://659004bd47948def6b33498d8d1474f28c2e98986754f61af012267f3984269e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6d2dcd8df64e44df8cdb4c3a022f03fccb39451631ca35b30d4c7c35271ff0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:50Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.918968 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg454" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81411ec7-f865-48e2-b464-ab6213ef5e52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a366b314408eb8453ab7f970a243c511a2716b5472d92b03c81061135e4e59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a366b314408eb8453ab7f970a243c511a2716b5472d92b03c81061135e4e59c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f841cf4a886c5e17d9919787305ce297d7910b2692be53169324a56bd545be5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f841cf4a886c5e17d9919787305ce297d7910b2692be53169324a56bd545be5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e7825cb1d51e854b5218b09bca98a38c75b356ac9d3d6c206147e98b050dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21e7825cb1d51e854b5218b09bca98a38c75b356ac9d3d6c206147e98b050dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5881047b9fa64b7504f1d62b697a6f352b05d14eedd61e6b64d5ff99d91edda2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5881047b9fa64b7504f1d62b697a6f352b05d14eedd61e6b64d5ff99d91edda2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg454\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:50Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.929312 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.929372 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.929388 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.929408 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.929419 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:50Z","lastTransitionTime":"2025-11-25T15:02:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:50 crc kubenswrapper[4890]: I1125 15:02:50.933120 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4txd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4322dbd78237bf443a985ce3c79bca4cba2103862d81177f5474a76f4905eed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfcq7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4txd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:50Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.031919 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.031971 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.031983 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.032005 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.032017 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:51Z","lastTransitionTime":"2025-11-25T15:02:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.135175 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.135234 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.135249 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.135272 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.135285 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:51Z","lastTransitionTime":"2025-11-25T15:02:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.237791 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.237840 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.237848 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.237867 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.237880 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:51Z","lastTransitionTime":"2025-11-25T15:02:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.341875 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.341952 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.341966 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.341989 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.342012 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:51Z","lastTransitionTime":"2025-11-25T15:02:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.445487 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.445542 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.445551 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.445571 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.445582 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:51Z","lastTransitionTime":"2025-11-25T15:02:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.453254 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zg454" event={"ID":"81411ec7-f865-48e2-b464-ab6213ef5e52","Type":"ContainerStarted","Data":"db59b77d323079e0b588a6adc0f247dda84b5497de17264390366124934cdca8"} Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.453290 4890 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.493450 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86e031a7-b4c8-4030-a078-cdca7ccc9dde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4c4921f0973b03f6dade29c2ef32684a481cd1c3e1e23b64ed636943c9bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f9285acc7ac3cbb503a1e546728748c42971c8c9019cc8ca837076f3e55558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa240816a6e9d2821006ce240eccf27d8f57a7f0c5394314fa7177488da4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://659004bd47948def6b33498d8d1474f28c2e98986754f61af012267f3984269e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6d2dcd8df64e44df8cdb4c3a022f03fccb39451631ca35b30d4c7c35271ff0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:51Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.516847 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg454" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81411ec7-f865-48e2-b464-ab6213ef5e52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db59b77d323079e0b588a6adc0f247dda84b5497de17264390366124934cdca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a366b314408eb8453ab7f970a243c511a2716b5472d92b03c81061135e4e59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a366b314408eb8453ab7f970a243c511a2716b5472d92b03c81061135e4e59c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f841cf4a886c5e17d9919787305ce297d7910b2692be53169324a56bd545be5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f841cf4a886c5e17d9919787305ce297d7910b2692be53169324a56bd545be5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e7825cb1d51e854b5218b09bca98a38c75b356ac9d3d6c206147e98b050dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21e7825cb1d51e854b5218b09bca98a38c75b356ac9d3d6c206147e98b050dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5881047b9fa64b7504f1d62b697a6f352b05d14eedd61e6b64d5ff99d91edda2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5881047b9fa64b7504f1d62b697a6f352b05d14eedd61e6b64d5ff99d91edda2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg454\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:51Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.535479 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4txd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4322dbd78237bf443a985ce3c79bca4cba2103862d81177f5474a76f4905eed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfcq7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4txd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:51Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.548870 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.548964 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.548980 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.549002 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.549019 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:51Z","lastTransitionTime":"2025-11-25T15:02:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.549691 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583964bf462fa0005d471f82f6827552a60659ab97984eb1faf074c7b82c2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:51Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.563665 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:51Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.575386 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7g5s8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a7dc232-835b-491c-afc9-9eb14f257af3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9a7ac317fbb9c589573e80babe484c21bef4da3d9ab79dfb7ebf5b8300c472c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4cs2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7g5s8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:51Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.593266 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a404958-dffa-4a6a-99fc-3e4950614bd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84364aff9c933d3c3e94ba252568db3ce9296b02f8290b087631173b7500c202\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zdk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:51Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.605811 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e4f849d-f239-4727-a73e-18327856929a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://490f1595bc77030d2ee789197fa3748324e1b7a0c9d9180ba08383bbac43adc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0902d6f8e56aecc5a2a13a7148f94a8a00ad139027b3c036aba0215ba0a67f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xx2tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:51Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.620200 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd5e648-6322-404d-bf1b-3b59a8092291\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76d277b4001db9d63b0af88b98afcc9019321101429c9be47b3022d3575d4e06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e3a186b14ea199d1aeba5751290c266ec918cb055120cfc5c8a2ae03b1469e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a860a998c40aee3e94e4cfd748f54391b3096fd3ed75c84c7ea2a5a7cefa33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://081a6d623cd043c0919d04e2ca93314b2b85fe33048634aa2637297d54b52cee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f506ec649ce24f4df6fd33e4aeafd13e42167dd71323524a5a3164517659fcc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:51Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.635215 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d5a9ccb-13ba-4dd5-9904-42aac5eca374\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d19bc48acb392051f0f6c5a6820135bf29ebe6d2afd098697553298933cf280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93325a6407389c4a7db0ba3c1a9c84a6980db3c3277cdf3860583a564b562bbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20505e91fc60158eeb3a8cfa0aafc5b4e4436d5def5e735a3a58b0f9f3621058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3281314d85a20945d56b75ab7b7aef1e382a73e74dff03773d118d03eb0ee16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:51Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.648368 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b33b805b76abbcace899e527d5ee3eaeb9e480fb03dc8eacd3f6dd90d2dd6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:51Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.651107 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.651148 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.651178 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.651193 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.651202 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:51Z","lastTransitionTime":"2025-11-25T15:02:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.665299 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:51Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.680577 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:51Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.695223 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3358a356ae5b422d6bba72bc516374e14ccdace80726aa71b10e25102e0f1100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba3acdae961f49f797f66bbebe350e34a4df5882bcf2a4fe8d37f2ca95865c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:51Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.709502 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8tdrh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1097de4-74e4-4178-b9b4-14a22539da9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfa0291da1598e930aeb4a76f9f4e8b0728001a770acd8f10215465539f8d8b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8tdrh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:51Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.753208 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.753246 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.753257 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.753274 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.753284 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:51Z","lastTransitionTime":"2025-11-25T15:02:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.856240 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.856296 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.856307 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.856327 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.856341 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:51Z","lastTransitionTime":"2025-11-25T15:02:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.959741 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.959797 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.959810 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.959832 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:51 crc kubenswrapper[4890]: I1125 15:02:51.959845 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:51Z","lastTransitionTime":"2025-11-25T15:02:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.062992 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.063038 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.063047 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.063061 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.063072 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:52Z","lastTransitionTime":"2025-11-25T15:02:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.165911 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.165973 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.165992 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.166017 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.166034 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:52Z","lastTransitionTime":"2025-11-25T15:02:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.173242 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.173296 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 15:02:52 crc kubenswrapper[4890]: E1125 15:02:52.173383 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 15:02:52 crc kubenswrapper[4890]: E1125 15:02:52.173481 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.173295 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 15:02:52 crc kubenswrapper[4890]: E1125 15:02:52.173665 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.204193 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a404958-dffa-4a6a-99fc-3e4950614bd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84364aff9c933d3c3e94ba252568db3ce9296b02f8290b087631173b7500c202\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zdk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:52Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.218912 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e4f849d-f239-4727-a73e-18327856929a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://490f1595bc77030d2ee789197fa3748324e1b7a0c9d9180ba08383bbac43adc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0902d6f8e56aecc5a2a13a7148f94a8a00ad139027b3c036aba0215ba0a67f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xx2tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:52Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.234087 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583964bf462fa0005d471f82f6827552a60659ab97984eb1faf074c7b82c2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:52Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.248079 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:52Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.264338 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7g5s8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a7dc232-835b-491c-afc9-9eb14f257af3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9a7ac317fbb9c589573e80babe484c21bef4da3d9ab79dfb7ebf5b8300c472c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4cs2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7g5s8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:52Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.268456 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.268537 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.268553 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.268577 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.268593 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:52Z","lastTransitionTime":"2025-11-25T15:02:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.279996 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b33b805b76abbcace899e527d5ee3eaeb9e480fb03dc8eacd3f6dd90d2dd6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:52Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.298827 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd5e648-6322-404d-bf1b-3b59a8092291\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76d277b4001db9d63b0af88b98afcc9019321101429c9be47b3022d3575d4e06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e3a186b14ea199d1aeba5751290c266ec918cb055120cfc5c8a2ae03b1469e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a860a998c40aee3e94e4cfd748f54391b3096fd3ed75c84c7ea2a5a7cefa33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://081a6d623cd043c0919d04e2ca93314b2b85fe33048634aa2637297d54b52cee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f506ec649ce24f4df6fd33e4aeafd13e42167dd71323524a5a3164517659fcc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:52Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.318312 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d5a9ccb-13ba-4dd5-9904-42aac5eca374\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d19bc48acb392051f0f6c5a6820135bf29ebe6d2afd098697553298933cf280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93325a6407389c4a7db0ba3c1a9c84a6980db3c3277cdf3860583a564b562bbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20505e91fc60158eeb3a8cfa0aafc5b4e4436d5def5e735a3a58b0f9f3621058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3281314d85a20945d56b75ab7b7aef1e382a73e74dff03773d118d03eb0ee16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:52Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.335824 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3358a356ae5b422d6bba72bc516374e14ccdace80726aa71b10e25102e0f1100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba3acdae961f49f797f66bbebe350e34a4df5882bcf2a4fe8d37f2ca95865c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:52Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.346624 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8tdrh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1097de4-74e4-4178-b9b4-14a22539da9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfa0291da1598e930aeb4a76f9f4e8b0728001a770acd8f10215465539f8d8b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8tdrh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:52Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.358665 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:52Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.371141 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:52Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.371944 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.371971 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.371980 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.371995 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.372005 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:52Z","lastTransitionTime":"2025-11-25T15:02:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.419566 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86e031a7-b4c8-4030-a078-cdca7ccc9dde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4c4921f0973b03f6dade29c2ef32684a481cd1c3e1e23b64ed636943c9bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f9285acc7ac3cbb503a1e546728748c42971c8c9019cc8ca837076f3e55558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa240816a6e9d2821006ce240eccf27d8f57a7f0c5394314fa7177488da4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://659004bd47948def6b33498d8d1474f28c2e98986754f61af012267f3984269e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6d2dcd8df64e44df8cdb4c3a022f03fccb39451631ca35b30d4c7c35271ff0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:52Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.443076 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg454" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81411ec7-f865-48e2-b464-ab6213ef5e52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db59b77d323079e0b588a6adc0f247dda84b5497de17264390366124934cdca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a366b314408eb8453ab7f970a243c511a2716b5472d92b03c81061135e4e59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a366b314408eb8453ab7f970a243c511a2716b5472d92b03c81061135e4e59c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f841cf4a886c5e17d9919787305ce297d7910b2692be53169324a56bd545be5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f841cf4a886c5e17d9919787305ce297d7910b2692be53169324a56bd545be5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e7825cb1d51e854b5218b09bca98a38c75b356ac9d3d6c206147e98b050dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21e7825cb1d51e854b5218b09bca98a38c75b356ac9d3d6c206147e98b050dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5881047b9fa64b7504f1d62b697a6f352b05d14eedd61e6b64d5ff99d91edda2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5881047b9fa64b7504f1d62b697a6f352b05d14eedd61e6b64d5ff99d91edda2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg454\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:52Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.456614 4890 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.465487 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4txd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4322dbd78237bf443a985ce3c79bca4cba2103862d81177f5474a76f4905eed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfcq7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4txd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:52Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.474819 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.474899 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.474931 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.474955 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.474968 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:52Z","lastTransitionTime":"2025-11-25T15:02:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.578320 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.578362 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.578374 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.578393 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.578405 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:52Z","lastTransitionTime":"2025-11-25T15:02:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.681491 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.681547 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.681556 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.681575 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.681587 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:52Z","lastTransitionTime":"2025-11-25T15:02:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.784482 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.784543 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.784561 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.784583 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.784599 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:52Z","lastTransitionTime":"2025-11-25T15:02:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.888027 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.888109 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.888119 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.888175 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.888193 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:52Z","lastTransitionTime":"2025-11-25T15:02:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.991009 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.991093 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.991125 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.991149 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:52 crc kubenswrapper[4890]: I1125 15:02:52.991192 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:52Z","lastTransitionTime":"2025-11-25T15:02:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.094589 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.094672 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.094687 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.094709 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.094723 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:53Z","lastTransitionTime":"2025-11-25T15:02:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.197564 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.197604 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.197618 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.197634 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.197644 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:53Z","lastTransitionTime":"2025-11-25T15:02:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.304418 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.304469 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.304482 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.304500 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.304510 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:53Z","lastTransitionTime":"2025-11-25T15:02:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.406714 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.406750 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.406758 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.406771 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.406779 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:53Z","lastTransitionTime":"2025-11-25T15:02:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.463014 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7zdk8_9a404958-dffa-4a6a-99fc-3e4950614bd4/ovnkube-controller/0.log" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.467028 4890 generic.go:334] "Generic (PLEG): container finished" podID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerID="84364aff9c933d3c3e94ba252568db3ce9296b02f8290b087631173b7500c202" exitCode=1 Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.467104 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" event={"ID":"9a404958-dffa-4a6a-99fc-3e4950614bd4","Type":"ContainerDied","Data":"84364aff9c933d3c3e94ba252568db3ce9296b02f8290b087631173b7500c202"} Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.469989 4890 scope.go:117] "RemoveContainer" containerID="84364aff9c933d3c3e94ba252568db3ce9296b02f8290b087631173b7500c202" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.488567 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:53Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.509861 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.510362 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.510378 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.510398 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.510414 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:53Z","lastTransitionTime":"2025-11-25T15:02:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.511588 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:53Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.529601 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3358a356ae5b422d6bba72bc516374e14ccdace80726aa71b10e25102e0f1100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba3acdae961f49f797f66bbebe350e34a4df5882bcf2a4fe8d37f2ca95865c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:53Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.542219 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8tdrh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1097de4-74e4-4178-b9b4-14a22539da9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfa0291da1598e930aeb4a76f9f4e8b0728001a770acd8f10215465539f8d8b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8tdrh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:53Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.565856 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86e031a7-b4c8-4030-a078-cdca7ccc9dde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4c4921f0973b03f6dade29c2ef32684a481cd1c3e1e23b64ed636943c9bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f9285acc7ac3cbb503a1e546728748c42971c8c9019cc8ca837076f3e55558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa240816a6e9d2821006ce240eccf27d8f57a7f0c5394314fa7177488da4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://659004bd47948def6b33498d8d1474f28c2e98986754f61af012267f3984269e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6d2dcd8df64e44df8cdb4c3a022f03fccb39451631ca35b30d4c7c35271ff0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:53Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.581371 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg454" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81411ec7-f865-48e2-b464-ab6213ef5e52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db59b77d323079e0b588a6adc0f247dda84b5497de17264390366124934cdca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a366b314408eb8453ab7f970a243c511a2716b5472d92b03c81061135e4e59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a366b314408eb8453ab7f970a243c511a2716b5472d92b03c81061135e4e59c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f841cf4a886c5e17d9919787305ce297d7910b2692be53169324a56bd545be5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f841cf4a886c5e17d9919787305ce297d7910b2692be53169324a56bd545be5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e7825cb1d51e854b5218b09bca98a38c75b356ac9d3d6c206147e98b050dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21e7825cb1d51e854b5218b09bca98a38c75b356ac9d3d6c206147e98b050dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5881047b9fa64b7504f1d62b697a6f352b05d14eedd61e6b64d5ff99d91edda2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5881047b9fa64b7504f1d62b697a6f352b05d14eedd61e6b64d5ff99d91edda2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg454\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:53Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.593488 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4txd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4322dbd78237bf443a985ce3c79bca4cba2103862d81177f5474a76f4905eed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfcq7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4txd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:53Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.608011 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583964bf462fa0005d471f82f6827552a60659ab97984eb1faf074c7b82c2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:53Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.613304 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.613353 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.613366 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.613385 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.613397 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:53Z","lastTransitionTime":"2025-11-25T15:02:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.620874 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:53Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.635498 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7g5s8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a7dc232-835b-491c-afc9-9eb14f257af3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9a7ac317fbb9c589573e80babe484c21bef4da3d9ab79dfb7ebf5b8300c472c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4cs2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7g5s8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:53Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.655785 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a404958-dffa-4a6a-99fc-3e4950614bd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84364aff9c933d3c3e94ba252568db3ce9296b02f8290b087631173b7500c202\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84364aff9c933d3c3e94ba252568db3ce9296b02f8290b087631173b7500c202\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T15:02:52Z\\\",\\\"message\\\":\\\".io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1125 15:02:52.386129 6142 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1125 15:02:52.389228 6142 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 15:02:52.389548 6142 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1125 15:02:52.389576 6142 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1125 15:02:52.389596 6142 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 15:02:52.389628 6142 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 15:02:52.400291 6142 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 15:02:52.400526 6142 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 15:02:52.400551 6142 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 15:02:52.400574 6142 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 15:02:52.400597 6142 factory.go:656] Stopping watch factory\\\\nI1125 15:02:52.400612 6142 ovnkube.go:599] Stopped ovnkube\\\\nI1125 15:02:52.400642 6142 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 15:02:52.400649 6142 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 15:02:52.400655 6142 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 15:02:5\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zdk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:53Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.668190 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e4f849d-f239-4727-a73e-18327856929a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://490f1595bc77030d2ee789197fa3748324e1b7a0c9d9180ba08383bbac43adc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0902d6f8e56aecc5a2a13a7148f94a8a00ad139027b3c036aba0215ba0a67f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xx2tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:53Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.683607 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd5e648-6322-404d-bf1b-3b59a8092291\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76d277b4001db9d63b0af88b98afcc9019321101429c9be47b3022d3575d4e06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e3a186b14ea199d1aeba5751290c266ec918cb055120cfc5c8a2ae03b1469e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a860a998c40aee3e94e4cfd748f54391b3096fd3ed75c84c7ea2a5a7cefa33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://081a6d623cd043c0919d04e2ca93314b2b85fe33048634aa2637297d54b52cee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f506ec649ce24f4df6fd33e4aeafd13e42167dd71323524a5a3164517659fcc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:53Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.699619 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d5a9ccb-13ba-4dd5-9904-42aac5eca374\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d19bc48acb392051f0f6c5a6820135bf29ebe6d2afd098697553298933cf280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93325a6407389c4a7db0ba3c1a9c84a6980db3c3277cdf3860583a564b562bbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20505e91fc60158eeb3a8cfa0aafc5b4e4436d5def5e735a3a58b0f9f3621058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3281314d85a20945d56b75ab7b7aef1e382a73e74dff03773d118d03eb0ee16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:53Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.715890 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b33b805b76abbcace899e527d5ee3eaeb9e480fb03dc8eacd3f6dd90d2dd6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:53Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.716280 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.716324 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.716341 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.716367 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.716386 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:53Z","lastTransitionTime":"2025-11-25T15:02:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.819182 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.819230 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.819240 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.819281 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.819297 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:53Z","lastTransitionTime":"2025-11-25T15:02:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.922670 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.922724 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.922736 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.922756 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:53 crc kubenswrapper[4890]: I1125 15:02:53.922794 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:53Z","lastTransitionTime":"2025-11-25T15:02:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.025902 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.025993 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.026024 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.026062 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.026082 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:54Z","lastTransitionTime":"2025-11-25T15:02:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.128315 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.128369 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.128380 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.128395 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.128405 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:54Z","lastTransitionTime":"2025-11-25T15:02:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.172647 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.172755 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 15:02:54 crc kubenswrapper[4890]: E1125 15:02:54.172824 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 15:02:54 crc kubenswrapper[4890]: E1125 15:02:54.172970 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.173045 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 15:02:54 crc kubenswrapper[4890]: E1125 15:02:54.173138 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.233498 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.233573 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.233586 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.233610 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.233624 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:54Z","lastTransitionTime":"2025-11-25T15:02:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.337480 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.337529 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.337539 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.337556 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.337617 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:54Z","lastTransitionTime":"2025-11-25T15:02:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.441223 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.441302 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.441324 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.441356 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.441376 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:54Z","lastTransitionTime":"2025-11-25T15:02:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.474084 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7zdk8_9a404958-dffa-4a6a-99fc-3e4950614bd4/ovnkube-controller/0.log" Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.477655 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" event={"ID":"9a404958-dffa-4a6a-99fc-3e4950614bd4","Type":"ContainerStarted","Data":"f2f3af2cdb804a47cddb121d619a7545a54af0888b120babca0aa2217ff01d7b"} Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.544231 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.544282 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.544293 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.544313 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.544323 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:54Z","lastTransitionTime":"2025-11-25T15:02:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.646976 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.647026 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.647039 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.647054 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.647066 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:54Z","lastTransitionTime":"2025-11-25T15:02:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.749890 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.749955 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.749969 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.749991 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.750006 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:54Z","lastTransitionTime":"2025-11-25T15:02:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.853245 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.853300 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.853313 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.853335 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.853349 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:54Z","lastTransitionTime":"2025-11-25T15:02:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.956215 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.956268 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.956425 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.956456 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.956470 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:54Z","lastTransitionTime":"2025-11-25T15:02:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.966201 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s"] Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.966891 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s" Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.969262 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.970182 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.982640 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:54Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:54 crc kubenswrapper[4890]: I1125 15:02:54.998474 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:54Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.019585 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3358a356ae5b422d6bba72bc516374e14ccdace80726aa71b10e25102e0f1100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba3acdae961f49f797f66bbebe350e34a4df5882bcf2a4fe8d37f2ca95865c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:55Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.025843 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvcr5\" (UniqueName: \"kubernetes.io/projected/a299b8e0-37bc-48d9-a7e8-64ced2d92d40-kube-api-access-rvcr5\") pod \"ovnkube-control-plane-749d76644c-6wr7s\" (UID: \"a299b8e0-37bc-48d9-a7e8-64ced2d92d40\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.025938 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a299b8e0-37bc-48d9-a7e8-64ced2d92d40-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-6wr7s\" (UID: \"a299b8e0-37bc-48d9-a7e8-64ced2d92d40\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.025965 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a299b8e0-37bc-48d9-a7e8-64ced2d92d40-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-6wr7s\" (UID: \"a299b8e0-37bc-48d9-a7e8-64ced2d92d40\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.025998 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a299b8e0-37bc-48d9-a7e8-64ced2d92d40-env-overrides\") pod \"ovnkube-control-plane-749d76644c-6wr7s\" (UID: \"a299b8e0-37bc-48d9-a7e8-64ced2d92d40\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.033180 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8tdrh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1097de4-74e4-4178-b9b4-14a22539da9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfa0291da1598e930aeb4a76f9f4e8b0728001a770acd8f10215465539f8d8b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8tdrh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:55Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.055941 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86e031a7-b4c8-4030-a078-cdca7ccc9dde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4c4921f0973b03f6dade29c2ef32684a481cd1c3e1e23b64ed636943c9bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f9285acc7ac3cbb503a1e546728748c42971c8c9019cc8ca837076f3e55558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa240816a6e9d2821006ce240eccf27d8f57a7f0c5394314fa7177488da4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://659004bd47948def6b33498d8d1474f28c2e98986754f61af012267f3984269e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6d2dcd8df64e44df8cdb4c3a022f03fccb39451631ca35b30d4c7c35271ff0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:55Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.058600 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.058636 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.058645 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.058660 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.058670 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:55Z","lastTransitionTime":"2025-11-25T15:02:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.072786 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg454" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81411ec7-f865-48e2-b464-ab6213ef5e52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db59b77d323079e0b588a6adc0f247dda84b5497de17264390366124934cdca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a366b314408eb8453ab7f970a243c511a2716b5472d92b03c81061135e4e59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a366b314408eb8453ab7f970a243c511a2716b5472d92b03c81061135e4e59c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f841cf4a886c5e17d9919787305ce297d7910b2692be53169324a56bd545be5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f841cf4a886c5e17d9919787305ce297d7910b2692be53169324a56bd545be5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e7825cb1d51e854b5218b09bca98a38c75b356ac9d3d6c206147e98b050dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21e7825cb1d51e854b5218b09bca98a38c75b356ac9d3d6c206147e98b050dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5881047b9fa64b7504f1d62b697a6f352b05d14eedd61e6b64d5ff99d91edda2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5881047b9fa64b7504f1d62b697a6f352b05d14eedd61e6b64d5ff99d91edda2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg454\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:55Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.088425 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4txd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4322dbd78237bf443a985ce3c79bca4cba2103862d81177f5474a76f4905eed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfcq7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4txd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:55Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.102585 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a299b8e0-37bc-48d9-a7e8-64ced2d92d40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvcr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvcr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6wr7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:55Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.119651 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583964bf462fa0005d471f82f6827552a60659ab97984eb1faf074c7b82c2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:55Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.126814 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a299b8e0-37bc-48d9-a7e8-64ced2d92d40-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-6wr7s\" (UID: \"a299b8e0-37bc-48d9-a7e8-64ced2d92d40\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.126850 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a299b8e0-37bc-48d9-a7e8-64ced2d92d40-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-6wr7s\" (UID: \"a299b8e0-37bc-48d9-a7e8-64ced2d92d40\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.126868 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a299b8e0-37bc-48d9-a7e8-64ced2d92d40-env-overrides\") pod \"ovnkube-control-plane-749d76644c-6wr7s\" (UID: \"a299b8e0-37bc-48d9-a7e8-64ced2d92d40\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.126913 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvcr5\" (UniqueName: \"kubernetes.io/projected/a299b8e0-37bc-48d9-a7e8-64ced2d92d40-kube-api-access-rvcr5\") pod \"ovnkube-control-plane-749d76644c-6wr7s\" (UID: \"a299b8e0-37bc-48d9-a7e8-64ced2d92d40\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.128423 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a299b8e0-37bc-48d9-a7e8-64ced2d92d40-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-6wr7s\" (UID: \"a299b8e0-37bc-48d9-a7e8-64ced2d92d40\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.128540 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a299b8e0-37bc-48d9-a7e8-64ced2d92d40-env-overrides\") pod \"ovnkube-control-plane-749d76644c-6wr7s\" (UID: \"a299b8e0-37bc-48d9-a7e8-64ced2d92d40\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.134143 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a299b8e0-37bc-48d9-a7e8-64ced2d92d40-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-6wr7s\" (UID: \"a299b8e0-37bc-48d9-a7e8-64ced2d92d40\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.135563 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:55Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.146105 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvcr5\" (UniqueName: \"kubernetes.io/projected/a299b8e0-37bc-48d9-a7e8-64ced2d92d40-kube-api-access-rvcr5\") pod \"ovnkube-control-plane-749d76644c-6wr7s\" (UID: \"a299b8e0-37bc-48d9-a7e8-64ced2d92d40\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.150993 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7g5s8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a7dc232-835b-491c-afc9-9eb14f257af3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9a7ac317fbb9c589573e80babe484c21bef4da3d9ab79dfb7ebf5b8300c472c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4cs2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7g5s8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:55Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.160590 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.160623 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.160631 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.160646 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.160654 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:55Z","lastTransitionTime":"2025-11-25T15:02:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.174182 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a404958-dffa-4a6a-99fc-3e4950614bd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84364aff9c933d3c3e94ba252568db3ce9296b02f8290b087631173b7500c202\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84364aff9c933d3c3e94ba252568db3ce9296b02f8290b087631173b7500c202\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T15:02:52Z\\\",\\\"message\\\":\\\".io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1125 15:02:52.386129 6142 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1125 15:02:52.389228 6142 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 15:02:52.389548 6142 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1125 15:02:52.389576 6142 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1125 15:02:52.389596 6142 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 15:02:52.389628 6142 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 15:02:52.400291 6142 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 15:02:52.400526 6142 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 15:02:52.400551 6142 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 15:02:52.400574 6142 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 15:02:52.400597 6142 factory.go:656] Stopping watch factory\\\\nI1125 15:02:52.400612 6142 ovnkube.go:599] Stopped ovnkube\\\\nI1125 15:02:52.400642 6142 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 15:02:52.400649 6142 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 15:02:52.400655 6142 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 15:02:5\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zdk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:55Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.188312 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e4f849d-f239-4727-a73e-18327856929a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://490f1595bc77030d2ee789197fa3748324e1b7a0c9d9180ba08383bbac43adc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0902d6f8e56aecc5a2a13a7148f94a8a00ad139027b3c036aba0215ba0a67f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xx2tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:55Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.203420 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd5e648-6322-404d-bf1b-3b59a8092291\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76d277b4001db9d63b0af88b98afcc9019321101429c9be47b3022d3575d4e06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e3a186b14ea199d1aeba5751290c266ec918cb055120cfc5c8a2ae03b1469e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a860a998c40aee3e94e4cfd748f54391b3096fd3ed75c84c7ea2a5a7cefa33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://081a6d623cd043c0919d04e2ca93314b2b85fe33048634aa2637297d54b52cee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f506ec649ce24f4df6fd33e4aeafd13e42167dd71323524a5a3164517659fcc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:55Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.217602 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d5a9ccb-13ba-4dd5-9904-42aac5eca374\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d19bc48acb392051f0f6c5a6820135bf29ebe6d2afd098697553298933cf280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93325a6407389c4a7db0ba3c1a9c84a6980db3c3277cdf3860583a564b562bbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20505e91fc60158eeb3a8cfa0aafc5b4e4436d5def5e735a3a58b0f9f3621058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3281314d85a20945d56b75ab7b7aef1e382a73e74dff03773d118d03eb0ee16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:55Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.231394 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b33b805b76abbcace899e527d5ee3eaeb9e480fb03dc8eacd3f6dd90d2dd6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:55Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.263348 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.263399 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.263411 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.263433 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.263447 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:55Z","lastTransitionTime":"2025-11-25T15:02:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.280636 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s" Nov 25 15:02:55 crc kubenswrapper[4890]: W1125 15:02:55.300065 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda299b8e0_37bc_48d9_a7e8_64ced2d92d40.slice/crio-20cf210fa9aaa376f43fd301cb5c2734dd5d592a561272a27824d2a914582951 WatchSource:0}: Error finding container 20cf210fa9aaa376f43fd301cb5c2734dd5d592a561272a27824d2a914582951: Status 404 returned error can't find the container with id 20cf210fa9aaa376f43fd301cb5c2734dd5d592a561272a27824d2a914582951 Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.366288 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.366335 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.366346 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.366361 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.366370 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:55Z","lastTransitionTime":"2025-11-25T15:02:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.469574 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.469602 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.469610 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.469622 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.469630 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:55Z","lastTransitionTime":"2025-11-25T15:02:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.480370 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s" event={"ID":"a299b8e0-37bc-48d9-a7e8-64ced2d92d40","Type":"ContainerStarted","Data":"20cf210fa9aaa376f43fd301cb5c2734dd5d592a561272a27824d2a914582951"} Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.481611 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7zdk8_9a404958-dffa-4a6a-99fc-3e4950614bd4/ovnkube-controller/1.log" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.481981 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7zdk8_9a404958-dffa-4a6a-99fc-3e4950614bd4/ovnkube-controller/0.log" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.485145 4890 generic.go:334] "Generic (PLEG): container finished" podID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerID="f2f3af2cdb804a47cddb121d619a7545a54af0888b120babca0aa2217ff01d7b" exitCode=1 Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.485186 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" event={"ID":"9a404958-dffa-4a6a-99fc-3e4950614bd4","Type":"ContainerDied","Data":"f2f3af2cdb804a47cddb121d619a7545a54af0888b120babca0aa2217ff01d7b"} Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.485227 4890 scope.go:117] "RemoveContainer" containerID="84364aff9c933d3c3e94ba252568db3ce9296b02f8290b087631173b7500c202" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.485809 4890 scope.go:117] "RemoveContainer" containerID="f2f3af2cdb804a47cddb121d619a7545a54af0888b120babca0aa2217ff01d7b" Nov 25 15:02:55 crc kubenswrapper[4890]: E1125 15:02:55.485948 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7zdk8_openshift-ovn-kubernetes(9a404958-dffa-4a6a-99fc-3e4950614bd4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" podUID="9a404958-dffa-4a6a-99fc-3e4950614bd4" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.499018 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a299b8e0-37bc-48d9-a7e8-64ced2d92d40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvcr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvcr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6wr7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:55Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.521246 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86e031a7-b4c8-4030-a078-cdca7ccc9dde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4c4921f0973b03f6dade29c2ef32684a481cd1c3e1e23b64ed636943c9bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f9285acc7ac3cbb503a1e546728748c42971c8c9019cc8ca837076f3e55558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa240816a6e9d2821006ce240eccf27d8f57a7f0c5394314fa7177488da4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://659004bd47948def6b33498d8d1474f28c2e98986754f61af012267f3984269e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6d2dcd8df64e44df8cdb4c3a022f03fccb39451631ca35b30d4c7c35271ff0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:55Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.542361 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg454" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81411ec7-f865-48e2-b464-ab6213ef5e52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db59b77d323079e0b588a6adc0f247dda84b5497de17264390366124934cdca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a366b314408eb8453ab7f970a243c511a2716b5472d92b03c81061135e4e59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a366b314408eb8453ab7f970a243c511a2716b5472d92b03c81061135e4e59c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f841cf4a886c5e17d9919787305ce297d7910b2692be53169324a56bd545be5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f841cf4a886c5e17d9919787305ce297d7910b2692be53169324a56bd545be5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e7825cb1d51e854b5218b09bca98a38c75b356ac9d3d6c206147e98b050dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21e7825cb1d51e854b5218b09bca98a38c75b356ac9d3d6c206147e98b050dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5881047b9fa64b7504f1d62b697a6f352b05d14eedd61e6b64d5ff99d91edda2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5881047b9fa64b7504f1d62b697a6f352b05d14eedd61e6b64d5ff99d91edda2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg454\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:55Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.559360 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4txd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4322dbd78237bf443a985ce3c79bca4cba2103862d81177f5474a76f4905eed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfcq7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4txd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:55Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.573154 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.573207 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.573216 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.573228 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.573239 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:55Z","lastTransitionTime":"2025-11-25T15:02:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.581322 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a404958-dffa-4a6a-99fc-3e4950614bd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2f3af2cdb804a47cddb121d619a7545a54af0888b120babca0aa2217ff01d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84364aff9c933d3c3e94ba252568db3ce9296b02f8290b087631173b7500c202\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T15:02:52Z\\\",\\\"message\\\":\\\".io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1125 15:02:52.386129 6142 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1125 15:02:52.389228 6142 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 15:02:52.389548 6142 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1125 15:02:52.389576 6142 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1125 15:02:52.389596 6142 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 15:02:52.389628 6142 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 15:02:52.400291 6142 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 15:02:52.400526 6142 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 15:02:52.400551 6142 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 15:02:52.400574 6142 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 15:02:52.400597 6142 factory.go:656] Stopping watch factory\\\\nI1125 15:02:52.400612 6142 ovnkube.go:599] Stopped ovnkube\\\\nI1125 15:02:52.400642 6142 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 15:02:52.400649 6142 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 15:02:52.400655 6142 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 15:02:5\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2f3af2cdb804a47cddb121d619a7545a54af0888b120babca0aa2217ff01d7b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"message\\\":\\\"42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-authentication/oauth-openshift]} name:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 15:02:55.414271 6305 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1125 15:02:55.414314 6305 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 15:02:55.414334 6305 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 15:02:55.414375 6305 factory.go:656] Stopping watch factory\\\\nI1125 15:02:55.414396 6305 ovnkube.go:599] Stopped ovnkube\\\\nI1125 15:02:55.414398 6305 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 15:02:55.414425 6305 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 15:02:55.414443 6305 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 15:02:55.414540 6305 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zdk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:55Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.592092 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e4f849d-f239-4727-a73e-18327856929a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://490f1595bc77030d2ee789197fa3748324e1b7a0c9d9180ba08383bbac43adc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0902d6f8e56aecc5a2a13a7148f94a8a00ad139027b3c036aba0215ba0a67f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xx2tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:55Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.606460 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583964bf462fa0005d471f82f6827552a60659ab97984eb1faf074c7b82c2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:55Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.619057 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:55Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.631281 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7g5s8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a7dc232-835b-491c-afc9-9eb14f257af3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9a7ac317fbb9c589573e80babe484c21bef4da3d9ab79dfb7ebf5b8300c472c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4cs2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7g5s8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:55Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.644271 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b33b805b76abbcace899e527d5ee3eaeb9e480fb03dc8eacd3f6dd90d2dd6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:55Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.662459 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd5e648-6322-404d-bf1b-3b59a8092291\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76d277b4001db9d63b0af88b98afcc9019321101429c9be47b3022d3575d4e06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e3a186b14ea199d1aeba5751290c266ec918cb055120cfc5c8a2ae03b1469e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a860a998c40aee3e94e4cfd748f54391b3096fd3ed75c84c7ea2a5a7cefa33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://081a6d623cd043c0919d04e2ca93314b2b85fe33048634aa2637297d54b52cee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f506ec649ce24f4df6fd33e4aeafd13e42167dd71323524a5a3164517659fcc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:55Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.675071 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.675114 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.675125 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.675141 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.675150 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:55Z","lastTransitionTime":"2025-11-25T15:02:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.682251 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d5a9ccb-13ba-4dd5-9904-42aac5eca374\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d19bc48acb392051f0f6c5a6820135bf29ebe6d2afd098697553298933cf280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93325a6407389c4a7db0ba3c1a9c84a6980db3c3277cdf3860583a564b562bbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20505e91fc60158eeb3a8cfa0aafc5b4e4436d5def5e735a3a58b0f9f3621058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3281314d85a20945d56b75ab7b7aef1e382a73e74dff03773d118d03eb0ee16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:55Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.697047 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3358a356ae5b422d6bba72bc516374e14ccdace80726aa71b10e25102e0f1100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba3acdae961f49f797f66bbebe350e34a4df5882bcf2a4fe8d37f2ca95865c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:55Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.708792 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8tdrh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1097de4-74e4-4178-b9b4-14a22539da9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfa0291da1598e930aeb4a76f9f4e8b0728001a770acd8f10215465539f8d8b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8tdrh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:55Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.725469 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:55Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.732584 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-2nlkp"] Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.733043 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nlkp" Nov 25 15:02:55 crc kubenswrapper[4890]: E1125 15:02:55.733107 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nlkp" podUID="16f52653-c067-4186-bc40-9c0f7c383b64" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.751335 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:55Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.762520 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7g5s8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a7dc232-835b-491c-afc9-9eb14f257af3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9a7ac317fbb9c589573e80babe484c21bef4da3d9ab79dfb7ebf5b8300c472c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4cs2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7g5s8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:55Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.777075 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.777118 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.777127 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.777143 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.777152 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:55Z","lastTransitionTime":"2025-11-25T15:02:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.780845 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a404958-dffa-4a6a-99fc-3e4950614bd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2f3af2cdb804a47cddb121d619a7545a54af0888b120babca0aa2217ff01d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84364aff9c933d3c3e94ba252568db3ce9296b02f8290b087631173b7500c202\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T15:02:52Z\\\",\\\"message\\\":\\\".io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1125 15:02:52.386129 6142 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1125 15:02:52.389228 6142 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 15:02:52.389548 6142 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1125 15:02:52.389576 6142 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1125 15:02:52.389596 6142 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 15:02:52.389628 6142 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 15:02:52.400291 6142 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 15:02:52.400526 6142 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 15:02:52.400551 6142 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 15:02:52.400574 6142 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 15:02:52.400597 6142 factory.go:656] Stopping watch factory\\\\nI1125 15:02:52.400612 6142 ovnkube.go:599] Stopped ovnkube\\\\nI1125 15:02:52.400642 6142 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 15:02:52.400649 6142 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 15:02:52.400655 6142 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 15:02:5\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2f3af2cdb804a47cddb121d619a7545a54af0888b120babca0aa2217ff01d7b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"message\\\":\\\"42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-authentication/oauth-openshift]} name:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 15:02:55.414271 6305 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1125 15:02:55.414314 6305 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 15:02:55.414334 6305 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 15:02:55.414375 6305 factory.go:656] Stopping watch factory\\\\nI1125 15:02:55.414396 6305 ovnkube.go:599] Stopped ovnkube\\\\nI1125 15:02:55.414398 6305 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 15:02:55.414425 6305 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 15:02:55.414443 6305 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 15:02:55.414540 6305 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zdk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:55Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.791080 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e4f849d-f239-4727-a73e-18327856929a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://490f1595bc77030d2ee789197fa3748324e1b7a0c9d9180ba08383bbac43adc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0902d6f8e56aecc5a2a13a7148f94a8a00ad139027b3c036aba0215ba0a67f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xx2tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:55Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.801416 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583964bf462fa0005d471f82f6827552a60659ab97984eb1faf074c7b82c2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:55Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.810971 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:55Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.819983 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b33b805b76abbcace899e527d5ee3eaeb9e480fb03dc8eacd3f6dd90d2dd6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:55Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.829645 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2nlkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16f52653-c067-4186-bc40-9c0f7c383b64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmlnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmlnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2nlkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:55Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.833009 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmlnl\" (UniqueName: \"kubernetes.io/projected/16f52653-c067-4186-bc40-9c0f7c383b64-kube-api-access-kmlnl\") pod \"network-metrics-daemon-2nlkp\" (UID: \"16f52653-c067-4186-bc40-9c0f7c383b64\") " pod="openshift-multus/network-metrics-daemon-2nlkp" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.833207 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/16f52653-c067-4186-bc40-9c0f7c383b64-metrics-certs\") pod \"network-metrics-daemon-2nlkp\" (UID: \"16f52653-c067-4186-bc40-9c0f7c383b64\") " pod="openshift-multus/network-metrics-daemon-2nlkp" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.851880 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd5e648-6322-404d-bf1b-3b59a8092291\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76d277b4001db9d63b0af88b98afcc9019321101429c9be47b3022d3575d4e06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e3a186b14ea199d1aeba5751290c266ec918cb055120cfc5c8a2ae03b1469e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a860a998c40aee3e94e4cfd748f54391b3096fd3ed75c84c7ea2a5a7cefa33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://081a6d623cd043c0919d04e2ca93314b2b85fe33048634aa2637297d54b52cee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f506ec649ce24f4df6fd33e4aeafd13e42167dd71323524a5a3164517659fcc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:55Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.864559 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d5a9ccb-13ba-4dd5-9904-42aac5eca374\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d19bc48acb392051f0f6c5a6820135bf29ebe6d2afd098697553298933cf280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93325a6407389c4a7db0ba3c1a9c84a6980db3c3277cdf3860583a564b562bbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20505e91fc60158eeb3a8cfa0aafc5b4e4436d5def5e735a3a58b0f9f3621058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3281314d85a20945d56b75ab7b7aef1e382a73e74dff03773d118d03eb0ee16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:55Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.875126 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:55Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.882189 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.882231 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.882244 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.882260 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.882273 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:55Z","lastTransitionTime":"2025-11-25T15:02:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.891661 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3358a356ae5b422d6bba72bc516374e14ccdace80726aa71b10e25102e0f1100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba3acdae961f49f797f66bbebe350e34a4df5882bcf2a4fe8d37f2ca95865c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:55Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.900778 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8tdrh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1097de4-74e4-4178-b9b4-14a22539da9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfa0291da1598e930aeb4a76f9f4e8b0728001a770acd8f10215465539f8d8b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8tdrh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:55Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.910876 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:55Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.923979 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4txd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4322dbd78237bf443a985ce3c79bca4cba2103862d81177f5474a76f4905eed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfcq7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4txd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:55Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.934630 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmlnl\" (UniqueName: \"kubernetes.io/projected/16f52653-c067-4186-bc40-9c0f7c383b64-kube-api-access-kmlnl\") pod \"network-metrics-daemon-2nlkp\" (UID: \"16f52653-c067-4186-bc40-9c0f7c383b64\") " pod="openshift-multus/network-metrics-daemon-2nlkp" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.934701 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/16f52653-c067-4186-bc40-9c0f7c383b64-metrics-certs\") pod \"network-metrics-daemon-2nlkp\" (UID: \"16f52653-c067-4186-bc40-9c0f7c383b64\") " pod="openshift-multus/network-metrics-daemon-2nlkp" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.934779 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a299b8e0-37bc-48d9-a7e8-64ced2d92d40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvcr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvcr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6wr7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:55Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:55 crc kubenswrapper[4890]: E1125 15:02:55.934844 4890 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 15:02:55 crc kubenswrapper[4890]: E1125 15:02:55.934893 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/16f52653-c067-4186-bc40-9c0f7c383b64-metrics-certs podName:16f52653-c067-4186-bc40-9c0f7c383b64 nodeName:}" failed. No retries permitted until 2025-11-25 15:02:56.434876846 +0000 UTC m=+34.877339456 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/16f52653-c067-4186-bc40-9c0f7c383b64-metrics-certs") pod "network-metrics-daemon-2nlkp" (UID: "16f52653-c067-4186-bc40-9c0f7c383b64") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.950280 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmlnl\" (UniqueName: \"kubernetes.io/projected/16f52653-c067-4186-bc40-9c0f7c383b64-kube-api-access-kmlnl\") pod \"network-metrics-daemon-2nlkp\" (UID: \"16f52653-c067-4186-bc40-9c0f7c383b64\") " pod="openshift-multus/network-metrics-daemon-2nlkp" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.954432 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86e031a7-b4c8-4030-a078-cdca7ccc9dde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4c4921f0973b03f6dade29c2ef32684a481cd1c3e1e23b64ed636943c9bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f9285acc7ac3cbb503a1e546728748c42971c8c9019cc8ca837076f3e55558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa240816a6e9d2821006ce240eccf27d8f57a7f0c5394314fa7177488da4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://659004bd47948def6b33498d8d1474f28c2e98986754f61af012267f3984269e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6d2dcd8df64e44df8cdb4c3a022f03fccb39451631ca35b30d4c7c35271ff0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:55Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.968225 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg454" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81411ec7-f865-48e2-b464-ab6213ef5e52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db59b77d323079e0b588a6adc0f247dda84b5497de17264390366124934cdca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a366b314408eb8453ab7f970a243c511a2716b5472d92b03c81061135e4e59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a366b314408eb8453ab7f970a243c511a2716b5472d92b03c81061135e4e59c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f841cf4a886c5e17d9919787305ce297d7910b2692be53169324a56bd545be5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f841cf4a886c5e17d9919787305ce297d7910b2692be53169324a56bd545be5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e7825cb1d51e854b5218b09bca98a38c75b356ac9d3d6c206147e98b050dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21e7825cb1d51e854b5218b09bca98a38c75b356ac9d3d6c206147e98b050dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5881047b9fa64b7504f1d62b697a6f352b05d14eedd61e6b64d5ff99d91edda2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5881047b9fa64b7504f1d62b697a6f352b05d14eedd61e6b64d5ff99d91edda2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg454\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:55Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.984760 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.984825 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.984836 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.984850 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:55 crc kubenswrapper[4890]: I1125 15:02:55.984860 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:55Z","lastTransitionTime":"2025-11-25T15:02:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.087028 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.087110 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.087146 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.087179 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.087193 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:56Z","lastTransitionTime":"2025-11-25T15:02:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.173091 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.173132 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.173119 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 15:02:56 crc kubenswrapper[4890]: E1125 15:02:56.173274 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 15:02:56 crc kubenswrapper[4890]: E1125 15:02:56.173457 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 15:02:56 crc kubenswrapper[4890]: E1125 15:02:56.173586 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.189492 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.189560 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.189572 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.189596 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.189613 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:56Z","lastTransitionTime":"2025-11-25T15:02:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.292636 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.292686 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.292698 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.292715 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.292727 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:56Z","lastTransitionTime":"2025-11-25T15:02:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.395225 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.395329 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.395345 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.395369 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.395384 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:56Z","lastTransitionTime":"2025-11-25T15:02:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.440066 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/16f52653-c067-4186-bc40-9c0f7c383b64-metrics-certs\") pod \"network-metrics-daemon-2nlkp\" (UID: \"16f52653-c067-4186-bc40-9c0f7c383b64\") " pod="openshift-multus/network-metrics-daemon-2nlkp" Nov 25 15:02:56 crc kubenswrapper[4890]: E1125 15:02:56.440223 4890 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 15:02:56 crc kubenswrapper[4890]: E1125 15:02:56.440301 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/16f52653-c067-4186-bc40-9c0f7c383b64-metrics-certs podName:16f52653-c067-4186-bc40-9c0f7c383b64 nodeName:}" failed. No retries permitted until 2025-11-25 15:02:57.440288261 +0000 UTC m=+35.882750871 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/16f52653-c067-4186-bc40-9c0f7c383b64-metrics-certs") pod "network-metrics-daemon-2nlkp" (UID: "16f52653-c067-4186-bc40-9c0f7c383b64") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.490407 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s" event={"ID":"a299b8e0-37bc-48d9-a7e8-64ced2d92d40","Type":"ContainerStarted","Data":"90eb11dd969896d6769a18a39ac3c87633b1d42e110093a32bc77bd6f8a7e332"} Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.490471 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s" event={"ID":"a299b8e0-37bc-48d9-a7e8-64ced2d92d40","Type":"ContainerStarted","Data":"c5bf7fec98590eb53dd0ecce057c364d4b943bdc0fd152f49a6dfc9586b056f8"} Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.492596 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7zdk8_9a404958-dffa-4a6a-99fc-3e4950614bd4/ovnkube-controller/1.log" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.497199 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.497272 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.497297 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.497326 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.497349 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:56Z","lastTransitionTime":"2025-11-25T15:02:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.500491 4890 scope.go:117] "RemoveContainer" containerID="f2f3af2cdb804a47cddb121d619a7545a54af0888b120babca0aa2217ff01d7b" Nov 25 15:02:56 crc kubenswrapper[4890]: E1125 15:02:56.500827 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7zdk8_openshift-ovn-kubernetes(9a404958-dffa-4a6a-99fc-3e4950614bd4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" podUID="9a404958-dffa-4a6a-99fc-3e4950614bd4" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.522181 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86e031a7-b4c8-4030-a078-cdca7ccc9dde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4c4921f0973b03f6dade29c2ef32684a481cd1c3e1e23b64ed636943c9bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f9285acc7ac3cbb503a1e546728748c42971c8c9019cc8ca837076f3e55558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa240816a6e9d2821006ce240eccf27d8f57a7f0c5394314fa7177488da4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://659004bd47948def6b33498d8d1474f28c2e98986754f61af012267f3984269e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6d2dcd8df64e44df8cdb4c3a022f03fccb39451631ca35b30d4c7c35271ff0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:56Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.543903 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg454" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81411ec7-f865-48e2-b464-ab6213ef5e52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db59b77d323079e0b588a6adc0f247dda84b5497de17264390366124934cdca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a366b314408eb8453ab7f970a243c511a2716b5472d92b03c81061135e4e59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a366b314408eb8453ab7f970a243c511a2716b5472d92b03c81061135e4e59c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f841cf4a886c5e17d9919787305ce297d7910b2692be53169324a56bd545be5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f841cf4a886c5e17d9919787305ce297d7910b2692be53169324a56bd545be5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e7825cb1d51e854b5218b09bca98a38c75b356ac9d3d6c206147e98b050dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21e7825cb1d51e854b5218b09bca98a38c75b356ac9d3d6c206147e98b050dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5881047b9fa64b7504f1d62b697a6f352b05d14eedd61e6b64d5ff99d91edda2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5881047b9fa64b7504f1d62b697a6f352b05d14eedd61e6b64d5ff99d91edda2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg454\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:56Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.556712 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4txd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4322dbd78237bf443a985ce3c79bca4cba2103862d81177f5474a76f4905eed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfcq7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4txd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:56Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.567804 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a299b8e0-37bc-48d9-a7e8-64ced2d92d40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5bf7fec98590eb53dd0ecce057c364d4b943bdc0fd152f49a6dfc9586b056f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvcr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90eb11dd969896d6769a18a39ac3c87633b1d42e110093a32bc77bd6f8a7e332\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvcr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6wr7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:56Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.580742 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583964bf462fa0005d471f82f6827552a60659ab97984eb1faf074c7b82c2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:56Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.593393 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:56Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.599401 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.599442 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.599454 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.599469 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.599482 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:56Z","lastTransitionTime":"2025-11-25T15:02:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.602980 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7g5s8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a7dc232-835b-491c-afc9-9eb14f257af3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9a7ac317fbb9c589573e80babe484c21bef4da3d9ab79dfb7ebf5b8300c472c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4cs2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7g5s8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:56Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.619534 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a404958-dffa-4a6a-99fc-3e4950614bd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2f3af2cdb804a47cddb121d619a7545a54af0888b120babca0aa2217ff01d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84364aff9c933d3c3e94ba252568db3ce9296b02f8290b087631173b7500c202\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T15:02:52Z\\\",\\\"message\\\":\\\".io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1125 15:02:52.386129 6142 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1125 15:02:52.389228 6142 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 15:02:52.389548 6142 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1125 15:02:52.389576 6142 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1125 15:02:52.389596 6142 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 15:02:52.389628 6142 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 15:02:52.400291 6142 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 15:02:52.400526 6142 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 15:02:52.400551 6142 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 15:02:52.400574 6142 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 15:02:52.400597 6142 factory.go:656] Stopping watch factory\\\\nI1125 15:02:52.400612 6142 ovnkube.go:599] Stopped ovnkube\\\\nI1125 15:02:52.400642 6142 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 15:02:52.400649 6142 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 15:02:52.400655 6142 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 15:02:5\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2f3af2cdb804a47cddb121d619a7545a54af0888b120babca0aa2217ff01d7b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"message\\\":\\\"42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-authentication/oauth-openshift]} name:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 15:02:55.414271 6305 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1125 15:02:55.414314 6305 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 15:02:55.414334 6305 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 15:02:55.414375 6305 factory.go:656] Stopping watch factory\\\\nI1125 15:02:55.414396 6305 ovnkube.go:599] Stopped ovnkube\\\\nI1125 15:02:55.414398 6305 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 15:02:55.414425 6305 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 15:02:55.414443 6305 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 15:02:55.414540 6305 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zdk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:56Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.629086 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e4f849d-f239-4727-a73e-18327856929a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://490f1595bc77030d2ee789197fa3748324e1b7a0c9d9180ba08383bbac43adc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0902d6f8e56aecc5a2a13a7148f94a8a00ad139027b3c036aba0215ba0a67f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xx2tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:56Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.639453 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd5e648-6322-404d-bf1b-3b59a8092291\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76d277b4001db9d63b0af88b98afcc9019321101429c9be47b3022d3575d4e06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e3a186b14ea199d1aeba5751290c266ec918cb055120cfc5c8a2ae03b1469e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a860a998c40aee3e94e4cfd748f54391b3096fd3ed75c84c7ea2a5a7cefa33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://081a6d623cd043c0919d04e2ca93314b2b85fe33048634aa2637297d54b52cee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f506ec649ce24f4df6fd33e4aeafd13e42167dd71323524a5a3164517659fcc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:56Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.650754 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d5a9ccb-13ba-4dd5-9904-42aac5eca374\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d19bc48acb392051f0f6c5a6820135bf29ebe6d2afd098697553298933cf280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93325a6407389c4a7db0ba3c1a9c84a6980db3c3277cdf3860583a564b562bbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20505e91fc60158eeb3a8cfa0aafc5b4e4436d5def5e735a3a58b0f9f3621058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3281314d85a20945d56b75ab7b7aef1e382a73e74dff03773d118d03eb0ee16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:56Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.662906 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b33b805b76abbcace899e527d5ee3eaeb9e480fb03dc8eacd3f6dd90d2dd6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:56Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.673153 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2nlkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16f52653-c067-4186-bc40-9c0f7c383b64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmlnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmlnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2nlkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:56Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.682482 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:56Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.692516 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:56Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.701958 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.702046 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.702060 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.702077 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.702089 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:56Z","lastTransitionTime":"2025-11-25T15:02:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.708691 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3358a356ae5b422d6bba72bc516374e14ccdace80726aa71b10e25102e0f1100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba3acdae961f49f797f66bbebe350e34a4df5882bcf2a4fe8d37f2ca95865c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:56Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.720914 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8tdrh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1097de4-74e4-4178-b9b4-14a22539da9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfa0291da1598e930aeb4a76f9f4e8b0728001a770acd8f10215465539f8d8b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8tdrh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:56Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.731386 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b33b805b76abbcace899e527d5ee3eaeb9e480fb03dc8eacd3f6dd90d2dd6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:56Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.742043 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2nlkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16f52653-c067-4186-bc40-9c0f7c383b64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmlnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmlnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2nlkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:56Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.757990 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd5e648-6322-404d-bf1b-3b59a8092291\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76d277b4001db9d63b0af88b98afcc9019321101429c9be47b3022d3575d4e06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e3a186b14ea199d1aeba5751290c266ec918cb055120cfc5c8a2ae03b1469e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a860a998c40aee3e94e4cfd748f54391b3096fd3ed75c84c7ea2a5a7cefa33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://081a6d623cd043c0919d04e2ca93314b2b85fe33048634aa2637297d54b52cee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f506ec649ce24f4df6fd33e4aeafd13e42167dd71323524a5a3164517659fcc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:56Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.774669 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d5a9ccb-13ba-4dd5-9904-42aac5eca374\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d19bc48acb392051f0f6c5a6820135bf29ebe6d2afd098697553298933cf280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93325a6407389c4a7db0ba3c1a9c84a6980db3c3277cdf3860583a564b562bbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20505e91fc60158eeb3a8cfa0aafc5b4e4436d5def5e735a3a58b0f9f3621058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3281314d85a20945d56b75ab7b7aef1e382a73e74dff03773d118d03eb0ee16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:56Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.788412 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:56Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.803033 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3358a356ae5b422d6bba72bc516374e14ccdace80726aa71b10e25102e0f1100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba3acdae961f49f797f66bbebe350e34a4df5882bcf2a4fe8d37f2ca95865c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:56Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.805643 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.805705 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.805717 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.805740 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.805752 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:56Z","lastTransitionTime":"2025-11-25T15:02:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.814805 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8tdrh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1097de4-74e4-4178-b9b4-14a22539da9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfa0291da1598e930aeb4a76f9f4e8b0728001a770acd8f10215465539f8d8b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8tdrh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:56Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.834125 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:56Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.851722 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4txd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4322dbd78237bf443a985ce3c79bca4cba2103862d81177f5474a76f4905eed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfcq7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4txd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:56Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.866205 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a299b8e0-37bc-48d9-a7e8-64ced2d92d40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5bf7fec98590eb53dd0ecce057c364d4b943bdc0fd152f49a6dfc9586b056f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvcr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90eb11dd969896d6769a18a39ac3c87633b1d42e110093a32bc77bd6f8a7e332\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvcr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6wr7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:56Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.892421 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86e031a7-b4c8-4030-a078-cdca7ccc9dde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4c4921f0973b03f6dade29c2ef32684a481cd1c3e1e23b64ed636943c9bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f9285acc7ac3cbb503a1e546728748c42971c8c9019cc8ca837076f3e55558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa240816a6e9d2821006ce240eccf27d8f57a7f0c5394314fa7177488da4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://659004bd47948def6b33498d8d1474f28c2e98986754f61af012267f3984269e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6d2dcd8df64e44df8cdb4c3a022f03fccb39451631ca35b30d4c7c35271ff0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:56Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.909066 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.909138 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.909177 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.909204 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.909227 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:56Z","lastTransitionTime":"2025-11-25T15:02:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.909325 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg454" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81411ec7-f865-48e2-b464-ab6213ef5e52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db59b77d323079e0b588a6adc0f247dda84b5497de17264390366124934cdca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a366b314408eb8453ab7f970a243c511a2716b5472d92b03c81061135e4e59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a366b314408eb8453ab7f970a243c511a2716b5472d92b03c81061135e4e59c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f841cf4a886c5e17d9919787305ce297d7910b2692be53169324a56bd545be5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f841cf4a886c5e17d9919787305ce297d7910b2692be53169324a56bd545be5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e7825cb1d51e854b5218b09bca98a38c75b356ac9d3d6c206147e98b050dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21e7825cb1d51e854b5218b09bca98a38c75b356ac9d3d6c206147e98b050dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5881047b9fa64b7504f1d62b697a6f352b05d14eedd61e6b64d5ff99d91edda2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5881047b9fa64b7504f1d62b697a6f352b05d14eedd61e6b64d5ff99d91edda2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg454\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:56Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.924678 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7g5s8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a7dc232-835b-491c-afc9-9eb14f257af3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9a7ac317fbb9c589573e80babe484c21bef4da3d9ab79dfb7ebf5b8300c472c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4cs2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7g5s8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:56Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.947540 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a404958-dffa-4a6a-99fc-3e4950614bd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2f3af2cdb804a47cddb121d619a7545a54af0888b120babca0aa2217ff01d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2f3af2cdb804a47cddb121d619a7545a54af0888b120babca0aa2217ff01d7b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"message\\\":\\\"42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-authentication/oauth-openshift]} name:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 15:02:55.414271 6305 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1125 15:02:55.414314 6305 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 15:02:55.414334 6305 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 15:02:55.414375 6305 factory.go:656] Stopping watch factory\\\\nI1125 15:02:55.414396 6305 ovnkube.go:599] Stopped ovnkube\\\\nI1125 15:02:55.414398 6305 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 15:02:55.414425 6305 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 15:02:55.414443 6305 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 15:02:55.414540 6305 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7zdk8_openshift-ovn-kubernetes(9a404958-dffa-4a6a-99fc-3e4950614bd4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zdk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:56Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.961964 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e4f849d-f239-4727-a73e-18327856929a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://490f1595bc77030d2ee789197fa3748324e1b7a0c9d9180ba08383bbac43adc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0902d6f8e56aecc5a2a13a7148f94a8a00ad139027b3c036aba0215ba0a67f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xx2tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:56Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.976295 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583964bf462fa0005d471f82f6827552a60659ab97984eb1faf074c7b82c2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:56Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:56 crc kubenswrapper[4890]: I1125 15:02:56.989016 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:56Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.012639 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.012712 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.012733 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.012760 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.012784 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:57Z","lastTransitionTime":"2025-11-25T15:02:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.116619 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.116737 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.116751 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.116773 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.116789 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:57Z","lastTransitionTime":"2025-11-25T15:02:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.172665 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nlkp" Nov 25 15:02:57 crc kubenswrapper[4890]: E1125 15:02:57.172947 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nlkp" podUID="16f52653-c067-4186-bc40-9c0f7c383b64" Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.220055 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.220123 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.220144 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.220205 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.220237 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:57Z","lastTransitionTime":"2025-11-25T15:02:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.248798 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:02:57 crc kubenswrapper[4890]: E1125 15:02:57.249000 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:03:13.248959913 +0000 UTC m=+51.691422523 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.249074 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.249156 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 15:02:57 crc kubenswrapper[4890]: E1125 15:02:57.249249 4890 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.249273 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.249317 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:02:57 crc kubenswrapper[4890]: E1125 15:02:57.249325 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 15:03:13.249307521 +0000 UTC m=+51.691770131 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 15:02:57 crc kubenswrapper[4890]: E1125 15:02:57.249378 4890 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 15:02:57 crc kubenswrapper[4890]: E1125 15:02:57.249428 4890 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 15:02:57 crc kubenswrapper[4890]: E1125 15:02:57.249442 4890 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 15:02:57 crc kubenswrapper[4890]: E1125 15:02:57.249500 4890 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 15:02:57 crc kubenswrapper[4890]: E1125 15:02:57.249518 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 15:03:13.249500856 +0000 UTC m=+51.691963666 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 15:02:57 crc kubenswrapper[4890]: E1125 15:02:57.249527 4890 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 15:02:57 crc kubenswrapper[4890]: E1125 15:02:57.249445 4890 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 15:02:57 crc kubenswrapper[4890]: E1125 15:02:57.249544 4890 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 15:02:57 crc kubenswrapper[4890]: E1125 15:02:57.249609 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 15:03:13.249595028 +0000 UTC m=+51.692057848 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 15:02:57 crc kubenswrapper[4890]: E1125 15:02:57.249637 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 15:03:13.249624159 +0000 UTC m=+51.692086999 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.323912 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.323968 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.323982 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.324005 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.324021 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:57Z","lastTransitionTime":"2025-11-25T15:02:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.428948 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.429004 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.429070 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.429095 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.429108 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:57Z","lastTransitionTime":"2025-11-25T15:02:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.450983 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/16f52653-c067-4186-bc40-9c0f7c383b64-metrics-certs\") pod \"network-metrics-daemon-2nlkp\" (UID: \"16f52653-c067-4186-bc40-9c0f7c383b64\") " pod="openshift-multus/network-metrics-daemon-2nlkp" Nov 25 15:02:57 crc kubenswrapper[4890]: E1125 15:02:57.451270 4890 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 15:02:57 crc kubenswrapper[4890]: E1125 15:02:57.451429 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/16f52653-c067-4186-bc40-9c0f7c383b64-metrics-certs podName:16f52653-c067-4186-bc40-9c0f7c383b64 nodeName:}" failed. No retries permitted until 2025-11-25 15:02:59.45139534 +0000 UTC m=+37.893857980 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/16f52653-c067-4186-bc40-9c0f7c383b64-metrics-certs") pod "network-metrics-daemon-2nlkp" (UID: "16f52653-c067-4186-bc40-9c0f7c383b64") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.533084 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.533716 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.533728 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.533750 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.533765 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:57Z","lastTransitionTime":"2025-11-25T15:02:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.636790 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.636863 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.636882 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.636916 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.636938 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:57Z","lastTransitionTime":"2025-11-25T15:02:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.739420 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.739462 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.739472 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.739488 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.739501 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:57Z","lastTransitionTime":"2025-11-25T15:02:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.842837 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.842895 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.842908 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.842926 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.843244 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:57Z","lastTransitionTime":"2025-11-25T15:02:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.945934 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.945984 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.946000 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.946017 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.946028 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:57Z","lastTransitionTime":"2025-11-25T15:02:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.985929 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.985978 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.985987 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.986007 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:57 crc kubenswrapper[4890]: I1125 15:02:57.986018 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:57Z","lastTransitionTime":"2025-11-25T15:02:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:57 crc kubenswrapper[4890]: E1125 15:02:57.997852 4890 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:02:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:02:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:02:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:02:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7c7b4592-d206-4ee7-bf93-ada4b8bc755a\\\",\\\"systemUUID\\\":\\\"5350de30-707e-4bde-b990-1fa65ae3a01e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:57Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.002424 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.002486 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.002500 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.002517 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.002528 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:58Z","lastTransitionTime":"2025-11-25T15:02:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:58 crc kubenswrapper[4890]: E1125 15:02:58.018255 4890 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:02:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:02:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:02:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:02:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7c7b4592-d206-4ee7-bf93-ada4b8bc755a\\\",\\\"systemUUID\\\":\\\"5350de30-707e-4bde-b990-1fa65ae3a01e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:58Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.022120 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.022169 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.022178 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.022191 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.022200 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:58Z","lastTransitionTime":"2025-11-25T15:02:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:58 crc kubenswrapper[4890]: E1125 15:02:58.036550 4890 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:02:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:02:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:02:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:02:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7c7b4592-d206-4ee7-bf93-ada4b8bc755a\\\",\\\"systemUUID\\\":\\\"5350de30-707e-4bde-b990-1fa65ae3a01e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:58Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.042491 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.042537 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.042548 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.042569 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.042580 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:58Z","lastTransitionTime":"2025-11-25T15:02:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:58 crc kubenswrapper[4890]: E1125 15:02:58.055043 4890 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:02:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:02:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:02:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:02:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7c7b4592-d206-4ee7-bf93-ada4b8bc755a\\\",\\\"systemUUID\\\":\\\"5350de30-707e-4bde-b990-1fa65ae3a01e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:58Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.059619 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.059662 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.059673 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.059690 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.059702 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:58Z","lastTransitionTime":"2025-11-25T15:02:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:58 crc kubenswrapper[4890]: E1125 15:02:58.072048 4890 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:02:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:02:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:02:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:02:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7c7b4592-d206-4ee7-bf93-ada4b8bc755a\\\",\\\"systemUUID\\\":\\\"5350de30-707e-4bde-b990-1fa65ae3a01e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:02:58Z is after 2025-08-24T17:21:41Z" Nov 25 15:02:58 crc kubenswrapper[4890]: E1125 15:02:58.072202 4890 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.074144 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.074199 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.074212 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.074231 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.074243 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:58Z","lastTransitionTime":"2025-11-25T15:02:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.173042 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.173118 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.173042 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 15:02:58 crc kubenswrapper[4890]: E1125 15:02:58.173297 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 15:02:58 crc kubenswrapper[4890]: E1125 15:02:58.173389 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 15:02:58 crc kubenswrapper[4890]: E1125 15:02:58.173487 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.178110 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.178153 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.178181 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.178197 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.178207 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:58Z","lastTransitionTime":"2025-11-25T15:02:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.280695 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.280741 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.280751 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.280767 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.280778 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:58Z","lastTransitionTime":"2025-11-25T15:02:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.383699 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.383769 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.383779 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.383794 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.383806 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:58Z","lastTransitionTime":"2025-11-25T15:02:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.486702 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.486814 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.486832 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.486857 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.486874 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:58Z","lastTransitionTime":"2025-11-25T15:02:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.588859 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.588902 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.588912 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.588926 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.588936 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:58Z","lastTransitionTime":"2025-11-25T15:02:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.691609 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.691650 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.691661 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.691677 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.691689 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:58Z","lastTransitionTime":"2025-11-25T15:02:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.794236 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.794282 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.794290 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.794305 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.794314 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:58Z","lastTransitionTime":"2025-11-25T15:02:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.897403 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.897439 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.897450 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.897468 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:58 crc kubenswrapper[4890]: I1125 15:02:58.897483 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:58Z","lastTransitionTime":"2025-11-25T15:02:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.000597 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.000645 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.000653 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.000669 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.000681 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:59Z","lastTransitionTime":"2025-11-25T15:02:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.103763 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.103820 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.103829 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.103849 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.103860 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:59Z","lastTransitionTime":"2025-11-25T15:02:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.173055 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nlkp" Nov 25 15:02:59 crc kubenswrapper[4890]: E1125 15:02:59.173277 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nlkp" podUID="16f52653-c067-4186-bc40-9c0f7c383b64" Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.206918 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.206969 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.206980 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.207001 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.207015 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:59Z","lastTransitionTime":"2025-11-25T15:02:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.309396 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.309437 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.309447 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.309462 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.309471 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:59Z","lastTransitionTime":"2025-11-25T15:02:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.412488 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.412535 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.412545 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.412561 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.412572 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:59Z","lastTransitionTime":"2025-11-25T15:02:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.473199 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/16f52653-c067-4186-bc40-9c0f7c383b64-metrics-certs\") pod \"network-metrics-daemon-2nlkp\" (UID: \"16f52653-c067-4186-bc40-9c0f7c383b64\") " pod="openshift-multus/network-metrics-daemon-2nlkp" Nov 25 15:02:59 crc kubenswrapper[4890]: E1125 15:02:59.473383 4890 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 15:02:59 crc kubenswrapper[4890]: E1125 15:02:59.473458 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/16f52653-c067-4186-bc40-9c0f7c383b64-metrics-certs podName:16f52653-c067-4186-bc40-9c0f7c383b64 nodeName:}" failed. No retries permitted until 2025-11-25 15:03:03.473441671 +0000 UTC m=+41.915904271 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/16f52653-c067-4186-bc40-9c0f7c383b64-metrics-certs") pod "network-metrics-daemon-2nlkp" (UID: "16f52653-c067-4186-bc40-9c0f7c383b64") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.517905 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.517958 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.517972 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.517993 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.518006 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:59Z","lastTransitionTime":"2025-11-25T15:02:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.621077 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.621120 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.621129 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.621144 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.621172 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:59Z","lastTransitionTime":"2025-11-25T15:02:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.724205 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.724266 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.724281 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.724301 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.724316 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:59Z","lastTransitionTime":"2025-11-25T15:02:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.826370 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.826399 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.826409 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.826420 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.826429 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:59Z","lastTransitionTime":"2025-11-25T15:02:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.928152 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.928204 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.928213 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.928227 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:02:59 crc kubenswrapper[4890]: I1125 15:02:59.928236 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:02:59Z","lastTransitionTime":"2025-11-25T15:02:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.031725 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.031795 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.031808 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.031835 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.031850 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:00Z","lastTransitionTime":"2025-11-25T15:03:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.134691 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.134765 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.134779 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.134799 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.134811 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:00Z","lastTransitionTime":"2025-11-25T15:03:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.172844 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.172952 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 15:03:00 crc kubenswrapper[4890]: E1125 15:03:00.173115 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.173243 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 15:03:00 crc kubenswrapper[4890]: E1125 15:03:00.173423 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 15:03:00 crc kubenswrapper[4890]: E1125 15:03:00.173597 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.238124 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.238193 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.238203 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.238222 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.238236 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:00Z","lastTransitionTime":"2025-11-25T15:03:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.340778 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.340858 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.340870 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.340889 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.340900 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:00Z","lastTransitionTime":"2025-11-25T15:03:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.443603 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.443648 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.443658 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.443673 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.443685 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:00Z","lastTransitionTime":"2025-11-25T15:03:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.546294 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.546325 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.546333 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.546350 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.546360 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:00Z","lastTransitionTime":"2025-11-25T15:03:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.648492 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.648558 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.648572 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.648594 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.648606 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:00Z","lastTransitionTime":"2025-11-25T15:03:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.752300 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.752359 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.752369 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.752387 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.752398 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:00Z","lastTransitionTime":"2025-11-25T15:03:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.855854 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.855897 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.855907 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.855923 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.855934 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:00Z","lastTransitionTime":"2025-11-25T15:03:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.959411 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.959493 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.959506 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.959524 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:00 crc kubenswrapper[4890]: I1125 15:03:00.959554 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:00Z","lastTransitionTime":"2025-11-25T15:03:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.062716 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.062783 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.062805 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.062836 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.062859 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:01Z","lastTransitionTime":"2025-11-25T15:03:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.166029 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.166123 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.166132 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.166145 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.166153 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:01Z","lastTransitionTime":"2025-11-25T15:03:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.172605 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nlkp" Nov 25 15:03:01 crc kubenswrapper[4890]: E1125 15:03:01.172763 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nlkp" podUID="16f52653-c067-4186-bc40-9c0f7c383b64" Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.269630 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.269703 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.269726 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.269804 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.269832 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:01Z","lastTransitionTime":"2025-11-25T15:03:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.372625 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.372704 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.372722 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.372756 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.372776 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:01Z","lastTransitionTime":"2025-11-25T15:03:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.475944 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.476075 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.476100 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.476219 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.476292 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:01Z","lastTransitionTime":"2025-11-25T15:03:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.580827 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.580904 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.580923 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.580956 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.580974 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:01Z","lastTransitionTime":"2025-11-25T15:03:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.684730 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.684790 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.684801 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.684822 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.684836 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:01Z","lastTransitionTime":"2025-11-25T15:03:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.788150 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.788231 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.788249 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.788276 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.788295 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:01Z","lastTransitionTime":"2025-11-25T15:03:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.892538 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.892650 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.892667 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.892706 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.892722 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:01Z","lastTransitionTime":"2025-11-25T15:03:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.995917 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.995950 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.995959 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.995973 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:01 crc kubenswrapper[4890]: I1125 15:03:01.995983 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:01Z","lastTransitionTime":"2025-11-25T15:03:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.099213 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.099258 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.099268 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.099294 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.099309 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:02Z","lastTransitionTime":"2025-11-25T15:03:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.172659 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.172690 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:03:02 crc kubenswrapper[4890]: E1125 15:03:02.172909 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 15:03:02 crc kubenswrapper[4890]: E1125 15:03:02.172994 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.173076 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 15:03:02 crc kubenswrapper[4890]: E1125 15:03:02.173150 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.202982 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.203054 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.203070 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.203096 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.203116 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:02Z","lastTransitionTime":"2025-11-25T15:03:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.210430 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86e031a7-b4c8-4030-a078-cdca7ccc9dde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4c4921f0973b03f6dade29c2ef32684a481cd1c3e1e23b64ed636943c9bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f9285acc7ac3cbb503a1e546728748c42971c8c9019cc8ca837076f3e55558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa240816a6e9d2821006ce240eccf27d8f57a7f0c5394314fa7177488da4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://659004bd47948def6b33498d8d1474f28c2e98986754f61af012267f3984269e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6d2dcd8df64e44df8cdb4c3a022f03fccb39451631ca35b30d4c7c35271ff0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:02Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.236823 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg454" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81411ec7-f865-48e2-b464-ab6213ef5e52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db59b77d323079e0b588a6adc0f247dda84b5497de17264390366124934cdca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a366b314408eb8453ab7f970a243c511a2716b5472d92b03c81061135e4e59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a366b314408eb8453ab7f970a243c511a2716b5472d92b03c81061135e4e59c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f841cf4a886c5e17d9919787305ce297d7910b2692be53169324a56bd545be5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f841cf4a886c5e17d9919787305ce297d7910b2692be53169324a56bd545be5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e7825cb1d51e854b5218b09bca98a38c75b356ac9d3d6c206147e98b050dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21e7825cb1d51e854b5218b09bca98a38c75b356ac9d3d6c206147e98b050dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5881047b9fa64b7504f1d62b697a6f352b05d14eedd61e6b64d5ff99d91edda2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5881047b9fa64b7504f1d62b697a6f352b05d14eedd61e6b64d5ff99d91edda2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg454\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:02Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.261270 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4txd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4322dbd78237bf443a985ce3c79bca4cba2103862d81177f5474a76f4905eed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfcq7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4txd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:02Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.280140 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a299b8e0-37bc-48d9-a7e8-64ced2d92d40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5bf7fec98590eb53dd0ecce057c364d4b943bdc0fd152f49a6dfc9586b056f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvcr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90eb11dd969896d6769a18a39ac3c87633b1d42e110093a32bc77bd6f8a7e332\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvcr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6wr7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:02Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.296509 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e4f849d-f239-4727-a73e-18327856929a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://490f1595bc77030d2ee789197fa3748324e1b7a0c9d9180ba08383bbac43adc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0902d6f8e56aecc5a2a13a7148f94a8a00ad139027b3c036aba0215ba0a67f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xx2tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:02Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.306380 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.306524 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.306618 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.306704 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.306726 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:02Z","lastTransitionTime":"2025-11-25T15:03:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.310369 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583964bf462fa0005d471f82f6827552a60659ab97984eb1faf074c7b82c2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:02Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.331659 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:02Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.343645 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7g5s8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a7dc232-835b-491c-afc9-9eb14f257af3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9a7ac317fbb9c589573e80babe484c21bef4da3d9ab79dfb7ebf5b8300c472c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4cs2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7g5s8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:02Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.365723 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a404958-dffa-4a6a-99fc-3e4950614bd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2f3af2cdb804a47cddb121d619a7545a54af0888b120babca0aa2217ff01d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2f3af2cdb804a47cddb121d619a7545a54af0888b120babca0aa2217ff01d7b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"message\\\":\\\"42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-authentication/oauth-openshift]} name:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 15:02:55.414271 6305 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1125 15:02:55.414314 6305 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 15:02:55.414334 6305 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 15:02:55.414375 6305 factory.go:656] Stopping watch factory\\\\nI1125 15:02:55.414396 6305 ovnkube.go:599] Stopped ovnkube\\\\nI1125 15:02:55.414398 6305 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 15:02:55.414425 6305 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 15:02:55.414443 6305 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 15:02:55.414540 6305 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7zdk8_openshift-ovn-kubernetes(9a404958-dffa-4a6a-99fc-3e4950614bd4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zdk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:02Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.380842 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2nlkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16f52653-c067-4186-bc40-9c0f7c383b64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmlnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmlnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2nlkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:02Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.395448 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd5e648-6322-404d-bf1b-3b59a8092291\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76d277b4001db9d63b0af88b98afcc9019321101429c9be47b3022d3575d4e06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e3a186b14ea199d1aeba5751290c266ec918cb055120cfc5c8a2ae03b1469e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a860a998c40aee3e94e4cfd748f54391b3096fd3ed75c84c7ea2a5a7cefa33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://081a6d623cd043c0919d04e2ca93314b2b85fe33048634aa2637297d54b52cee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f506ec649ce24f4df6fd33e4aeafd13e42167dd71323524a5a3164517659fcc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:02Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.407679 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d5a9ccb-13ba-4dd5-9904-42aac5eca374\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d19bc48acb392051f0f6c5a6820135bf29ebe6d2afd098697553298933cf280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93325a6407389c4a7db0ba3c1a9c84a6980db3c3277cdf3860583a564b562bbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20505e91fc60158eeb3a8cfa0aafc5b4e4436d5def5e735a3a58b0f9f3621058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3281314d85a20945d56b75ab7b7aef1e382a73e74dff03773d118d03eb0ee16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:02Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.410445 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.410495 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.410508 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.410529 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.410541 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:02Z","lastTransitionTime":"2025-11-25T15:03:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.421928 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b33b805b76abbcace899e527d5ee3eaeb9e480fb03dc8eacd3f6dd90d2dd6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:02Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.437368 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8tdrh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1097de4-74e4-4178-b9b4-14a22539da9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfa0291da1598e930aeb4a76f9f4e8b0728001a770acd8f10215465539f8d8b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8tdrh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:02Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.458848 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:02Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.475497 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:02Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.492845 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3358a356ae5b422d6bba72bc516374e14ccdace80726aa71b10e25102e0f1100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba3acdae961f49f797f66bbebe350e34a4df5882bcf2a4fe8d37f2ca95865c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:02Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.518977 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.519016 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.519025 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.519042 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.519054 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:02Z","lastTransitionTime":"2025-11-25T15:03:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.621776 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.621807 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.621815 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.621827 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.621836 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:02Z","lastTransitionTime":"2025-11-25T15:03:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.723807 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.723846 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.723855 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.723870 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.723881 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:02Z","lastTransitionTime":"2025-11-25T15:03:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.826095 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.826145 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.826184 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.826203 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.826218 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:02Z","lastTransitionTime":"2025-11-25T15:03:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.929371 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.929417 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.929425 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.929440 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:02 crc kubenswrapper[4890]: I1125 15:03:02.929449 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:02Z","lastTransitionTime":"2025-11-25T15:03:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:03 crc kubenswrapper[4890]: I1125 15:03:03.033028 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:03 crc kubenswrapper[4890]: I1125 15:03:03.033078 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:03 crc kubenswrapper[4890]: I1125 15:03:03.033090 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:03 crc kubenswrapper[4890]: I1125 15:03:03.033113 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:03 crc kubenswrapper[4890]: I1125 15:03:03.033126 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:03Z","lastTransitionTime":"2025-11-25T15:03:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:03 crc kubenswrapper[4890]: I1125 15:03:03.136208 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:03 crc kubenswrapper[4890]: I1125 15:03:03.136279 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:03 crc kubenswrapper[4890]: I1125 15:03:03.136293 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:03 crc kubenswrapper[4890]: I1125 15:03:03.136318 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:03 crc kubenswrapper[4890]: I1125 15:03:03.136336 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:03Z","lastTransitionTime":"2025-11-25T15:03:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:03 crc kubenswrapper[4890]: I1125 15:03:03.172958 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nlkp" Nov 25 15:03:03 crc kubenswrapper[4890]: E1125 15:03:03.173201 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nlkp" podUID="16f52653-c067-4186-bc40-9c0f7c383b64" Nov 25 15:03:03 crc kubenswrapper[4890]: I1125 15:03:03.239983 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:03 crc kubenswrapper[4890]: I1125 15:03:03.240050 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:03 crc kubenswrapper[4890]: I1125 15:03:03.240068 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:03 crc kubenswrapper[4890]: I1125 15:03:03.240089 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:03 crc kubenswrapper[4890]: I1125 15:03:03.240101 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:03Z","lastTransitionTime":"2025-11-25T15:03:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:03 crc kubenswrapper[4890]: I1125 15:03:03.343016 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:03 crc kubenswrapper[4890]: I1125 15:03:03.343086 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:03 crc kubenswrapper[4890]: I1125 15:03:03.343104 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:03 crc kubenswrapper[4890]: I1125 15:03:03.343133 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:03 crc kubenswrapper[4890]: I1125 15:03:03.343149 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:03Z","lastTransitionTime":"2025-11-25T15:03:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.061961 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/16f52653-c067-4186-bc40-9c0f7c383b64-metrics-certs\") pod \"network-metrics-daemon-2nlkp\" (UID: \"16f52653-c067-4186-bc40-9c0f7c383b64\") " pod="openshift-multus/network-metrics-daemon-2nlkp" Nov 25 15:03:04 crc kubenswrapper[4890]: E1125 15:03:04.062312 4890 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 15:03:04 crc kubenswrapper[4890]: E1125 15:03:04.062360 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/16f52653-c067-4186-bc40-9c0f7c383b64-metrics-certs podName:16f52653-c067-4186-bc40-9c0f7c383b64 nodeName:}" failed. No retries permitted until 2025-11-25 15:03:12.062346126 +0000 UTC m=+50.504808736 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/16f52653-c067-4186-bc40-9c0f7c383b64-metrics-certs") pod "network-metrics-daemon-2nlkp" (UID: "16f52653-c067-4186-bc40-9c0f7c383b64") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.062508 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.064264 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.064422 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.064536 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.064832 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.065073 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:04Z","lastTransitionTime":"2025-11-25T15:03:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.065179 4890 scope.go:117] "RemoveContainer" containerID="f2f3af2cdb804a47cddb121d619a7545a54af0888b120babca0aa2217ff01d7b" Nov 25 15:03:04 crc kubenswrapper[4890]: E1125 15:03:04.066233 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7zdk8_openshift-ovn-kubernetes(9a404958-dffa-4a6a-99fc-3e4950614bd4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" podUID="9a404958-dffa-4a6a-99fc-3e4950614bd4" Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.167646 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.167687 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.167696 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.167715 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.167727 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:04Z","lastTransitionTime":"2025-11-25T15:03:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.174710 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 15:03:04 crc kubenswrapper[4890]: E1125 15:03:04.174831 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.174892 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:03:04 crc kubenswrapper[4890]: E1125 15:03:04.174938 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.174980 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 15:03:04 crc kubenswrapper[4890]: E1125 15:03:04.175019 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.270846 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.270898 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.270909 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.270927 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.270940 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:04Z","lastTransitionTime":"2025-11-25T15:03:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.374050 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.374131 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.374143 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.374197 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.374212 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:04Z","lastTransitionTime":"2025-11-25T15:03:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.477127 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.477187 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.477199 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.477216 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.477229 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:04Z","lastTransitionTime":"2025-11-25T15:03:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.579996 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.580039 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.580048 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.580061 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.580071 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:04Z","lastTransitionTime":"2025-11-25T15:03:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.684344 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.684408 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.684418 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.684436 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.684447 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:04Z","lastTransitionTime":"2025-11-25T15:03:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.786893 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.786938 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.786948 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.786963 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.786976 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:04Z","lastTransitionTime":"2025-11-25T15:03:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.890227 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.890303 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.890316 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.890340 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.890356 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:04Z","lastTransitionTime":"2025-11-25T15:03:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.993121 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.993218 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.993234 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.993253 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:04 crc kubenswrapper[4890]: I1125 15:03:04.993263 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:04Z","lastTransitionTime":"2025-11-25T15:03:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:05 crc kubenswrapper[4890]: I1125 15:03:05.096583 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:05 crc kubenswrapper[4890]: I1125 15:03:05.096642 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:05 crc kubenswrapper[4890]: I1125 15:03:05.096655 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:05 crc kubenswrapper[4890]: I1125 15:03:05.096681 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:05 crc kubenswrapper[4890]: I1125 15:03:05.096694 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:05Z","lastTransitionTime":"2025-11-25T15:03:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:05 crc kubenswrapper[4890]: I1125 15:03:05.173137 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nlkp" Nov 25 15:03:05 crc kubenswrapper[4890]: E1125 15:03:05.173411 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nlkp" podUID="16f52653-c067-4186-bc40-9c0f7c383b64" Nov 25 15:03:05 crc kubenswrapper[4890]: I1125 15:03:05.199468 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:05 crc kubenswrapper[4890]: I1125 15:03:05.199550 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:05 crc kubenswrapper[4890]: I1125 15:03:05.199560 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:05 crc kubenswrapper[4890]: I1125 15:03:05.199582 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:05 crc kubenswrapper[4890]: I1125 15:03:05.199602 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:05Z","lastTransitionTime":"2025-11-25T15:03:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:05 crc kubenswrapper[4890]: I1125 15:03:05.302731 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:05 crc kubenswrapper[4890]: I1125 15:03:05.302775 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:05 crc kubenswrapper[4890]: I1125 15:03:05.302783 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:05 crc kubenswrapper[4890]: I1125 15:03:05.302798 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:05 crc kubenswrapper[4890]: I1125 15:03:05.302808 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:05Z","lastTransitionTime":"2025-11-25T15:03:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:05 crc kubenswrapper[4890]: I1125 15:03:05.408196 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:05 crc kubenswrapper[4890]: I1125 15:03:05.408653 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:05 crc kubenswrapper[4890]: I1125 15:03:05.408738 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:05 crc kubenswrapper[4890]: I1125 15:03:05.408834 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:05 crc kubenswrapper[4890]: I1125 15:03:05.408907 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:05Z","lastTransitionTime":"2025-11-25T15:03:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:05 crc kubenswrapper[4890]: I1125 15:03:05.512432 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:05 crc kubenswrapper[4890]: I1125 15:03:05.512494 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:05 crc kubenswrapper[4890]: I1125 15:03:05.512511 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:05 crc kubenswrapper[4890]: I1125 15:03:05.512536 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:05 crc kubenswrapper[4890]: I1125 15:03:05.512552 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:05Z","lastTransitionTime":"2025-11-25T15:03:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:05 crc kubenswrapper[4890]: I1125 15:03:05.615830 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:05 crc kubenswrapper[4890]: I1125 15:03:05.615878 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:05 crc kubenswrapper[4890]: I1125 15:03:05.615887 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:05 crc kubenswrapper[4890]: I1125 15:03:05.615904 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:05 crc kubenswrapper[4890]: I1125 15:03:05.615914 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:05Z","lastTransitionTime":"2025-11-25T15:03:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:05 crc kubenswrapper[4890]: I1125 15:03:05.719527 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:05 crc kubenswrapper[4890]: I1125 15:03:05.719601 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:05 crc kubenswrapper[4890]: I1125 15:03:05.719628 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:05 crc kubenswrapper[4890]: I1125 15:03:05.719657 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:05 crc kubenswrapper[4890]: I1125 15:03:05.719680 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:05Z","lastTransitionTime":"2025-11-25T15:03:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:05 crc kubenswrapper[4890]: I1125 15:03:05.822900 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:05 crc kubenswrapper[4890]: I1125 15:03:05.822942 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:05 crc kubenswrapper[4890]: I1125 15:03:05.822952 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:05 crc kubenswrapper[4890]: I1125 15:03:05.822965 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:05 crc kubenswrapper[4890]: I1125 15:03:05.822975 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:05Z","lastTransitionTime":"2025-11-25T15:03:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:05 crc kubenswrapper[4890]: I1125 15:03:05.926976 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:05 crc kubenswrapper[4890]: I1125 15:03:05.927071 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:05 crc kubenswrapper[4890]: I1125 15:03:05.927229 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:05 crc kubenswrapper[4890]: I1125 15:03:05.927278 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:05 crc kubenswrapper[4890]: I1125 15:03:05.927308 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:05Z","lastTransitionTime":"2025-11-25T15:03:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.031074 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.031130 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.031142 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.031177 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.031189 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:06Z","lastTransitionTime":"2025-11-25T15:03:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.134397 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.134444 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.134453 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.134471 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.134482 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:06Z","lastTransitionTime":"2025-11-25T15:03:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.173067 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.173313 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:03:06 crc kubenswrapper[4890]: E1125 15:03:06.173500 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.173547 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 15:03:06 crc kubenswrapper[4890]: E1125 15:03:06.173631 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 15:03:06 crc kubenswrapper[4890]: E1125 15:03:06.173684 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.236758 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.236817 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.236829 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.237030 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.237043 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:06Z","lastTransitionTime":"2025-11-25T15:03:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.339447 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.339519 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.339529 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.339542 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.339550 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:06Z","lastTransitionTime":"2025-11-25T15:03:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.441392 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.441425 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.441433 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.441447 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.441457 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:06Z","lastTransitionTime":"2025-11-25T15:03:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.544479 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.544582 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.544604 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.544625 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.544636 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:06Z","lastTransitionTime":"2025-11-25T15:03:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.648292 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.648363 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.648385 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.648414 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.648430 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:06Z","lastTransitionTime":"2025-11-25T15:03:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.750926 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.751011 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.751035 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.751062 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.751081 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:06Z","lastTransitionTime":"2025-11-25T15:03:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.853598 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.853643 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.853655 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.853671 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.853682 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:06Z","lastTransitionTime":"2025-11-25T15:03:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.957527 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.957573 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.957587 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.957605 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:06 crc kubenswrapper[4890]: I1125 15:03:06.957620 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:06Z","lastTransitionTime":"2025-11-25T15:03:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.061216 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.061288 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.061317 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.061351 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.061376 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:07Z","lastTransitionTime":"2025-11-25T15:03:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.164901 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.164952 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.164962 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.164979 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.164989 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:07Z","lastTransitionTime":"2025-11-25T15:03:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.172379 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nlkp" Nov 25 15:03:07 crc kubenswrapper[4890]: E1125 15:03:07.172598 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nlkp" podUID="16f52653-c067-4186-bc40-9c0f7c383b64" Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.267526 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.267582 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.267592 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.267611 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.267622 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:07Z","lastTransitionTime":"2025-11-25T15:03:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.371155 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.371260 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.371278 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.371309 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.371328 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:07Z","lastTransitionTime":"2025-11-25T15:03:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.474970 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.475035 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.475047 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.475068 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.475080 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:07Z","lastTransitionTime":"2025-11-25T15:03:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.577673 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.577731 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.577750 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.577777 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.577797 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:07Z","lastTransitionTime":"2025-11-25T15:03:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.681536 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.681619 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.681636 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.681663 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.681686 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:07Z","lastTransitionTime":"2025-11-25T15:03:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.784570 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.784642 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.784659 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.784690 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.784709 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:07Z","lastTransitionTime":"2025-11-25T15:03:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.888510 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.888590 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.888615 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.888651 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.888678 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:07Z","lastTransitionTime":"2025-11-25T15:03:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.992574 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.992625 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.992640 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.992660 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:07 crc kubenswrapper[4890]: I1125 15:03:07.992672 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:07Z","lastTransitionTime":"2025-11-25T15:03:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.095876 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.095946 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.096618 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.096687 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.096718 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:08Z","lastTransitionTime":"2025-11-25T15:03:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.172595 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.172648 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.172749 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 15:03:08 crc kubenswrapper[4890]: E1125 15:03:08.172788 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 15:03:08 crc kubenswrapper[4890]: E1125 15:03:08.172961 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 15:03:08 crc kubenswrapper[4890]: E1125 15:03:08.173017 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.200878 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.200937 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.200952 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.200970 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.200985 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:08Z","lastTransitionTime":"2025-11-25T15:03:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.304090 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.304148 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.304176 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.304196 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.304208 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:08Z","lastTransitionTime":"2025-11-25T15:03:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.391683 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.391745 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.391756 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.391782 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.391795 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:08Z","lastTransitionTime":"2025-11-25T15:03:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:08 crc kubenswrapper[4890]: E1125 15:03:08.406823 4890 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7c7b4592-d206-4ee7-bf93-ada4b8bc755a\\\",\\\"systemUUID\\\":\\\"5350de30-707e-4bde-b990-1fa65ae3a01e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:08Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.411840 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.411869 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.411877 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.411894 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.411904 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:08Z","lastTransitionTime":"2025-11-25T15:03:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:08 crc kubenswrapper[4890]: E1125 15:03:08.427660 4890 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7c7b4592-d206-4ee7-bf93-ada4b8bc755a\\\",\\\"systemUUID\\\":\\\"5350de30-707e-4bde-b990-1fa65ae3a01e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:08Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.431991 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.432027 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.432036 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.432057 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.432069 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:08Z","lastTransitionTime":"2025-11-25T15:03:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:08 crc kubenswrapper[4890]: E1125 15:03:08.445936 4890 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7c7b4592-d206-4ee7-bf93-ada4b8bc755a\\\",\\\"systemUUID\\\":\\\"5350de30-707e-4bde-b990-1fa65ae3a01e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:08Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.450472 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.450500 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.450509 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.450532 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.450543 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:08Z","lastTransitionTime":"2025-11-25T15:03:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:08 crc kubenswrapper[4890]: E1125 15:03:08.464573 4890 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7c7b4592-d206-4ee7-bf93-ada4b8bc755a\\\",\\\"systemUUID\\\":\\\"5350de30-707e-4bde-b990-1fa65ae3a01e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:08Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.469427 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.469453 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.469461 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.469479 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.469491 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:08Z","lastTransitionTime":"2025-11-25T15:03:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:08 crc kubenswrapper[4890]: E1125 15:03:08.484873 4890 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7c7b4592-d206-4ee7-bf93-ada4b8bc755a\\\",\\\"systemUUID\\\":\\\"5350de30-707e-4bde-b990-1fa65ae3a01e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:08Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:08 crc kubenswrapper[4890]: E1125 15:03:08.485014 4890 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.486885 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.486933 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.486947 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.486963 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.486975 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:08Z","lastTransitionTime":"2025-11-25T15:03:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.590922 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.590980 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.590995 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.591017 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.591033 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:08Z","lastTransitionTime":"2025-11-25T15:03:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.694669 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.694728 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.694744 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.694774 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.694788 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:08Z","lastTransitionTime":"2025-11-25T15:03:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.797773 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.797837 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.797851 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.797889 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.797903 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:08Z","lastTransitionTime":"2025-11-25T15:03:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.901311 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.901393 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.901411 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.901435 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:08 crc kubenswrapper[4890]: I1125 15:03:08.901449 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:08Z","lastTransitionTime":"2025-11-25T15:03:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.003960 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.003999 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.004009 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.004027 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.004038 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:09Z","lastTransitionTime":"2025-11-25T15:03:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.107143 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.107196 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.107207 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.107225 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.107236 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:09Z","lastTransitionTime":"2025-11-25T15:03:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.172581 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nlkp" Nov 25 15:03:09 crc kubenswrapper[4890]: E1125 15:03:09.172832 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nlkp" podUID="16f52653-c067-4186-bc40-9c0f7c383b64" Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.212006 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.212090 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.212117 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.212153 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.212217 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:09Z","lastTransitionTime":"2025-11-25T15:03:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.316375 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.316450 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.316478 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.316514 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.316534 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:09Z","lastTransitionTime":"2025-11-25T15:03:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.419819 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.419884 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.419904 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.419935 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.419957 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:09Z","lastTransitionTime":"2025-11-25T15:03:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.523139 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.523239 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.523257 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.523286 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.523306 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:09Z","lastTransitionTime":"2025-11-25T15:03:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.627016 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.627125 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.627152 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.627240 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.627274 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:09Z","lastTransitionTime":"2025-11-25T15:03:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.730870 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.730922 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.730934 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.730967 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.731041 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:09Z","lastTransitionTime":"2025-11-25T15:03:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.835113 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.835183 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.835195 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.835214 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.835227 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:09Z","lastTransitionTime":"2025-11-25T15:03:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.938954 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.939018 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.939031 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.939049 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:09 crc kubenswrapper[4890]: I1125 15:03:09.939060 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:09Z","lastTransitionTime":"2025-11-25T15:03:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.042706 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.042759 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.042769 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.042791 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.042805 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:10Z","lastTransitionTime":"2025-11-25T15:03:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.146000 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.146052 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.146062 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.146090 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.146112 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:10Z","lastTransitionTime":"2025-11-25T15:03:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.172684 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.172780 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.172825 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:03:10 crc kubenswrapper[4890]: E1125 15:03:10.172941 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 15:03:10 crc kubenswrapper[4890]: E1125 15:03:10.173047 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 15:03:10 crc kubenswrapper[4890]: E1125 15:03:10.173135 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.249032 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.249215 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.249238 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.249265 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.249286 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:10Z","lastTransitionTime":"2025-11-25T15:03:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.352453 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.352539 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.352566 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.352600 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.352631 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:10Z","lastTransitionTime":"2025-11-25T15:03:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.455941 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.456009 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.456024 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.456050 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.456061 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:10Z","lastTransitionTime":"2025-11-25T15:03:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.559144 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.559215 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.559225 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.559243 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.559255 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:10Z","lastTransitionTime":"2025-11-25T15:03:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.662644 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.662732 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.662742 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.662760 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.662772 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:10Z","lastTransitionTime":"2025-11-25T15:03:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.765248 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.765286 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.765295 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.765311 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.765322 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:10Z","lastTransitionTime":"2025-11-25T15:03:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.868503 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.868547 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.868557 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.868574 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.868588 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:10Z","lastTransitionTime":"2025-11-25T15:03:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.971602 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.971661 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.971674 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.971698 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:10 crc kubenswrapper[4890]: I1125 15:03:10.971714 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:10Z","lastTransitionTime":"2025-11-25T15:03:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:11 crc kubenswrapper[4890]: I1125 15:03:11.074573 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:11 crc kubenswrapper[4890]: I1125 15:03:11.074623 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:11 crc kubenswrapper[4890]: I1125 15:03:11.074634 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:11 crc kubenswrapper[4890]: I1125 15:03:11.074651 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:11 crc kubenswrapper[4890]: I1125 15:03:11.074666 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:11Z","lastTransitionTime":"2025-11-25T15:03:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:11 crc kubenswrapper[4890]: I1125 15:03:11.173271 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nlkp" Nov 25 15:03:11 crc kubenswrapper[4890]: E1125 15:03:11.173474 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nlkp" podUID="16f52653-c067-4186-bc40-9c0f7c383b64" Nov 25 15:03:11 crc kubenswrapper[4890]: I1125 15:03:11.177858 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:11 crc kubenswrapper[4890]: I1125 15:03:11.177918 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:11 crc kubenswrapper[4890]: I1125 15:03:11.177937 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:11 crc kubenswrapper[4890]: I1125 15:03:11.177959 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:11 crc kubenswrapper[4890]: I1125 15:03:11.177978 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:11Z","lastTransitionTime":"2025-11-25T15:03:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:11 crc kubenswrapper[4890]: I1125 15:03:11.280727 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:11 crc kubenswrapper[4890]: I1125 15:03:11.280767 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:11 crc kubenswrapper[4890]: I1125 15:03:11.280778 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:11 crc kubenswrapper[4890]: I1125 15:03:11.280793 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:11 crc kubenswrapper[4890]: I1125 15:03:11.280805 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:11Z","lastTransitionTime":"2025-11-25T15:03:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:11 crc kubenswrapper[4890]: I1125 15:03:11.383525 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:11 crc kubenswrapper[4890]: I1125 15:03:11.383559 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:11 crc kubenswrapper[4890]: I1125 15:03:11.383569 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:11 crc kubenswrapper[4890]: I1125 15:03:11.383584 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:11 crc kubenswrapper[4890]: I1125 15:03:11.383594 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:11Z","lastTransitionTime":"2025-11-25T15:03:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:11 crc kubenswrapper[4890]: I1125 15:03:11.485721 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:11 crc kubenswrapper[4890]: I1125 15:03:11.485757 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:11 crc kubenswrapper[4890]: I1125 15:03:11.485765 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:11 crc kubenswrapper[4890]: I1125 15:03:11.485779 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:11 crc kubenswrapper[4890]: I1125 15:03:11.485788 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:11Z","lastTransitionTime":"2025-11-25T15:03:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:11 crc kubenswrapper[4890]: I1125 15:03:11.588144 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:11 crc kubenswrapper[4890]: I1125 15:03:11.588210 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:11 crc kubenswrapper[4890]: I1125 15:03:11.588227 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:11 crc kubenswrapper[4890]: I1125 15:03:11.588248 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:11 crc kubenswrapper[4890]: I1125 15:03:11.588267 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:11Z","lastTransitionTime":"2025-11-25T15:03:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:11 crc kubenswrapper[4890]: I1125 15:03:11.691311 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:11 crc kubenswrapper[4890]: I1125 15:03:11.691364 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:11 crc kubenswrapper[4890]: I1125 15:03:11.691380 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:11 crc kubenswrapper[4890]: I1125 15:03:11.691397 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:11 crc kubenswrapper[4890]: I1125 15:03:11.691412 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:11Z","lastTransitionTime":"2025-11-25T15:03:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:11 crc kubenswrapper[4890]: I1125 15:03:11.794808 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:11 crc kubenswrapper[4890]: I1125 15:03:11.794880 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:11 crc kubenswrapper[4890]: I1125 15:03:11.795569 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:11 crc kubenswrapper[4890]: I1125 15:03:11.795607 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:11 crc kubenswrapper[4890]: I1125 15:03:11.795622 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:11Z","lastTransitionTime":"2025-11-25T15:03:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:11 crc kubenswrapper[4890]: I1125 15:03:11.898208 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:11 crc kubenswrapper[4890]: I1125 15:03:11.898303 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:11 crc kubenswrapper[4890]: I1125 15:03:11.898327 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:11 crc kubenswrapper[4890]: I1125 15:03:11.898387 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:11 crc kubenswrapper[4890]: I1125 15:03:11.898426 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:11Z","lastTransitionTime":"2025-11-25T15:03:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.001025 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.001071 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.001080 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.001097 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.001108 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:12Z","lastTransitionTime":"2025-11-25T15:03:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.103388 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.103427 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.103437 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.103452 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.103463 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:12Z","lastTransitionTime":"2025-11-25T15:03:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.155243 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/16f52653-c067-4186-bc40-9c0f7c383b64-metrics-certs\") pod \"network-metrics-daemon-2nlkp\" (UID: \"16f52653-c067-4186-bc40-9c0f7c383b64\") " pod="openshift-multus/network-metrics-daemon-2nlkp" Nov 25 15:03:12 crc kubenswrapper[4890]: E1125 15:03:12.155465 4890 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 15:03:12 crc kubenswrapper[4890]: E1125 15:03:12.155588 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/16f52653-c067-4186-bc40-9c0f7c383b64-metrics-certs podName:16f52653-c067-4186-bc40-9c0f7c383b64 nodeName:}" failed. No retries permitted until 2025-11-25 15:03:28.155566444 +0000 UTC m=+66.598029054 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/16f52653-c067-4186-bc40-9c0f7c383b64-metrics-certs") pod "network-metrics-daemon-2nlkp" (UID: "16f52653-c067-4186-bc40-9c0f7c383b64") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.172258 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.172763 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 15:03:12 crc kubenswrapper[4890]: E1125 15:03:12.173049 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.173148 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 15:03:12 crc kubenswrapper[4890]: E1125 15:03:12.173793 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 15:03:12 crc kubenswrapper[4890]: E1125 15:03:12.173914 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.190363 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd5e648-6322-404d-bf1b-3b59a8092291\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76d277b4001db9d63b0af88b98afcc9019321101429c9be47b3022d3575d4e06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e3a186b14ea199d1aeba5751290c266ec918cb055120cfc5c8a2ae03b1469e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a860a998c40aee3e94e4cfd748f54391b3096fd3ed75c84c7ea2a5a7cefa33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://081a6d623cd043c0919d04e2ca93314b2b85fe33048634aa2637297d54b52cee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f506ec649ce24f4df6fd33e4aeafd13e42167dd71323524a5a3164517659fcc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:12Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.205888 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.205941 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.205952 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.205967 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.205977 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:12Z","lastTransitionTime":"2025-11-25T15:03:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.206336 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d5a9ccb-13ba-4dd5-9904-42aac5eca374\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d19bc48acb392051f0f6c5a6820135bf29ebe6d2afd098697553298933cf280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93325a6407389c4a7db0ba3c1a9c84a6980db3c3277cdf3860583a564b562bbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20505e91fc60158eeb3a8cfa0aafc5b4e4436d5def5e735a3a58b0f9f3621058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3281314d85a20945d56b75ab7b7aef1e382a73e74dff03773d118d03eb0ee16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:12Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.222857 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b33b805b76abbcace899e527d5ee3eaeb9e480fb03dc8eacd3f6dd90d2dd6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:12Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.236736 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2nlkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16f52653-c067-4186-bc40-9c0f7c383b64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmlnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmlnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2nlkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:12Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.255232 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:12Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.272625 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:12Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.286894 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3358a356ae5b422d6bba72bc516374e14ccdace80726aa71b10e25102e0f1100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba3acdae961f49f797f66bbebe350e34a4df5882bcf2a4fe8d37f2ca95865c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:12Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.299957 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8tdrh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1097de4-74e4-4178-b9b4-14a22539da9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfa0291da1598e930aeb4a76f9f4e8b0728001a770acd8f10215465539f8d8b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8tdrh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:12Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.308847 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.308903 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.308922 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.308949 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.308961 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:12Z","lastTransitionTime":"2025-11-25T15:03:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.323986 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86e031a7-b4c8-4030-a078-cdca7ccc9dde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4c4921f0973b03f6dade29c2ef32684a481cd1c3e1e23b64ed636943c9bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f9285acc7ac3cbb503a1e546728748c42971c8c9019cc8ca837076f3e55558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa240816a6e9d2821006ce240eccf27d8f57a7f0c5394314fa7177488da4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://659004bd47948def6b33498d8d1474f28c2e98986754f61af012267f3984269e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6d2dcd8df64e44df8cdb4c3a022f03fccb39451631ca35b30d4c7c35271ff0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:12Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.346430 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg454" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81411ec7-f865-48e2-b464-ab6213ef5e52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db59b77d323079e0b588a6adc0f247dda84b5497de17264390366124934cdca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a366b314408eb8453ab7f970a243c511a2716b5472d92b03c81061135e4e59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a366b314408eb8453ab7f970a243c511a2716b5472d92b03c81061135e4e59c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f841cf4a886c5e17d9919787305ce297d7910b2692be53169324a56bd545be5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f841cf4a886c5e17d9919787305ce297d7910b2692be53169324a56bd545be5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e7825cb1d51e854b5218b09bca98a38c75b356ac9d3d6c206147e98b050dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21e7825cb1d51e854b5218b09bca98a38c75b356ac9d3d6c206147e98b050dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5881047b9fa64b7504f1d62b697a6f352b05d14eedd61e6b64d5ff99d91edda2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5881047b9fa64b7504f1d62b697a6f352b05d14eedd61e6b64d5ff99d91edda2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg454\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:12Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.369028 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4txd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4322dbd78237bf443a985ce3c79bca4cba2103862d81177f5474a76f4905eed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfcq7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4txd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:12Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.385500 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a299b8e0-37bc-48d9-a7e8-64ced2d92d40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5bf7fec98590eb53dd0ecce057c364d4b943bdc0fd152f49a6dfc9586b056f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvcr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90eb11dd969896d6769a18a39ac3c87633b1d42e110093a32bc77bd6f8a7e332\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvcr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6wr7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:12Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.402722 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583964bf462fa0005d471f82f6827552a60659ab97984eb1faf074c7b82c2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:12Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.412746 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.412808 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.412820 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.412856 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.412870 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:12Z","lastTransitionTime":"2025-11-25T15:03:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.417913 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:12Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.429599 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7g5s8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a7dc232-835b-491c-afc9-9eb14f257af3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9a7ac317fbb9c589573e80babe484c21bef4da3d9ab79dfb7ebf5b8300c472c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4cs2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7g5s8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:12Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.450330 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a404958-dffa-4a6a-99fc-3e4950614bd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2f3af2cdb804a47cddb121d619a7545a54af0888b120babca0aa2217ff01d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2f3af2cdb804a47cddb121d619a7545a54af0888b120babca0aa2217ff01d7b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"message\\\":\\\"42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-authentication/oauth-openshift]} name:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 15:02:55.414271 6305 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1125 15:02:55.414314 6305 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 15:02:55.414334 6305 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 15:02:55.414375 6305 factory.go:656] Stopping watch factory\\\\nI1125 15:02:55.414396 6305 ovnkube.go:599] Stopped ovnkube\\\\nI1125 15:02:55.414398 6305 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 15:02:55.414425 6305 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 15:02:55.414443 6305 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 15:02:55.414540 6305 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7zdk8_openshift-ovn-kubernetes(9a404958-dffa-4a6a-99fc-3e4950614bd4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zdk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:12Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.462765 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e4f849d-f239-4727-a73e-18327856929a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://490f1595bc77030d2ee789197fa3748324e1b7a0c9d9180ba08383bbac43adc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0902d6f8e56aecc5a2a13a7148f94a8a00ad139027b3c036aba0215ba0a67f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xx2tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:12Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.515604 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.515639 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.515650 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.515665 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.515676 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:12Z","lastTransitionTime":"2025-11-25T15:03:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.618002 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.618043 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.618054 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.618069 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.618080 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:12Z","lastTransitionTime":"2025-11-25T15:03:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.722950 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.723013 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.723030 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.723052 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.723075 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:12Z","lastTransitionTime":"2025-11-25T15:03:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.826274 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.826314 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.826324 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.826339 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.826348 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:12Z","lastTransitionTime":"2025-11-25T15:03:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.929282 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.929316 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.929327 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.929343 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:12 crc kubenswrapper[4890]: I1125 15:03:12.929356 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:12Z","lastTransitionTime":"2025-11-25T15:03:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.032309 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.032355 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.032388 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.032407 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.032421 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:13Z","lastTransitionTime":"2025-11-25T15:03:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.134636 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.134684 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.134697 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.134713 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.134724 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:13Z","lastTransitionTime":"2025-11-25T15:03:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.172892 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nlkp" Nov 25 15:03:13 crc kubenswrapper[4890]: E1125 15:03:13.173062 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nlkp" podUID="16f52653-c067-4186-bc40-9c0f7c383b64" Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.236513 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.236612 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.236626 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.236646 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.236656 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:13Z","lastTransitionTime":"2025-11-25T15:03:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.268104 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.268248 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 15:03:13 crc kubenswrapper[4890]: E1125 15:03:13.268268 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:03:45.268243059 +0000 UTC m=+83.710705669 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.268310 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.268368 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:03:13 crc kubenswrapper[4890]: E1125 15:03:13.268382 4890 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 15:03:13 crc kubenswrapper[4890]: E1125 15:03:13.268394 4890 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 15:03:13 crc kubenswrapper[4890]: E1125 15:03:13.268404 4890 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 15:03:13 crc kubenswrapper[4890]: E1125 15:03:13.268420 4890 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.268404 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 15:03:13 crc kubenswrapper[4890]: E1125 15:03:13.268440 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 15:03:45.268431394 +0000 UTC m=+83.710894004 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 15:03:13 crc kubenswrapper[4890]: E1125 15:03:13.268482 4890 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 15:03:13 crc kubenswrapper[4890]: E1125 15:03:13.268482 4890 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 15:03:13 crc kubenswrapper[4890]: E1125 15:03:13.268557 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 15:03:45.268510025 +0000 UTC m=+83.710972705 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 15:03:13 crc kubenswrapper[4890]: E1125 15:03:13.268559 4890 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 15:03:13 crc kubenswrapper[4890]: E1125 15:03:13.268581 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 15:03:45.268571327 +0000 UTC m=+83.711034067 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 15:03:13 crc kubenswrapper[4890]: E1125 15:03:13.268582 4890 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 15:03:13 crc kubenswrapper[4890]: E1125 15:03:13.268623 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 15:03:45.268611388 +0000 UTC m=+83.711074008 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.339720 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.339754 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.339764 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.339776 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.339786 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:13Z","lastTransitionTime":"2025-11-25T15:03:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.442559 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.442624 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.442635 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.442653 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.442663 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:13Z","lastTransitionTime":"2025-11-25T15:03:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.545129 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.545217 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.545239 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.545264 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.545281 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:13Z","lastTransitionTime":"2025-11-25T15:03:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.648543 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.648580 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.648623 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.648638 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.648650 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:13Z","lastTransitionTime":"2025-11-25T15:03:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.751820 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.751872 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.751886 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.751906 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.751922 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:13Z","lastTransitionTime":"2025-11-25T15:03:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.854224 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.854270 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.854285 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.854305 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.854319 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:13Z","lastTransitionTime":"2025-11-25T15:03:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.956893 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.957275 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.957285 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.957300 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:13 crc kubenswrapper[4890]: I1125 15:03:13.957309 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:13Z","lastTransitionTime":"2025-11-25T15:03:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.059760 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.059799 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.059808 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.059822 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.059830 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:14Z","lastTransitionTime":"2025-11-25T15:03:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.161631 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.161667 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.161677 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.161692 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.161703 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:14Z","lastTransitionTime":"2025-11-25T15:03:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.172152 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.172396 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.172152 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:03:14 crc kubenswrapper[4890]: E1125 15:03:14.172496 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 15:03:14 crc kubenswrapper[4890]: E1125 15:03:14.172582 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 15:03:14 crc kubenswrapper[4890]: E1125 15:03:14.172647 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.188670 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.198256 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.204298 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:14Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.218536 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:14Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.231330 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3358a356ae5b422d6bba72bc516374e14ccdace80726aa71b10e25102e0f1100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba3acdae961f49f797f66bbebe350e34a4df5882bcf2a4fe8d37f2ca95865c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:14Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.242662 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8tdrh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1097de4-74e4-4178-b9b4-14a22539da9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfa0291da1598e930aeb4a76f9f4e8b0728001a770acd8f10215465539f8d8b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8tdrh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:14Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.263582 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg454" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81411ec7-f865-48e2-b464-ab6213ef5e52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db59b77d323079e0b588a6adc0f247dda84b5497de17264390366124934cdca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a366b314408eb8453ab7f970a243c511a2716b5472d92b03c81061135e4e59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a366b314408eb8453ab7f970a243c511a2716b5472d92b03c81061135e4e59c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f841cf4a886c5e17d9919787305ce297d7910b2692be53169324a56bd545be5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f841cf4a886c5e17d9919787305ce297d7910b2692be53169324a56bd545be5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e7825cb1d51e854b5218b09bca98a38c75b356ac9d3d6c206147e98b050dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21e7825cb1d51e854b5218b09bca98a38c75b356ac9d3d6c206147e98b050dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5881047b9fa64b7504f1d62b697a6f352b05d14eedd61e6b64d5ff99d91edda2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5881047b9fa64b7504f1d62b697a6f352b05d14eedd61e6b64d5ff99d91edda2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg454\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:14Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.265468 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.265510 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.265519 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.265533 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.265669 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:14Z","lastTransitionTime":"2025-11-25T15:03:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.281107 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4txd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4322dbd78237bf443a985ce3c79bca4cba2103862d81177f5474a76f4905eed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfcq7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4txd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:14Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.294301 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a299b8e0-37bc-48d9-a7e8-64ced2d92d40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5bf7fec98590eb53dd0ecce057c364d4b943bdc0fd152f49a6dfc9586b056f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvcr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90eb11dd969896d6769a18a39ac3c87633b1d42e110093a32bc77bd6f8a7e332\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvcr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6wr7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:14Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.312811 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86e031a7-b4c8-4030-a078-cdca7ccc9dde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4c4921f0973b03f6dade29c2ef32684a481cd1c3e1e23b64ed636943c9bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f9285acc7ac3cbb503a1e546728748c42971c8c9019cc8ca837076f3e55558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa240816a6e9d2821006ce240eccf27d8f57a7f0c5394314fa7177488da4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://659004bd47948def6b33498d8d1474f28c2e98986754f61af012267f3984269e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6d2dcd8df64e44df8cdb4c3a022f03fccb39451631ca35b30d4c7c35271ff0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:14Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.329581 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:14Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.340124 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7g5s8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a7dc232-835b-491c-afc9-9eb14f257af3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9a7ac317fbb9c589573e80babe484c21bef4da3d9ab79dfb7ebf5b8300c472c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4cs2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7g5s8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:14Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.358694 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a404958-dffa-4a6a-99fc-3e4950614bd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2f3af2cdb804a47cddb121d619a7545a54af0888b120babca0aa2217ff01d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2f3af2cdb804a47cddb121d619a7545a54af0888b120babca0aa2217ff01d7b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"message\\\":\\\"42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-authentication/oauth-openshift]} name:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 15:02:55.414271 6305 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1125 15:02:55.414314 6305 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 15:02:55.414334 6305 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 15:02:55.414375 6305 factory.go:656] Stopping watch factory\\\\nI1125 15:02:55.414396 6305 ovnkube.go:599] Stopped ovnkube\\\\nI1125 15:02:55.414398 6305 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 15:02:55.414425 6305 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 15:02:55.414443 6305 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 15:02:55.414540 6305 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7zdk8_openshift-ovn-kubernetes(9a404958-dffa-4a6a-99fc-3e4950614bd4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zdk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:14Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.368849 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.368888 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.368896 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.368910 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.368920 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:14Z","lastTransitionTime":"2025-11-25T15:03:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.372271 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e4f849d-f239-4727-a73e-18327856929a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://490f1595bc77030d2ee789197fa3748324e1b7a0c9d9180ba08383bbac43adc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0902d6f8e56aecc5a2a13a7148f94a8a00ad139027b3c036aba0215ba0a67f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xx2tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:14Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.384917 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583964bf462fa0005d471f82f6827552a60659ab97984eb1faf074c7b82c2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:14Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.397280 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d5a9ccb-13ba-4dd5-9904-42aac5eca374\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d19bc48acb392051f0f6c5a6820135bf29ebe6d2afd098697553298933cf280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93325a6407389c4a7db0ba3c1a9c84a6980db3c3277cdf3860583a564b562bbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20505e91fc60158eeb3a8cfa0aafc5b4e4436d5def5e735a3a58b0f9f3621058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3281314d85a20945d56b75ab7b7aef1e382a73e74dff03773d118d03eb0ee16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:14Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.408832 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b33b805b76abbcace899e527d5ee3eaeb9e480fb03dc8eacd3f6dd90d2dd6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:14Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.419452 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2nlkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16f52653-c067-4186-bc40-9c0f7c383b64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmlnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmlnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2nlkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:14Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.431825 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd5e648-6322-404d-bf1b-3b59a8092291\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76d277b4001db9d63b0af88b98afcc9019321101429c9be47b3022d3575d4e06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e3a186b14ea199d1aeba5751290c266ec918cb055120cfc5c8a2ae03b1469e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a860a998c40aee3e94e4cfd748f54391b3096fd3ed75c84c7ea2a5a7cefa33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://081a6d623cd043c0919d04e2ca93314b2b85fe33048634aa2637297d54b52cee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f506ec649ce24f4df6fd33e4aeafd13e42167dd71323524a5a3164517659fcc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:14Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.471840 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.471877 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.471887 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.471906 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.471917 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:14Z","lastTransitionTime":"2025-11-25T15:03:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.574281 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.574351 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.574373 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.574405 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.574430 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:14Z","lastTransitionTime":"2025-11-25T15:03:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.676795 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.676841 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.676852 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.676868 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.676882 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:14Z","lastTransitionTime":"2025-11-25T15:03:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.779575 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.779627 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.779640 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.779672 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.779695 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:14Z","lastTransitionTime":"2025-11-25T15:03:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.882419 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.882459 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.882467 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.882479 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.882488 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:14Z","lastTransitionTime":"2025-11-25T15:03:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.984768 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.984842 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.984856 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.984876 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:14 crc kubenswrapper[4890]: I1125 15:03:14.984890 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:14Z","lastTransitionTime":"2025-11-25T15:03:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:15 crc kubenswrapper[4890]: I1125 15:03:15.087334 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:15 crc kubenswrapper[4890]: I1125 15:03:15.087376 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:15 crc kubenswrapper[4890]: I1125 15:03:15.087383 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:15 crc kubenswrapper[4890]: I1125 15:03:15.087398 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:15 crc kubenswrapper[4890]: I1125 15:03:15.087408 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:15Z","lastTransitionTime":"2025-11-25T15:03:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:15 crc kubenswrapper[4890]: I1125 15:03:15.172645 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nlkp" Nov 25 15:03:15 crc kubenswrapper[4890]: E1125 15:03:15.172828 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nlkp" podUID="16f52653-c067-4186-bc40-9c0f7c383b64" Nov 25 15:03:15 crc kubenswrapper[4890]: I1125 15:03:15.188995 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:15 crc kubenswrapper[4890]: I1125 15:03:15.189038 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:15 crc kubenswrapper[4890]: I1125 15:03:15.189049 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:15 crc kubenswrapper[4890]: I1125 15:03:15.189063 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:15 crc kubenswrapper[4890]: I1125 15:03:15.189074 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:15Z","lastTransitionTime":"2025-11-25T15:03:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:15 crc kubenswrapper[4890]: I1125 15:03:15.291737 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:15 crc kubenswrapper[4890]: I1125 15:03:15.291780 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:15 crc kubenswrapper[4890]: I1125 15:03:15.291790 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:15 crc kubenswrapper[4890]: I1125 15:03:15.291802 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:15 crc kubenswrapper[4890]: I1125 15:03:15.291810 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:15Z","lastTransitionTime":"2025-11-25T15:03:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:15 crc kubenswrapper[4890]: I1125 15:03:15.393768 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:15 crc kubenswrapper[4890]: I1125 15:03:15.393822 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:15 crc kubenswrapper[4890]: I1125 15:03:15.393834 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:15 crc kubenswrapper[4890]: I1125 15:03:15.393850 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:15 crc kubenswrapper[4890]: I1125 15:03:15.393862 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:15Z","lastTransitionTime":"2025-11-25T15:03:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:15 crc kubenswrapper[4890]: I1125 15:03:15.495772 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:15 crc kubenswrapper[4890]: I1125 15:03:15.495816 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:15 crc kubenswrapper[4890]: I1125 15:03:15.495827 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:15 crc kubenswrapper[4890]: I1125 15:03:15.495842 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:15 crc kubenswrapper[4890]: I1125 15:03:15.495855 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:15Z","lastTransitionTime":"2025-11-25T15:03:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:15 crc kubenswrapper[4890]: I1125 15:03:15.598267 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:15 crc kubenswrapper[4890]: I1125 15:03:15.598308 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:15 crc kubenswrapper[4890]: I1125 15:03:15.598321 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:15 crc kubenswrapper[4890]: I1125 15:03:15.598337 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:15 crc kubenswrapper[4890]: I1125 15:03:15.598348 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:15Z","lastTransitionTime":"2025-11-25T15:03:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:15 crc kubenswrapper[4890]: I1125 15:03:15.700515 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:15 crc kubenswrapper[4890]: I1125 15:03:15.700569 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:15 crc kubenswrapper[4890]: I1125 15:03:15.700631 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:15 crc kubenswrapper[4890]: I1125 15:03:15.700664 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:15 crc kubenswrapper[4890]: I1125 15:03:15.700681 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:15Z","lastTransitionTime":"2025-11-25T15:03:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:15 crc kubenswrapper[4890]: I1125 15:03:15.803596 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:15 crc kubenswrapper[4890]: I1125 15:03:15.803652 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:15 crc kubenswrapper[4890]: I1125 15:03:15.803664 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:15 crc kubenswrapper[4890]: I1125 15:03:15.803682 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:15 crc kubenswrapper[4890]: I1125 15:03:15.803695 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:15Z","lastTransitionTime":"2025-11-25T15:03:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:15 crc kubenswrapper[4890]: I1125 15:03:15.905809 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:15 crc kubenswrapper[4890]: I1125 15:03:15.905839 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:15 crc kubenswrapper[4890]: I1125 15:03:15.905848 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:15 crc kubenswrapper[4890]: I1125 15:03:15.905861 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:15 crc kubenswrapper[4890]: I1125 15:03:15.905872 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:15Z","lastTransitionTime":"2025-11-25T15:03:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.008629 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.008669 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.008680 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.008695 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.008705 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:16Z","lastTransitionTime":"2025-11-25T15:03:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.110326 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.110360 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.110368 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.110380 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.110390 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:16Z","lastTransitionTime":"2025-11-25T15:03:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.172321 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.172369 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.172321 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:03:16 crc kubenswrapper[4890]: E1125 15:03:16.172453 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 15:03:16 crc kubenswrapper[4890]: E1125 15:03:16.172529 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 15:03:16 crc kubenswrapper[4890]: E1125 15:03:16.172601 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.212734 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.212790 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.212799 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.212813 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.212822 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:16Z","lastTransitionTime":"2025-11-25T15:03:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.315149 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.315200 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.315209 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.315224 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.315235 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:16Z","lastTransitionTime":"2025-11-25T15:03:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.417763 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.417822 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.417875 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.417902 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.417924 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:16Z","lastTransitionTime":"2025-11-25T15:03:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.519867 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.519907 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.519920 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.519935 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.519945 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:16Z","lastTransitionTime":"2025-11-25T15:03:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.621939 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.621976 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.621987 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.622002 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.622014 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:16Z","lastTransitionTime":"2025-11-25T15:03:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.723829 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.723864 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.723875 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.723889 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.723927 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:16Z","lastTransitionTime":"2025-11-25T15:03:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.826271 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.826318 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.826329 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.826345 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.826359 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:16Z","lastTransitionTime":"2025-11-25T15:03:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.928740 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.928793 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.928808 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.928830 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:16 crc kubenswrapper[4890]: I1125 15:03:16.928845 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:16Z","lastTransitionTime":"2025-11-25T15:03:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.032330 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.032382 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.032393 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.032412 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.032426 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:17Z","lastTransitionTime":"2025-11-25T15:03:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.134859 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.134898 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.134910 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.134925 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.134936 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:17Z","lastTransitionTime":"2025-11-25T15:03:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.172593 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nlkp" Nov 25 15:03:17 crc kubenswrapper[4890]: E1125 15:03:17.172740 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nlkp" podUID="16f52653-c067-4186-bc40-9c0f7c383b64" Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.237575 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.237618 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.237631 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.237649 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.237661 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:17Z","lastTransitionTime":"2025-11-25T15:03:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.340673 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.340725 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.340738 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.340761 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.340774 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:17Z","lastTransitionTime":"2025-11-25T15:03:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.443418 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.443453 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.443461 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.443478 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.443489 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:17Z","lastTransitionTime":"2025-11-25T15:03:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.545599 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.545638 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.545649 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.545665 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.545674 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:17Z","lastTransitionTime":"2025-11-25T15:03:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.651527 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.651575 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.651587 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.651605 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.651623 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:17Z","lastTransitionTime":"2025-11-25T15:03:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.754219 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.754280 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.754298 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.754320 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.754343 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:17Z","lastTransitionTime":"2025-11-25T15:03:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.856005 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.856043 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.856051 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.856066 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.856078 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:17Z","lastTransitionTime":"2025-11-25T15:03:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.959299 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.959369 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.959456 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.959577 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:17 crc kubenswrapper[4890]: I1125 15:03:17.959603 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:17Z","lastTransitionTime":"2025-11-25T15:03:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.062691 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.062735 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.062745 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.062765 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.062776 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:18Z","lastTransitionTime":"2025-11-25T15:03:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.165476 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.165503 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.165510 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.165522 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.165531 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:18Z","lastTransitionTime":"2025-11-25T15:03:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.172459 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.172506 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 15:03:18 crc kubenswrapper[4890]: E1125 15:03:18.172574 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.172733 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:03:18 crc kubenswrapper[4890]: E1125 15:03:18.172892 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 15:03:18 crc kubenswrapper[4890]: E1125 15:03:18.173251 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.267739 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.267774 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.267783 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.267797 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.267806 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:18Z","lastTransitionTime":"2025-11-25T15:03:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.370111 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.370153 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.370198 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.370215 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.370226 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:18Z","lastTransitionTime":"2025-11-25T15:03:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.472978 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.473036 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.473057 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.473077 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.473091 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:18Z","lastTransitionTime":"2025-11-25T15:03:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.575255 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.575312 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.575328 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.575348 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.575365 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:18Z","lastTransitionTime":"2025-11-25T15:03:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.677999 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.678036 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.678045 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.678060 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.678070 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:18Z","lastTransitionTime":"2025-11-25T15:03:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.733020 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.733100 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.733120 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.733146 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.733199 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:18Z","lastTransitionTime":"2025-11-25T15:03:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:18 crc kubenswrapper[4890]: E1125 15:03:18.749550 4890 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7c7b4592-d206-4ee7-bf93-ada4b8bc755a\\\",\\\"systemUUID\\\":\\\"5350de30-707e-4bde-b990-1fa65ae3a01e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:18Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.753106 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.753141 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.753151 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.753181 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.753193 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:18Z","lastTransitionTime":"2025-11-25T15:03:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:18 crc kubenswrapper[4890]: E1125 15:03:18.768888 4890 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7c7b4592-d206-4ee7-bf93-ada4b8bc755a\\\",\\\"systemUUID\\\":\\\"5350de30-707e-4bde-b990-1fa65ae3a01e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:18Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.773149 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.773209 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.773218 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.773236 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.773246 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:18Z","lastTransitionTime":"2025-11-25T15:03:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:18 crc kubenswrapper[4890]: E1125 15:03:18.784378 4890 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7c7b4592-d206-4ee7-bf93-ada4b8bc755a\\\",\\\"systemUUID\\\":\\\"5350de30-707e-4bde-b990-1fa65ae3a01e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:18Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.787799 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.787839 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.787850 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.787863 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.787873 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:18Z","lastTransitionTime":"2025-11-25T15:03:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:18 crc kubenswrapper[4890]: E1125 15:03:18.804770 4890 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7c7b4592-d206-4ee7-bf93-ada4b8bc755a\\\",\\\"systemUUID\\\":\\\"5350de30-707e-4bde-b990-1fa65ae3a01e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:18Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.808018 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.808053 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.808067 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.808083 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.808097 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:18Z","lastTransitionTime":"2025-11-25T15:03:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:18 crc kubenswrapper[4890]: E1125 15:03:18.821879 4890 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7c7b4592-d206-4ee7-bf93-ada4b8bc755a\\\",\\\"systemUUID\\\":\\\"5350de30-707e-4bde-b990-1fa65ae3a01e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:18Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:18 crc kubenswrapper[4890]: E1125 15:03:18.822153 4890 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.824151 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.824214 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.824235 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.824254 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.824265 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:18Z","lastTransitionTime":"2025-11-25T15:03:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.928104 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.928191 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.928215 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.928245 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:18 crc kubenswrapper[4890]: I1125 15:03:18.928267 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:18Z","lastTransitionTime":"2025-11-25T15:03:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.030711 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.030751 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.030763 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.030779 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.030791 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:19Z","lastTransitionTime":"2025-11-25T15:03:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.133483 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.133523 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.133533 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.133548 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.133558 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:19Z","lastTransitionTime":"2025-11-25T15:03:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.173238 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nlkp" Nov 25 15:03:19 crc kubenswrapper[4890]: E1125 15:03:19.173537 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nlkp" podUID="16f52653-c067-4186-bc40-9c0f7c383b64" Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.173794 4890 scope.go:117] "RemoveContainer" containerID="f2f3af2cdb804a47cddb121d619a7545a54af0888b120babca0aa2217ff01d7b" Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.236495 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.236551 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.236561 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.236575 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.236591 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:19Z","lastTransitionTime":"2025-11-25T15:03:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.340279 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.340333 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.340345 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.340363 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.340377 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:19Z","lastTransitionTime":"2025-11-25T15:03:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.443268 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.443306 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.443316 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.443331 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.443343 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:19Z","lastTransitionTime":"2025-11-25T15:03:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.547128 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.547210 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.547227 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.547249 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.547264 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:19Z","lastTransitionTime":"2025-11-25T15:03:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.649518 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.649553 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.649560 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.649572 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.649581 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:19Z","lastTransitionTime":"2025-11-25T15:03:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.752034 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.752086 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.752099 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.752115 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.752127 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:19Z","lastTransitionTime":"2025-11-25T15:03:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.854680 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.854714 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.854724 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.854738 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.854763 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:19Z","lastTransitionTime":"2025-11-25T15:03:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.957473 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.957520 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.957529 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.957543 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:19 crc kubenswrapper[4890]: I1125 15:03:19.957555 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:19Z","lastTransitionTime":"2025-11-25T15:03:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.059835 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.059877 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.059886 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.059898 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.059908 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:20Z","lastTransitionTime":"2025-11-25T15:03:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.121518 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7zdk8_9a404958-dffa-4a6a-99fc-3e4950614bd4/ovnkube-controller/2.log" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.122097 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7zdk8_9a404958-dffa-4a6a-99fc-3e4950614bd4/ovnkube-controller/1.log" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.124998 4890 generic.go:334] "Generic (PLEG): container finished" podID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerID="fc71e729066adb3b4b41639d6028364bc27ae5a938d23ff9b00f7c813f207d8d" exitCode=1 Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.125042 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" event={"ID":"9a404958-dffa-4a6a-99fc-3e4950614bd4","Type":"ContainerDied","Data":"fc71e729066adb3b4b41639d6028364bc27ae5a938d23ff9b00f7c813f207d8d"} Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.125086 4890 scope.go:117] "RemoveContainer" containerID="f2f3af2cdb804a47cddb121d619a7545a54af0888b120babca0aa2217ff01d7b" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.125643 4890 scope.go:117] "RemoveContainer" containerID="fc71e729066adb3b4b41639d6028364bc27ae5a938d23ff9b00f7c813f207d8d" Nov 25 15:03:20 crc kubenswrapper[4890]: E1125 15:03:20.125885 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7zdk8_openshift-ovn-kubernetes(9a404958-dffa-4a6a-99fc-3e4950614bd4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" podUID="9a404958-dffa-4a6a-99fc-3e4950614bd4" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.138451 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8tdrh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1097de4-74e4-4178-b9b4-14a22539da9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfa0291da1598e930aeb4a76f9f4e8b0728001a770acd8f10215465539f8d8b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8tdrh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:20Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.149247 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9762ab15-c9d3-40ac-a789-7b885e5f779f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1154cddf4cb46b8a9db26e7f982ed9dd382d85b5ceb95e50325a475eed408f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca4d48ac675ee73e0946cc802994bb407b6770695320ce2380ddfce26f1b5b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://539bab27d80f8d76cc54a60f508a9985e8a3aa3d2f949ea5a5c2f0d92227da30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8092652bfe2886bf2ea5e6e0cbd6c84efc82b21182a4509a624b399943eede7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8092652bfe2886bf2ea5e6e0cbd6c84efc82b21182a4509a624b399943eede7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:20Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.159642 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:20Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.162308 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.162334 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.162341 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.162355 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.162364 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:20Z","lastTransitionTime":"2025-11-25T15:03:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.171065 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:20Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.172218 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.172257 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.172277 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 15:03:20 crc kubenswrapper[4890]: E1125 15:03:20.172347 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 15:03:20 crc kubenswrapper[4890]: E1125 15:03:20.172512 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 15:03:20 crc kubenswrapper[4890]: E1125 15:03:20.172553 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.183612 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3358a356ae5b422d6bba72bc516374e14ccdace80726aa71b10e25102e0f1100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba3acdae961f49f797f66bbebe350e34a4df5882bcf2a4fe8d37f2ca95865c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:20Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.202403 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86e031a7-b4c8-4030-a078-cdca7ccc9dde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4c4921f0973b03f6dade29c2ef32684a481cd1c3e1e23b64ed636943c9bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f9285acc7ac3cbb503a1e546728748c42971c8c9019cc8ca837076f3e55558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa240816a6e9d2821006ce240eccf27d8f57a7f0c5394314fa7177488da4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://659004bd47948def6b33498d8d1474f28c2e98986754f61af012267f3984269e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6d2dcd8df64e44df8cdb4c3a022f03fccb39451631ca35b30d4c7c35271ff0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:20Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.214783 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg454" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81411ec7-f865-48e2-b464-ab6213ef5e52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db59b77d323079e0b588a6adc0f247dda84b5497de17264390366124934cdca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a366b314408eb8453ab7f970a243c511a2716b5472d92b03c81061135e4e59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a366b314408eb8453ab7f970a243c511a2716b5472d92b03c81061135e4e59c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f841cf4a886c5e17d9919787305ce297d7910b2692be53169324a56bd545be5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f841cf4a886c5e17d9919787305ce297d7910b2692be53169324a56bd545be5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e7825cb1d51e854b5218b09bca98a38c75b356ac9d3d6c206147e98b050dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21e7825cb1d51e854b5218b09bca98a38c75b356ac9d3d6c206147e98b050dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5881047b9fa64b7504f1d62b697a6f352b05d14eedd61e6b64d5ff99d91edda2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5881047b9fa64b7504f1d62b697a6f352b05d14eedd61e6b64d5ff99d91edda2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg454\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:20Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.228707 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4txd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4322dbd78237bf443a985ce3c79bca4cba2103862d81177f5474a76f4905eed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfcq7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4txd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:20Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.239586 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a299b8e0-37bc-48d9-a7e8-64ced2d92d40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5bf7fec98590eb53dd0ecce057c364d4b943bdc0fd152f49a6dfc9586b056f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvcr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90eb11dd969896d6769a18a39ac3c87633b1d42e110093a32bc77bd6f8a7e332\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvcr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6wr7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:20Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.249759 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e4f849d-f239-4727-a73e-18327856929a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://490f1595bc77030d2ee789197fa3748324e1b7a0c9d9180ba08383bbac43adc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0902d6f8e56aecc5a2a13a7148f94a8a00ad139027b3c036aba0215ba0a67f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xx2tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:20Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.261508 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583964bf462fa0005d471f82f6827552a60659ab97984eb1faf074c7b82c2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:20Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.264024 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.264055 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.264068 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.264085 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.264094 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:20Z","lastTransitionTime":"2025-11-25T15:03:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.272281 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:20Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.280567 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7g5s8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a7dc232-835b-491c-afc9-9eb14f257af3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9a7ac317fbb9c589573e80babe484c21bef4da3d9ab79dfb7ebf5b8300c472c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4cs2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7g5s8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:20Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.295944 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a404958-dffa-4a6a-99fc-3e4950614bd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc71e729066adb3b4b41639d6028364bc27ae5a938d23ff9b00f7c813f207d8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2f3af2cdb804a47cddb121d619a7545a54af0888b120babca0aa2217ff01d7b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"message\\\":\\\"42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-authentication/oauth-openshift]} name:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 15:02:55.414271 6305 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1125 15:02:55.414314 6305 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 15:02:55.414334 6305 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 15:02:55.414375 6305 factory.go:656] Stopping watch factory\\\\nI1125 15:02:55.414396 6305 ovnkube.go:599] Stopped ovnkube\\\\nI1125 15:02:55.414398 6305 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 15:02:55.414425 6305 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 15:02:55.414443 6305 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 15:02:55.414540 6305 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc71e729066adb3b4b41639d6028364bc27ae5a938d23ff9b00f7c813f207d8d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T15:03:19Z\\\",\\\"message\\\":\\\"egistry/node-ca-8tdrh in node crc\\\\nI1125 15:03:19.917932 6586 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1125 15:03:19.917938 6586 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-8tdrh after 0 failed attempt(s)\\\\nI1125 15:03:19.917940 6586 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1125 15:03:19.917942 6586 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-8tdrh\\\\nI1125 15:03:19.917945 6586 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1125 15:03:19.917892 6586 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s in node crc\\\\nI1125 15:03:19.917955 6586 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s after 0 failed attempt(s)\\\\nI1125 15:03:19.917960 6586 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s\\\\nI1125 15:03:19.917799 6586 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-zg454\\\\nI1125 15:03:19.917970\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T15:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zdk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:20Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.305254 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2nlkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16f52653-c067-4186-bc40-9c0f7c383b64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmlnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmlnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2nlkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:20Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.316077 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd5e648-6322-404d-bf1b-3b59a8092291\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76d277b4001db9d63b0af88b98afcc9019321101429c9be47b3022d3575d4e06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e3a186b14ea199d1aeba5751290c266ec918cb055120cfc5c8a2ae03b1469e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a860a998c40aee3e94e4cfd748f54391b3096fd3ed75c84c7ea2a5a7cefa33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://081a6d623cd043c0919d04e2ca93314b2b85fe33048634aa2637297d54b52cee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f506ec649ce24f4df6fd33e4aeafd13e42167dd71323524a5a3164517659fcc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:20Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.350029 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d5a9ccb-13ba-4dd5-9904-42aac5eca374\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d19bc48acb392051f0f6c5a6820135bf29ebe6d2afd098697553298933cf280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93325a6407389c4a7db0ba3c1a9c84a6980db3c3277cdf3860583a564b562bbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20505e91fc60158eeb3a8cfa0aafc5b4e4436d5def5e735a3a58b0f9f3621058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3281314d85a20945d56b75ab7b7aef1e382a73e74dff03773d118d03eb0ee16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:20Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.367325 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.367365 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.367377 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.367393 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.367405 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:20Z","lastTransitionTime":"2025-11-25T15:03:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.374617 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b33b805b76abbcace899e527d5ee3eaeb9e480fb03dc8eacd3f6dd90d2dd6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:20Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.469742 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.469783 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.469792 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.469808 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.469816 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:20Z","lastTransitionTime":"2025-11-25T15:03:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.572594 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.572631 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.572638 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.572651 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.572661 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:20Z","lastTransitionTime":"2025-11-25T15:03:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.675062 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.675136 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.675202 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.675237 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.675264 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:20Z","lastTransitionTime":"2025-11-25T15:03:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.777332 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.777364 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.777372 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.777384 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.777394 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:20Z","lastTransitionTime":"2025-11-25T15:03:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.880153 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.880274 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.880300 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.880330 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.880352 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:20Z","lastTransitionTime":"2025-11-25T15:03:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.982932 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.982973 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.982982 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.982997 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:20 crc kubenswrapper[4890]: I1125 15:03:20.983006 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:20Z","lastTransitionTime":"2025-11-25T15:03:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:21 crc kubenswrapper[4890]: I1125 15:03:21.084970 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:21 crc kubenswrapper[4890]: I1125 15:03:21.085004 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:21 crc kubenswrapper[4890]: I1125 15:03:21.085020 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:21 crc kubenswrapper[4890]: I1125 15:03:21.085040 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:21 crc kubenswrapper[4890]: I1125 15:03:21.085054 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:21Z","lastTransitionTime":"2025-11-25T15:03:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:21 crc kubenswrapper[4890]: I1125 15:03:21.129669 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7zdk8_9a404958-dffa-4a6a-99fc-3e4950614bd4/ovnkube-controller/2.log" Nov 25 15:03:21 crc kubenswrapper[4890]: I1125 15:03:21.173040 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nlkp" Nov 25 15:03:21 crc kubenswrapper[4890]: E1125 15:03:21.173211 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nlkp" podUID="16f52653-c067-4186-bc40-9c0f7c383b64" Nov 25 15:03:21 crc kubenswrapper[4890]: I1125 15:03:21.187525 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:21 crc kubenswrapper[4890]: I1125 15:03:21.187559 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:21 crc kubenswrapper[4890]: I1125 15:03:21.187570 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:21 crc kubenswrapper[4890]: I1125 15:03:21.187584 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:21 crc kubenswrapper[4890]: I1125 15:03:21.187595 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:21Z","lastTransitionTime":"2025-11-25T15:03:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:21 crc kubenswrapper[4890]: I1125 15:03:21.289959 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:21 crc kubenswrapper[4890]: I1125 15:03:21.289996 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:21 crc kubenswrapper[4890]: I1125 15:03:21.290007 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:21 crc kubenswrapper[4890]: I1125 15:03:21.290023 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:21 crc kubenswrapper[4890]: I1125 15:03:21.290034 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:21Z","lastTransitionTime":"2025-11-25T15:03:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:21 crc kubenswrapper[4890]: I1125 15:03:21.392441 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:21 crc kubenswrapper[4890]: I1125 15:03:21.392486 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:21 crc kubenswrapper[4890]: I1125 15:03:21.392499 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:21 crc kubenswrapper[4890]: I1125 15:03:21.392516 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:21 crc kubenswrapper[4890]: I1125 15:03:21.392526 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:21Z","lastTransitionTime":"2025-11-25T15:03:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:21 crc kubenswrapper[4890]: I1125 15:03:21.494894 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:21 crc kubenswrapper[4890]: I1125 15:03:21.494946 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:21 crc kubenswrapper[4890]: I1125 15:03:21.494957 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:21 crc kubenswrapper[4890]: I1125 15:03:21.494977 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:21 crc kubenswrapper[4890]: I1125 15:03:21.494988 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:21Z","lastTransitionTime":"2025-11-25T15:03:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:21 crc kubenswrapper[4890]: I1125 15:03:21.597281 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:21 crc kubenswrapper[4890]: I1125 15:03:21.597311 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:21 crc kubenswrapper[4890]: I1125 15:03:21.597319 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:21 crc kubenswrapper[4890]: I1125 15:03:21.597331 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:21 crc kubenswrapper[4890]: I1125 15:03:21.597341 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:21Z","lastTransitionTime":"2025-11-25T15:03:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:21 crc kubenswrapper[4890]: I1125 15:03:21.699689 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:21 crc kubenswrapper[4890]: I1125 15:03:21.699728 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:21 crc kubenswrapper[4890]: I1125 15:03:21.699741 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:21 crc kubenswrapper[4890]: I1125 15:03:21.699759 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:21 crc kubenswrapper[4890]: I1125 15:03:21.699773 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:21Z","lastTransitionTime":"2025-11-25T15:03:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:21 crc kubenswrapper[4890]: I1125 15:03:21.802516 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:21 crc kubenswrapper[4890]: I1125 15:03:21.803125 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:21 crc kubenswrapper[4890]: I1125 15:03:21.803140 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:21 crc kubenswrapper[4890]: I1125 15:03:21.803201 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:21 crc kubenswrapper[4890]: I1125 15:03:21.803216 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:21Z","lastTransitionTime":"2025-11-25T15:03:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:21 crc kubenswrapper[4890]: I1125 15:03:21.905265 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:21 crc kubenswrapper[4890]: I1125 15:03:21.905310 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:21 crc kubenswrapper[4890]: I1125 15:03:21.905322 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:21 crc kubenswrapper[4890]: I1125 15:03:21.905338 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:21 crc kubenswrapper[4890]: I1125 15:03:21.905349 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:21Z","lastTransitionTime":"2025-11-25T15:03:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.007582 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.007636 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.007647 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.007667 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.007681 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:22Z","lastTransitionTime":"2025-11-25T15:03:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.114444 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.114515 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.114525 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.114540 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.114549 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:22Z","lastTransitionTime":"2025-11-25T15:03:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.173147 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.173213 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.173346 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 15:03:22 crc kubenswrapper[4890]: E1125 15:03:22.173484 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 15:03:22 crc kubenswrapper[4890]: E1125 15:03:22.173560 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 15:03:22 crc kubenswrapper[4890]: E1125 15:03:22.173671 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.192544 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86e031a7-b4c8-4030-a078-cdca7ccc9dde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4c4921f0973b03f6dade29c2ef32684a481cd1c3e1e23b64ed636943c9bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f9285acc7ac3cbb503a1e546728748c42971c8c9019cc8ca837076f3e55558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa240816a6e9d2821006ce240eccf27d8f57a7f0c5394314fa7177488da4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://659004bd47948def6b33498d8d1474f28c2e98986754f61af012267f3984269e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6d2dcd8df64e44df8cdb4c3a022f03fccb39451631ca35b30d4c7c35271ff0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:22Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.205213 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg454" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81411ec7-f865-48e2-b464-ab6213ef5e52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db59b77d323079e0b588a6adc0f247dda84b5497de17264390366124934cdca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a366b314408eb8453ab7f970a243c511a2716b5472d92b03c81061135e4e59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a366b314408eb8453ab7f970a243c511a2716b5472d92b03c81061135e4e59c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f841cf4a886c5e17d9919787305ce297d7910b2692be53169324a56bd545be5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f841cf4a886c5e17d9919787305ce297d7910b2692be53169324a56bd545be5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e7825cb1d51e854b5218b09bca98a38c75b356ac9d3d6c206147e98b050dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21e7825cb1d51e854b5218b09bca98a38c75b356ac9d3d6c206147e98b050dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5881047b9fa64b7504f1d62b697a6f352b05d14eedd61e6b64d5ff99d91edda2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5881047b9fa64b7504f1d62b697a6f352b05d14eedd61e6b64d5ff99d91edda2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg454\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:22Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.216847 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4txd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4322dbd78237bf443a985ce3c79bca4cba2103862d81177f5474a76f4905eed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfcq7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4txd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:22Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.217036 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.217069 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.217080 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.217095 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.217104 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:22Z","lastTransitionTime":"2025-11-25T15:03:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.226297 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a299b8e0-37bc-48d9-a7e8-64ced2d92d40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5bf7fec98590eb53dd0ecce057c364d4b943bdc0fd152f49a6dfc9586b056f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvcr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90eb11dd969896d6769a18a39ac3c87633b1d42e110093a32bc77bd6f8a7e332\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvcr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6wr7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:22Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.236559 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583964bf462fa0005d471f82f6827552a60659ab97984eb1faf074c7b82c2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:22Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.246932 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:22Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.257188 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7g5s8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a7dc232-835b-491c-afc9-9eb14f257af3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9a7ac317fbb9c589573e80babe484c21bef4da3d9ab79dfb7ebf5b8300c472c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4cs2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7g5s8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:22Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.274658 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a404958-dffa-4a6a-99fc-3e4950614bd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc71e729066adb3b4b41639d6028364bc27ae5a938d23ff9b00f7c813f207d8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2f3af2cdb804a47cddb121d619a7545a54af0888b120babca0aa2217ff01d7b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"message\\\":\\\"42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-authentication/oauth-openshift]} name:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 15:02:55.414271 6305 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1125 15:02:55.414314 6305 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 15:02:55.414334 6305 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 15:02:55.414375 6305 factory.go:656] Stopping watch factory\\\\nI1125 15:02:55.414396 6305 ovnkube.go:599] Stopped ovnkube\\\\nI1125 15:02:55.414398 6305 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 15:02:55.414425 6305 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 15:02:55.414443 6305 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 15:02:55.414540 6305 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc71e729066adb3b4b41639d6028364bc27ae5a938d23ff9b00f7c813f207d8d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T15:03:19Z\\\",\\\"message\\\":\\\"egistry/node-ca-8tdrh in node crc\\\\nI1125 15:03:19.917932 6586 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1125 15:03:19.917938 6586 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-8tdrh after 0 failed attempt(s)\\\\nI1125 15:03:19.917940 6586 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1125 15:03:19.917942 6586 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-8tdrh\\\\nI1125 15:03:19.917945 6586 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1125 15:03:19.917892 6586 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s in node crc\\\\nI1125 15:03:19.917955 6586 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s after 0 failed attempt(s)\\\\nI1125 15:03:19.917960 6586 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s\\\\nI1125 15:03:19.917799 6586 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-zg454\\\\nI1125 15:03:19.917970\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T15:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zdk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:22Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.285342 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e4f849d-f239-4727-a73e-18327856929a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://490f1595bc77030d2ee789197fa3748324e1b7a0c9d9180ba08383bbac43adc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0902d6f8e56aecc5a2a13a7148f94a8a00ad139027b3c036aba0215ba0a67f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xx2tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:22Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.300691 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd5e648-6322-404d-bf1b-3b59a8092291\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76d277b4001db9d63b0af88b98afcc9019321101429c9be47b3022d3575d4e06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e3a186b14ea199d1aeba5751290c266ec918cb055120cfc5c8a2ae03b1469e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a860a998c40aee3e94e4cfd748f54391b3096fd3ed75c84c7ea2a5a7cefa33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://081a6d623cd043c0919d04e2ca93314b2b85fe33048634aa2637297d54b52cee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f506ec649ce24f4df6fd33e4aeafd13e42167dd71323524a5a3164517659fcc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:22Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.314398 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d5a9ccb-13ba-4dd5-9904-42aac5eca374\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d19bc48acb392051f0f6c5a6820135bf29ebe6d2afd098697553298933cf280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93325a6407389c4a7db0ba3c1a9c84a6980db3c3277cdf3860583a564b562bbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20505e91fc60158eeb3a8cfa0aafc5b4e4436d5def5e735a3a58b0f9f3621058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3281314d85a20945d56b75ab7b7aef1e382a73e74dff03773d118d03eb0ee16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:22Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.319680 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.319736 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.319745 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.319762 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.319775 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:22Z","lastTransitionTime":"2025-11-25T15:03:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.326420 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b33b805b76abbcace899e527d5ee3eaeb9e480fb03dc8eacd3f6dd90d2dd6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:22Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.340106 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2nlkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16f52653-c067-4186-bc40-9c0f7c383b64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmlnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmlnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2nlkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:22Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.349779 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9762ab15-c9d3-40ac-a789-7b885e5f779f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1154cddf4cb46b8a9db26e7f982ed9dd382d85b5ceb95e50325a475eed408f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca4d48ac675ee73e0946cc802994bb407b6770695320ce2380ddfce26f1b5b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://539bab27d80f8d76cc54a60f508a9985e8a3aa3d2f949ea5a5c2f0d92227da30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8092652bfe2886bf2ea5e6e0cbd6c84efc82b21182a4509a624b399943eede7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8092652bfe2886bf2ea5e6e0cbd6c84efc82b21182a4509a624b399943eede7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:22Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.360744 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:22Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.373549 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:22Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.387038 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3358a356ae5b422d6bba72bc516374e14ccdace80726aa71b10e25102e0f1100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba3acdae961f49f797f66bbebe350e34a4df5882bcf2a4fe8d37f2ca95865c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:22Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.399015 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8tdrh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1097de4-74e4-4178-b9b4-14a22539da9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfa0291da1598e930aeb4a76f9f4e8b0728001a770acd8f10215465539f8d8b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8tdrh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:22Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.422060 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.422111 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.422124 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.422143 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.422176 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:22Z","lastTransitionTime":"2025-11-25T15:03:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.524343 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.524381 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.524391 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.524405 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.524416 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:22Z","lastTransitionTime":"2025-11-25T15:03:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.626785 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.626832 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.626847 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.626868 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.626884 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:22Z","lastTransitionTime":"2025-11-25T15:03:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.730017 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.730087 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.730103 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.730123 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.730136 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:22Z","lastTransitionTime":"2025-11-25T15:03:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.832137 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.832218 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.832230 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.832245 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.832255 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:22Z","lastTransitionTime":"2025-11-25T15:03:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.934354 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.934389 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.934401 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.934415 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:22 crc kubenswrapper[4890]: I1125 15:03:22.934426 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:22Z","lastTransitionTime":"2025-11-25T15:03:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.038960 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.039009 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.039034 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.039057 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.039071 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:23Z","lastTransitionTime":"2025-11-25T15:03:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.142034 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.142100 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.142119 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.142150 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.142215 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:23Z","lastTransitionTime":"2025-11-25T15:03:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.172771 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nlkp" Nov 25 15:03:23 crc kubenswrapper[4890]: E1125 15:03:23.173038 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nlkp" podUID="16f52653-c067-4186-bc40-9c0f7c383b64" Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.245785 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.245845 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.245858 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.245878 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.245891 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:23Z","lastTransitionTime":"2025-11-25T15:03:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.349287 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.349353 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.349364 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.349385 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.349397 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:23Z","lastTransitionTime":"2025-11-25T15:03:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.451970 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.452026 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.452037 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.452056 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.452068 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:23Z","lastTransitionTime":"2025-11-25T15:03:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.555119 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.555240 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.555256 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.555272 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.555288 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:23Z","lastTransitionTime":"2025-11-25T15:03:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.657875 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.657924 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.657939 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.657957 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.657967 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:23Z","lastTransitionTime":"2025-11-25T15:03:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.761243 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.761284 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.761292 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.761307 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.761316 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:23Z","lastTransitionTime":"2025-11-25T15:03:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.864107 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.864151 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.864177 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.864193 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.864209 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:23Z","lastTransitionTime":"2025-11-25T15:03:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.966838 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.966886 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.966897 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.966912 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:23 crc kubenswrapper[4890]: I1125 15:03:23.966923 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:23Z","lastTransitionTime":"2025-11-25T15:03:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.069386 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.069426 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.069436 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.069452 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.069462 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:24Z","lastTransitionTime":"2025-11-25T15:03:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.171478 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.171520 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.171531 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.171546 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.171556 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:24Z","lastTransitionTime":"2025-11-25T15:03:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.172824 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 15:03:24 crc kubenswrapper[4890]: E1125 15:03:24.172900 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.172967 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 15:03:24 crc kubenswrapper[4890]: E1125 15:03:24.173119 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.172967 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:03:24 crc kubenswrapper[4890]: E1125 15:03:24.173374 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.273917 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.273970 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.273985 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.274004 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.274016 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:24Z","lastTransitionTime":"2025-11-25T15:03:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.376630 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.376692 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.376712 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.376735 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.376751 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:24Z","lastTransitionTime":"2025-11-25T15:03:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.479626 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.479670 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.479684 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.479701 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.479711 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:24Z","lastTransitionTime":"2025-11-25T15:03:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.582544 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.582608 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.582623 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.582642 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.582658 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:24Z","lastTransitionTime":"2025-11-25T15:03:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.685064 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.685101 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.685115 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.685132 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.685176 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:24Z","lastTransitionTime":"2025-11-25T15:03:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.787742 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.787792 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.787803 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.787822 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.787834 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:24Z","lastTransitionTime":"2025-11-25T15:03:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.889994 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.890068 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.890081 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.890104 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.890122 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:24Z","lastTransitionTime":"2025-11-25T15:03:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.992356 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.992390 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.992398 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.992410 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:24 crc kubenswrapper[4890]: I1125 15:03:24.992421 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:24Z","lastTransitionTime":"2025-11-25T15:03:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:25 crc kubenswrapper[4890]: I1125 15:03:25.095125 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:25 crc kubenswrapper[4890]: I1125 15:03:25.095207 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:25 crc kubenswrapper[4890]: I1125 15:03:25.095220 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:25 crc kubenswrapper[4890]: I1125 15:03:25.095237 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:25 crc kubenswrapper[4890]: I1125 15:03:25.095250 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:25Z","lastTransitionTime":"2025-11-25T15:03:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:25 crc kubenswrapper[4890]: I1125 15:03:25.172410 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nlkp" Nov 25 15:03:25 crc kubenswrapper[4890]: E1125 15:03:25.172541 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nlkp" podUID="16f52653-c067-4186-bc40-9c0f7c383b64" Nov 25 15:03:25 crc kubenswrapper[4890]: I1125 15:03:25.198524 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:25 crc kubenswrapper[4890]: I1125 15:03:25.198583 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:25 crc kubenswrapper[4890]: I1125 15:03:25.198600 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:25 crc kubenswrapper[4890]: I1125 15:03:25.198621 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:25 crc kubenswrapper[4890]: I1125 15:03:25.198633 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:25Z","lastTransitionTime":"2025-11-25T15:03:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:25 crc kubenswrapper[4890]: I1125 15:03:25.300822 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:25 crc kubenswrapper[4890]: I1125 15:03:25.300870 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:25 crc kubenswrapper[4890]: I1125 15:03:25.300880 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:25 crc kubenswrapper[4890]: I1125 15:03:25.300895 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:25 crc kubenswrapper[4890]: I1125 15:03:25.300904 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:25Z","lastTransitionTime":"2025-11-25T15:03:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:25 crc kubenswrapper[4890]: I1125 15:03:25.403596 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:25 crc kubenswrapper[4890]: I1125 15:03:25.403636 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:25 crc kubenswrapper[4890]: I1125 15:03:25.403645 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:25 crc kubenswrapper[4890]: I1125 15:03:25.403659 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:25 crc kubenswrapper[4890]: I1125 15:03:25.403684 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:25Z","lastTransitionTime":"2025-11-25T15:03:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:25 crc kubenswrapper[4890]: I1125 15:03:25.505668 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:25 crc kubenswrapper[4890]: I1125 15:03:25.505717 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:25 crc kubenswrapper[4890]: I1125 15:03:25.505728 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:25 crc kubenswrapper[4890]: I1125 15:03:25.505745 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:25 crc kubenswrapper[4890]: I1125 15:03:25.505756 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:25Z","lastTransitionTime":"2025-11-25T15:03:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:25 crc kubenswrapper[4890]: I1125 15:03:25.608857 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:25 crc kubenswrapper[4890]: I1125 15:03:25.609067 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:25 crc kubenswrapper[4890]: I1125 15:03:25.609091 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:25 crc kubenswrapper[4890]: I1125 15:03:25.609126 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:25 crc kubenswrapper[4890]: I1125 15:03:25.609155 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:25Z","lastTransitionTime":"2025-11-25T15:03:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:25 crc kubenswrapper[4890]: I1125 15:03:25.711504 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:25 crc kubenswrapper[4890]: I1125 15:03:25.711539 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:25 crc kubenswrapper[4890]: I1125 15:03:25.711547 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:25 crc kubenswrapper[4890]: I1125 15:03:25.711562 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:25 crc kubenswrapper[4890]: I1125 15:03:25.711574 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:25Z","lastTransitionTime":"2025-11-25T15:03:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:25 crc kubenswrapper[4890]: I1125 15:03:25.814147 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:25 crc kubenswrapper[4890]: I1125 15:03:25.814197 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:25 crc kubenswrapper[4890]: I1125 15:03:25.814208 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:25 crc kubenswrapper[4890]: I1125 15:03:25.814223 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:25 crc kubenswrapper[4890]: I1125 15:03:25.814234 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:25Z","lastTransitionTime":"2025-11-25T15:03:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:25 crc kubenswrapper[4890]: I1125 15:03:25.916997 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:25 crc kubenswrapper[4890]: I1125 15:03:25.917028 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:25 crc kubenswrapper[4890]: I1125 15:03:25.917038 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:25 crc kubenswrapper[4890]: I1125 15:03:25.917053 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:25 crc kubenswrapper[4890]: I1125 15:03:25.917063 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:25Z","lastTransitionTime":"2025-11-25T15:03:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.019761 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.019834 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.019852 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.019875 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.019892 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:26Z","lastTransitionTime":"2025-11-25T15:03:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.122832 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.122868 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.122878 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.122892 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.122902 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:26Z","lastTransitionTime":"2025-11-25T15:03:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.172622 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.172673 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.173416 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:03:26 crc kubenswrapper[4890]: E1125 15:03:26.173552 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 15:03:26 crc kubenswrapper[4890]: E1125 15:03:26.173662 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 15:03:26 crc kubenswrapper[4890]: E1125 15:03:26.173787 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.225232 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.225274 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.225285 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.225303 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.225314 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:26Z","lastTransitionTime":"2025-11-25T15:03:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.327341 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.327375 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.327383 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.327398 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.327407 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:26Z","lastTransitionTime":"2025-11-25T15:03:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.430197 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.430242 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.430258 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.430275 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.430287 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:26Z","lastTransitionTime":"2025-11-25T15:03:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.533624 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.533659 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.533667 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.533680 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.533690 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:26Z","lastTransitionTime":"2025-11-25T15:03:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.636204 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.636249 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.636260 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.636276 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.636286 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:26Z","lastTransitionTime":"2025-11-25T15:03:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.739188 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.739235 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.739246 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.739262 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.739272 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:26Z","lastTransitionTime":"2025-11-25T15:03:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.841222 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.841256 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.841268 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.841285 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.841296 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:26Z","lastTransitionTime":"2025-11-25T15:03:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.943854 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.943900 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.943911 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.943931 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:26 crc kubenswrapper[4890]: I1125 15:03:26.943944 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:26Z","lastTransitionTime":"2025-11-25T15:03:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.046496 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.046542 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.046554 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.046572 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.046583 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:27Z","lastTransitionTime":"2025-11-25T15:03:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.148705 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.148741 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.148753 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.148768 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.148779 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:27Z","lastTransitionTime":"2025-11-25T15:03:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.172252 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nlkp" Nov 25 15:03:27 crc kubenswrapper[4890]: E1125 15:03:27.172404 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nlkp" podUID="16f52653-c067-4186-bc40-9c0f7c383b64" Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.251347 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.251377 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.251387 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.251399 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.251410 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:27Z","lastTransitionTime":"2025-11-25T15:03:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.353470 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.353513 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.353522 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.353535 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.353544 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:27Z","lastTransitionTime":"2025-11-25T15:03:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.456012 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.456050 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.456060 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.456081 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.456096 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:27Z","lastTransitionTime":"2025-11-25T15:03:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.558281 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.558318 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.558326 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.558339 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.558348 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:27Z","lastTransitionTime":"2025-11-25T15:03:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.660710 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.660764 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.660776 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.660794 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.660805 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:27Z","lastTransitionTime":"2025-11-25T15:03:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.763010 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.763071 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.763088 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.763112 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.763129 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:27Z","lastTransitionTime":"2025-11-25T15:03:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.865363 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.865391 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.865400 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.865416 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.865424 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:27Z","lastTransitionTime":"2025-11-25T15:03:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.967949 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.967992 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.968004 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.968016 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:27 crc kubenswrapper[4890]: I1125 15:03:27.968026 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:27Z","lastTransitionTime":"2025-11-25T15:03:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.070117 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.070178 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.070190 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.070207 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.070217 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:28Z","lastTransitionTime":"2025-11-25T15:03:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.172422 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.172540 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 15:03:28 crc kubenswrapper[4890]: E1125 15:03:28.172639 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.172638 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:03:28 crc kubenswrapper[4890]: E1125 15:03:28.172741 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 15:03:28 crc kubenswrapper[4890]: E1125 15:03:28.172724 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.172826 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.172862 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.172873 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.172890 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.172903 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:28Z","lastTransitionTime":"2025-11-25T15:03:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.219732 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/16f52653-c067-4186-bc40-9c0f7c383b64-metrics-certs\") pod \"network-metrics-daemon-2nlkp\" (UID: \"16f52653-c067-4186-bc40-9c0f7c383b64\") " pod="openshift-multus/network-metrics-daemon-2nlkp" Nov 25 15:03:28 crc kubenswrapper[4890]: E1125 15:03:28.219855 4890 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 15:03:28 crc kubenswrapper[4890]: E1125 15:03:28.219926 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/16f52653-c067-4186-bc40-9c0f7c383b64-metrics-certs podName:16f52653-c067-4186-bc40-9c0f7c383b64 nodeName:}" failed. No retries permitted until 2025-11-25 15:04:00.219907927 +0000 UTC m=+98.662370537 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/16f52653-c067-4186-bc40-9c0f7c383b64-metrics-certs") pod "network-metrics-daemon-2nlkp" (UID: "16f52653-c067-4186-bc40-9c0f7c383b64") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.275633 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.275663 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.275672 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.275686 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.275694 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:28Z","lastTransitionTime":"2025-11-25T15:03:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.378401 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.378434 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.378442 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.378458 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.378468 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:28Z","lastTransitionTime":"2025-11-25T15:03:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.481221 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.481258 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.481267 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.481281 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.481291 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:28Z","lastTransitionTime":"2025-11-25T15:03:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.584314 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.584348 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.584358 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.584372 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.584380 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:28Z","lastTransitionTime":"2025-11-25T15:03:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.686924 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.686976 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.686991 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.687012 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.687028 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:28Z","lastTransitionTime":"2025-11-25T15:03:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.789600 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.789622 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.789631 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.789644 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.789652 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:28Z","lastTransitionTime":"2025-11-25T15:03:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.888922 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.888970 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.888980 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.888994 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.889004 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:28Z","lastTransitionTime":"2025-11-25T15:03:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:28 crc kubenswrapper[4890]: E1125 15:03:28.901515 4890 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7c7b4592-d206-4ee7-bf93-ada4b8bc755a\\\",\\\"systemUUID\\\":\\\"5350de30-707e-4bde-b990-1fa65ae3a01e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:28Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.905248 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.905292 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.905306 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.905323 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.905337 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:28Z","lastTransitionTime":"2025-11-25T15:03:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:28 crc kubenswrapper[4890]: E1125 15:03:28.918400 4890 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7c7b4592-d206-4ee7-bf93-ada4b8bc755a\\\",\\\"systemUUID\\\":\\\"5350de30-707e-4bde-b990-1fa65ae3a01e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:28Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.922067 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.922109 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.922119 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.922136 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.922145 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:28Z","lastTransitionTime":"2025-11-25T15:03:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:28 crc kubenswrapper[4890]: E1125 15:03:28.937853 4890 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7c7b4592-d206-4ee7-bf93-ada4b8bc755a\\\",\\\"systemUUID\\\":\\\"5350de30-707e-4bde-b990-1fa65ae3a01e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:28Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.941761 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.941795 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.941805 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.941819 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.941830 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:28Z","lastTransitionTime":"2025-11-25T15:03:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:28 crc kubenswrapper[4890]: E1125 15:03:28.952544 4890 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7c7b4592-d206-4ee7-bf93-ada4b8bc755a\\\",\\\"systemUUID\\\":\\\"5350de30-707e-4bde-b990-1fa65ae3a01e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:28Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.956110 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.956149 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.956176 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.956191 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.956199 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:28Z","lastTransitionTime":"2025-11-25T15:03:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:28 crc kubenswrapper[4890]: E1125 15:03:28.966919 4890 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7c7b4592-d206-4ee7-bf93-ada4b8bc755a\\\",\\\"systemUUID\\\":\\\"5350de30-707e-4bde-b990-1fa65ae3a01e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:28Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:28 crc kubenswrapper[4890]: E1125 15:03:28.967027 4890 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.968297 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.968325 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.968334 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.968345 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:28 crc kubenswrapper[4890]: I1125 15:03:28.968353 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:28Z","lastTransitionTime":"2025-11-25T15:03:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.070764 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.070804 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.070813 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.070828 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.070838 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:29Z","lastTransitionTime":"2025-11-25T15:03:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.172281 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nlkp" Nov 25 15:03:29 crc kubenswrapper[4890]: E1125 15:03:29.172436 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nlkp" podUID="16f52653-c067-4186-bc40-9c0f7c383b64" Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.174008 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.174044 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.174053 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.175065 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.175091 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:29Z","lastTransitionTime":"2025-11-25T15:03:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.277548 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.277584 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.277595 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.277610 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.277622 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:29Z","lastTransitionTime":"2025-11-25T15:03:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.379296 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.379328 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.379336 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.379351 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.379360 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:29Z","lastTransitionTime":"2025-11-25T15:03:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.482769 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.482825 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.482837 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.482859 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.482875 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:29Z","lastTransitionTime":"2025-11-25T15:03:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.585589 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.585665 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.585683 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.585708 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.585729 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:29Z","lastTransitionTime":"2025-11-25T15:03:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.688757 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.688802 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.688814 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.688830 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.688845 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:29Z","lastTransitionTime":"2025-11-25T15:03:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.790955 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.790993 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.791005 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.791019 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.791029 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:29Z","lastTransitionTime":"2025-11-25T15:03:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.893317 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.893353 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.893361 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.893375 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.893394 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:29Z","lastTransitionTime":"2025-11-25T15:03:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.995731 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.995771 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.995782 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.995798 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:29 crc kubenswrapper[4890]: I1125 15:03:29.995811 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:29Z","lastTransitionTime":"2025-11-25T15:03:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:30 crc kubenswrapper[4890]: I1125 15:03:30.098541 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:30 crc kubenswrapper[4890]: I1125 15:03:30.098591 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:30 crc kubenswrapper[4890]: I1125 15:03:30.098605 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:30 crc kubenswrapper[4890]: I1125 15:03:30.098621 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:30 crc kubenswrapper[4890]: I1125 15:03:30.098632 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:30Z","lastTransitionTime":"2025-11-25T15:03:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:30 crc kubenswrapper[4890]: I1125 15:03:30.173042 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 15:03:30 crc kubenswrapper[4890]: I1125 15:03:30.173051 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:03:30 crc kubenswrapper[4890]: I1125 15:03:30.173068 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 15:03:30 crc kubenswrapper[4890]: E1125 15:03:30.173232 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 15:03:30 crc kubenswrapper[4890]: E1125 15:03:30.173453 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 15:03:30 crc kubenswrapper[4890]: E1125 15:03:30.173548 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 15:03:30 crc kubenswrapper[4890]: I1125 15:03:30.200630 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:30 crc kubenswrapper[4890]: I1125 15:03:30.200681 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:30 crc kubenswrapper[4890]: I1125 15:03:30.200693 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:30 crc kubenswrapper[4890]: I1125 15:03:30.200709 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:30 crc kubenswrapper[4890]: I1125 15:03:30.200720 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:30Z","lastTransitionTime":"2025-11-25T15:03:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:30 crc kubenswrapper[4890]: I1125 15:03:30.303652 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:30 crc kubenswrapper[4890]: I1125 15:03:30.303709 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:30 crc kubenswrapper[4890]: I1125 15:03:30.303718 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:30 crc kubenswrapper[4890]: I1125 15:03:30.303732 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:30 crc kubenswrapper[4890]: I1125 15:03:30.303742 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:30Z","lastTransitionTime":"2025-11-25T15:03:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:30 crc kubenswrapper[4890]: I1125 15:03:30.405641 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:30 crc kubenswrapper[4890]: I1125 15:03:30.406002 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:30 crc kubenswrapper[4890]: I1125 15:03:30.406014 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:30 crc kubenswrapper[4890]: I1125 15:03:30.406034 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:30 crc kubenswrapper[4890]: I1125 15:03:30.406048 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:30Z","lastTransitionTime":"2025-11-25T15:03:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:30 crc kubenswrapper[4890]: I1125 15:03:30.509210 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:30 crc kubenswrapper[4890]: I1125 15:03:30.509272 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:30 crc kubenswrapper[4890]: I1125 15:03:30.509283 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:30 crc kubenswrapper[4890]: I1125 15:03:30.509307 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:30 crc kubenswrapper[4890]: I1125 15:03:30.509319 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:30Z","lastTransitionTime":"2025-11-25T15:03:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:30 crc kubenswrapper[4890]: I1125 15:03:30.612619 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:30 crc kubenswrapper[4890]: I1125 15:03:30.612678 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:30 crc kubenswrapper[4890]: I1125 15:03:30.612690 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:30 crc kubenswrapper[4890]: I1125 15:03:30.612709 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:30 crc kubenswrapper[4890]: I1125 15:03:30.612723 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:30Z","lastTransitionTime":"2025-11-25T15:03:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:30 crc kubenswrapper[4890]: I1125 15:03:30.715589 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:30 crc kubenswrapper[4890]: I1125 15:03:30.715657 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:30 crc kubenswrapper[4890]: I1125 15:03:30.715683 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:30 crc kubenswrapper[4890]: I1125 15:03:30.715721 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:30 crc kubenswrapper[4890]: I1125 15:03:30.715736 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:30Z","lastTransitionTime":"2025-11-25T15:03:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:30 crc kubenswrapper[4890]: I1125 15:03:30.817795 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:30 crc kubenswrapper[4890]: I1125 15:03:30.817828 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:30 crc kubenswrapper[4890]: I1125 15:03:30.817836 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:30 crc kubenswrapper[4890]: I1125 15:03:30.817851 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:30 crc kubenswrapper[4890]: I1125 15:03:30.817859 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:30Z","lastTransitionTime":"2025-11-25T15:03:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:30 crc kubenswrapper[4890]: I1125 15:03:30.920065 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:30 crc kubenswrapper[4890]: I1125 15:03:30.920111 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:30 crc kubenswrapper[4890]: I1125 15:03:30.920128 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:30 crc kubenswrapper[4890]: I1125 15:03:30.920177 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:30 crc kubenswrapper[4890]: I1125 15:03:30.920197 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:30Z","lastTransitionTime":"2025-11-25T15:03:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.022691 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.022734 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.022746 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.022761 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.022770 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:31Z","lastTransitionTime":"2025-11-25T15:03:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.125377 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.125443 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.125454 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.125470 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.125480 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:31Z","lastTransitionTime":"2025-11-25T15:03:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.163439 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4txd2_0fcdeaed-8701-42ad-8534-a77b7f3f93f6/kube-multus/0.log" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.163524 4890 generic.go:334] "Generic (PLEG): container finished" podID="0fcdeaed-8701-42ad-8534-a77b7f3f93f6" containerID="b4322dbd78237bf443a985ce3c79bca4cba2103862d81177f5474a76f4905eed" exitCode=1 Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.163575 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4txd2" event={"ID":"0fcdeaed-8701-42ad-8534-a77b7f3f93f6","Type":"ContainerDied","Data":"b4322dbd78237bf443a985ce3c79bca4cba2103862d81177f5474a76f4905eed"} Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.164301 4890 scope.go:117] "RemoveContainer" containerID="b4322dbd78237bf443a985ce3c79bca4cba2103862d81177f5474a76f4905eed" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.172140 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nlkp" Nov 25 15:03:31 crc kubenswrapper[4890]: E1125 15:03:31.172380 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nlkp" podUID="16f52653-c067-4186-bc40-9c0f7c383b64" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.183900 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.184688 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583964bf462fa0005d471f82f6827552a60659ab97984eb1faf074c7b82c2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:31Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.198434 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:31Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.207601 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7g5s8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a7dc232-835b-491c-afc9-9eb14f257af3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9a7ac317fbb9c589573e80babe484c21bef4da3d9ab79dfb7ebf5b8300c472c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4cs2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7g5s8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:31Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.225703 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a404958-dffa-4a6a-99fc-3e4950614bd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc71e729066adb3b4b41639d6028364bc27ae5a938d23ff9b00f7c813f207d8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2f3af2cdb804a47cddb121d619a7545a54af0888b120babca0aa2217ff01d7b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"message\\\":\\\"42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-authentication/oauth-openshift]} name:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 15:02:55.414271 6305 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1125 15:02:55.414314 6305 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 15:02:55.414334 6305 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 15:02:55.414375 6305 factory.go:656] Stopping watch factory\\\\nI1125 15:02:55.414396 6305 ovnkube.go:599] Stopped ovnkube\\\\nI1125 15:02:55.414398 6305 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 15:02:55.414425 6305 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 15:02:55.414443 6305 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 15:02:55.414540 6305 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc71e729066adb3b4b41639d6028364bc27ae5a938d23ff9b00f7c813f207d8d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T15:03:19Z\\\",\\\"message\\\":\\\"egistry/node-ca-8tdrh in node crc\\\\nI1125 15:03:19.917932 6586 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1125 15:03:19.917938 6586 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-8tdrh after 0 failed attempt(s)\\\\nI1125 15:03:19.917940 6586 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1125 15:03:19.917942 6586 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-8tdrh\\\\nI1125 15:03:19.917945 6586 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1125 15:03:19.917892 6586 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s in node crc\\\\nI1125 15:03:19.917955 6586 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s after 0 failed attempt(s)\\\\nI1125 15:03:19.917960 6586 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s\\\\nI1125 15:03:19.917799 6586 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-zg454\\\\nI1125 15:03:19.917970\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T15:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zdk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:31Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.228598 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.228621 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.228628 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.228641 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.228650 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:31Z","lastTransitionTime":"2025-11-25T15:03:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.235893 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e4f849d-f239-4727-a73e-18327856929a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://490f1595bc77030d2ee789197fa3748324e1b7a0c9d9180ba08383bbac43adc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0902d6f8e56aecc5a2a13a7148f94a8a00ad139027b3c036aba0215ba0a67f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xx2tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:31Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.248007 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd5e648-6322-404d-bf1b-3b59a8092291\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76d277b4001db9d63b0af88b98afcc9019321101429c9be47b3022d3575d4e06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e3a186b14ea199d1aeba5751290c266ec918cb055120cfc5c8a2ae03b1469e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a860a998c40aee3e94e4cfd748f54391b3096fd3ed75c84c7ea2a5a7cefa33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://081a6d623cd043c0919d04e2ca93314b2b85fe33048634aa2637297d54b52cee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f506ec649ce24f4df6fd33e4aeafd13e42167dd71323524a5a3164517659fcc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:31Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.260341 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d5a9ccb-13ba-4dd5-9904-42aac5eca374\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d19bc48acb392051f0f6c5a6820135bf29ebe6d2afd098697553298933cf280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93325a6407389c4a7db0ba3c1a9c84a6980db3c3277cdf3860583a564b562bbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20505e91fc60158eeb3a8cfa0aafc5b4e4436d5def5e735a3a58b0f9f3621058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3281314d85a20945d56b75ab7b7aef1e382a73e74dff03773d118d03eb0ee16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:31Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.275305 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b33b805b76abbcace899e527d5ee3eaeb9e480fb03dc8eacd3f6dd90d2dd6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:31Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.287501 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2nlkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16f52653-c067-4186-bc40-9c0f7c383b64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmlnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmlnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2nlkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:31Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.301824 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9762ab15-c9d3-40ac-a789-7b885e5f779f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1154cddf4cb46b8a9db26e7f982ed9dd382d85b5ceb95e50325a475eed408f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca4d48ac675ee73e0946cc802994bb407b6770695320ce2380ddfce26f1b5b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://539bab27d80f8d76cc54a60f508a9985e8a3aa3d2f949ea5a5c2f0d92227da30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8092652bfe2886bf2ea5e6e0cbd6c84efc82b21182a4509a624b399943eede7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8092652bfe2886bf2ea5e6e0cbd6c84efc82b21182a4509a624b399943eede7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:31Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.316516 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:31Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.329812 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:31Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.330469 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.330497 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.330508 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.330524 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.330536 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:31Z","lastTransitionTime":"2025-11-25T15:03:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.341484 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3358a356ae5b422d6bba72bc516374e14ccdace80726aa71b10e25102e0f1100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba3acdae961f49f797f66bbebe350e34a4df5882bcf2a4fe8d37f2ca95865c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:31Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.349825 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8tdrh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1097de4-74e4-4178-b9b4-14a22539da9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfa0291da1598e930aeb4a76f9f4e8b0728001a770acd8f10215465539f8d8b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8tdrh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:31Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.367729 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86e031a7-b4c8-4030-a078-cdca7ccc9dde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4c4921f0973b03f6dade29c2ef32684a481cd1c3e1e23b64ed636943c9bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f9285acc7ac3cbb503a1e546728748c42971c8c9019cc8ca837076f3e55558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa240816a6e9d2821006ce240eccf27d8f57a7f0c5394314fa7177488da4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://659004bd47948def6b33498d8d1474f28c2e98986754f61af012267f3984269e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6d2dcd8df64e44df8cdb4c3a022f03fccb39451631ca35b30d4c7c35271ff0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:31Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.382018 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg454" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81411ec7-f865-48e2-b464-ab6213ef5e52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db59b77d323079e0b588a6adc0f247dda84b5497de17264390366124934cdca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a366b314408eb8453ab7f970a243c511a2716b5472d92b03c81061135e4e59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a366b314408eb8453ab7f970a243c511a2716b5472d92b03c81061135e4e59c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f841cf4a886c5e17d9919787305ce297d7910b2692be53169324a56bd545be5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f841cf4a886c5e17d9919787305ce297d7910b2692be53169324a56bd545be5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e7825cb1d51e854b5218b09bca98a38c75b356ac9d3d6c206147e98b050dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21e7825cb1d51e854b5218b09bca98a38c75b356ac9d3d6c206147e98b050dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5881047b9fa64b7504f1d62b697a6f352b05d14eedd61e6b64d5ff99d91edda2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5881047b9fa64b7504f1d62b697a6f352b05d14eedd61e6b64d5ff99d91edda2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg454\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:31Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.394455 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4txd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4322dbd78237bf443a985ce3c79bca4cba2103862d81177f5474a76f4905eed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4322dbd78237bf443a985ce3c79bca4cba2103862d81177f5474a76f4905eed\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T15:03:30Z\\\",\\\"message\\\":\\\"2025-11-25T15:02:44+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_bd112228-266e-450f-a5e1-285339caf6b6\\\\n2025-11-25T15:02:44+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_bd112228-266e-450f-a5e1-285339caf6b6 to /host/opt/cni/bin/\\\\n2025-11-25T15:02:45Z [verbose] multus-daemon started\\\\n2025-11-25T15:02:45Z [verbose] Readiness Indicator file check\\\\n2025-11-25T15:03:30Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfcq7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4txd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:31Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.406277 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a299b8e0-37bc-48d9-a7e8-64ced2d92d40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5bf7fec98590eb53dd0ecce057c364d4b943bdc0fd152f49a6dfc9586b056f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvcr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90eb11dd969896d6769a18a39ac3c87633b1d42e110093a32bc77bd6f8a7e332\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvcr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6wr7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:31Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.432713 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.432744 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.432752 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.432765 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.432774 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:31Z","lastTransitionTime":"2025-11-25T15:03:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.535445 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.535488 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.535497 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.535512 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.535522 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:31Z","lastTransitionTime":"2025-11-25T15:03:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.637600 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.637634 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.637647 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.637663 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.637673 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:31Z","lastTransitionTime":"2025-11-25T15:03:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.740120 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.740193 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.740204 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.740221 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.740230 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:31Z","lastTransitionTime":"2025-11-25T15:03:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.842888 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.842938 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.842948 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.842965 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.842974 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:31Z","lastTransitionTime":"2025-11-25T15:03:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.946242 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.946319 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.946340 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.946371 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:31 crc kubenswrapper[4890]: I1125 15:03:31.946391 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:31Z","lastTransitionTime":"2025-11-25T15:03:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.049392 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.049446 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.049456 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.049474 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.049490 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:32Z","lastTransitionTime":"2025-11-25T15:03:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.152071 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.152110 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.152121 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.152135 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.152144 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:32Z","lastTransitionTime":"2025-11-25T15:03:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.170842 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4txd2_0fcdeaed-8701-42ad-8534-a77b7f3f93f6/kube-multus/0.log" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.170971 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4txd2" event={"ID":"0fcdeaed-8701-42ad-8534-a77b7f3f93f6","Type":"ContainerStarted","Data":"4168b721ad650800e545d62906f69c441da51f6d8f11dd2537c8f2b2301afc2b"} Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.172689 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.172777 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.172776 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:03:32 crc kubenswrapper[4890]: E1125 15:03:32.172866 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 15:03:32 crc kubenswrapper[4890]: E1125 15:03:32.172962 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 15:03:32 crc kubenswrapper[4890]: E1125 15:03:32.173063 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.189856 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583964bf462fa0005d471f82f6827552a60659ab97984eb1faf074c7b82c2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:32Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.203486 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:32Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.216397 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7g5s8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a7dc232-835b-491c-afc9-9eb14f257af3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9a7ac317fbb9c589573e80babe484c21bef4da3d9ab79dfb7ebf5b8300c472c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4cs2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7g5s8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:32Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.239013 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a404958-dffa-4a6a-99fc-3e4950614bd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc71e729066adb3b4b41639d6028364bc27ae5a938d23ff9b00f7c813f207d8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2f3af2cdb804a47cddb121d619a7545a54af0888b120babca0aa2217ff01d7b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"message\\\":\\\"42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-authentication/oauth-openshift]} name:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 15:02:55.414271 6305 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1125 15:02:55.414314 6305 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 15:02:55.414334 6305 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 15:02:55.414375 6305 factory.go:656] Stopping watch factory\\\\nI1125 15:02:55.414396 6305 ovnkube.go:599] Stopped ovnkube\\\\nI1125 15:02:55.414398 6305 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 15:02:55.414425 6305 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 15:02:55.414443 6305 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 15:02:55.414540 6305 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc71e729066adb3b4b41639d6028364bc27ae5a938d23ff9b00f7c813f207d8d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T15:03:19Z\\\",\\\"message\\\":\\\"egistry/node-ca-8tdrh in node crc\\\\nI1125 15:03:19.917932 6586 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1125 15:03:19.917938 6586 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-8tdrh after 0 failed attempt(s)\\\\nI1125 15:03:19.917940 6586 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1125 15:03:19.917942 6586 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-8tdrh\\\\nI1125 15:03:19.917945 6586 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1125 15:03:19.917892 6586 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s in node crc\\\\nI1125 15:03:19.917955 6586 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s after 0 failed attempt(s)\\\\nI1125 15:03:19.917960 6586 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s\\\\nI1125 15:03:19.917799 6586 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-zg454\\\\nI1125 15:03:19.917970\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T15:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zdk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:32Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.253382 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e4f849d-f239-4727-a73e-18327856929a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://490f1595bc77030d2ee789197fa3748324e1b7a0c9d9180ba08383bbac43adc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0902d6f8e56aecc5a2a13a7148f94a8a00ad139027b3c036aba0215ba0a67f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xx2tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:32Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.255118 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.255144 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.255152 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.255182 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.255191 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:32Z","lastTransitionTime":"2025-11-25T15:03:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.277486 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd5e648-6322-404d-bf1b-3b59a8092291\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76d277b4001db9d63b0af88b98afcc9019321101429c9be47b3022d3575d4e06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e3a186b14ea199d1aeba5751290c266ec918cb055120cfc5c8a2ae03b1469e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a860a998c40aee3e94e4cfd748f54391b3096fd3ed75c84c7ea2a5a7cefa33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://081a6d623cd043c0919d04e2ca93314b2b85fe33048634aa2637297d54b52cee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f506ec649ce24f4df6fd33e4aeafd13e42167dd71323524a5a3164517659fcc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:32Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.290854 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d5a9ccb-13ba-4dd5-9904-42aac5eca374\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d19bc48acb392051f0f6c5a6820135bf29ebe6d2afd098697553298933cf280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93325a6407389c4a7db0ba3c1a9c84a6980db3c3277cdf3860583a564b562bbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20505e91fc60158eeb3a8cfa0aafc5b4e4436d5def5e735a3a58b0f9f3621058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3281314d85a20945d56b75ab7b7aef1e382a73e74dff03773d118d03eb0ee16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:32Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.303418 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2921ec70-5155-49f0-b2d8-a0e3e21b599c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83ee18494b4658ba014b0013673acea38ee70056ab1f6cd6aa4e173fe8531458\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a6ad38c7fe3f9eb0dfc09bbd22aba8c16d4b51b9bd9142f4204f25b914155c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a6ad38c7fe3f9eb0dfc09bbd22aba8c16d4b51b9bd9142f4204f25b914155c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:32Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.324395 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b33b805b76abbcace899e527d5ee3eaeb9e480fb03dc8eacd3f6dd90d2dd6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:32Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.339620 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2nlkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16f52653-c067-4186-bc40-9c0f7c383b64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmlnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmlnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2nlkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:32Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.351931 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9762ab15-c9d3-40ac-a789-7b885e5f779f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1154cddf4cb46b8a9db26e7f982ed9dd382d85b5ceb95e50325a475eed408f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca4d48ac675ee73e0946cc802994bb407b6770695320ce2380ddfce26f1b5b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://539bab27d80f8d76cc54a60f508a9985e8a3aa3d2f949ea5a5c2f0d92227da30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8092652bfe2886bf2ea5e6e0cbd6c84efc82b21182a4509a624b399943eede7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8092652bfe2886bf2ea5e6e0cbd6c84efc82b21182a4509a624b399943eede7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:32Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.357686 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.357723 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.357738 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.357755 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.357767 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:32Z","lastTransitionTime":"2025-11-25T15:03:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.369027 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:32Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.384311 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:32Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.397228 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3358a356ae5b422d6bba72bc516374e14ccdace80726aa71b10e25102e0f1100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba3acdae961f49f797f66bbebe350e34a4df5882bcf2a4fe8d37f2ca95865c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:32Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.411481 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8tdrh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1097de4-74e4-4178-b9b4-14a22539da9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfa0291da1598e930aeb4a76f9f4e8b0728001a770acd8f10215465539f8d8b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8tdrh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:32Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.432782 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86e031a7-b4c8-4030-a078-cdca7ccc9dde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4c4921f0973b03f6dade29c2ef32684a481cd1c3e1e23b64ed636943c9bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f9285acc7ac3cbb503a1e546728748c42971c8c9019cc8ca837076f3e55558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa240816a6e9d2821006ce240eccf27d8f57a7f0c5394314fa7177488da4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://659004bd47948def6b33498d8d1474f28c2e98986754f61af012267f3984269e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6d2dcd8df64e44df8cdb4c3a022f03fccb39451631ca35b30d4c7c35271ff0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:32Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.448821 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg454" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81411ec7-f865-48e2-b464-ab6213ef5e52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db59b77d323079e0b588a6adc0f247dda84b5497de17264390366124934cdca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a366b314408eb8453ab7f970a243c511a2716b5472d92b03c81061135e4e59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a366b314408eb8453ab7f970a243c511a2716b5472d92b03c81061135e4e59c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f841cf4a886c5e17d9919787305ce297d7910b2692be53169324a56bd545be5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f841cf4a886c5e17d9919787305ce297d7910b2692be53169324a56bd545be5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e7825cb1d51e854b5218b09bca98a38c75b356ac9d3d6c206147e98b050dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21e7825cb1d51e854b5218b09bca98a38c75b356ac9d3d6c206147e98b050dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5881047b9fa64b7504f1d62b697a6f352b05d14eedd61e6b64d5ff99d91edda2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5881047b9fa64b7504f1d62b697a6f352b05d14eedd61e6b64d5ff99d91edda2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg454\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:32Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.460755 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.460787 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.460796 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.460810 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.460820 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:32Z","lastTransitionTime":"2025-11-25T15:03:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.461143 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4txd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4168b721ad650800e545d62906f69c441da51f6d8f11dd2537c8f2b2301afc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4322dbd78237bf443a985ce3c79bca4cba2103862d81177f5474a76f4905eed\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T15:03:30Z\\\",\\\"message\\\":\\\"2025-11-25T15:02:44+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_bd112228-266e-450f-a5e1-285339caf6b6\\\\n2025-11-25T15:02:44+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_bd112228-266e-450f-a5e1-285339caf6b6 to /host/opt/cni/bin/\\\\n2025-11-25T15:02:45Z [verbose] multus-daemon started\\\\n2025-11-25T15:02:45Z [verbose] Readiness Indicator file check\\\\n2025-11-25T15:03:30Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfcq7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4txd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:32Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.473942 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a299b8e0-37bc-48d9-a7e8-64ced2d92d40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5bf7fec98590eb53dd0ecce057c364d4b943bdc0fd152f49a6dfc9586b056f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvcr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90eb11dd969896d6769a18a39ac3c87633b1d42e110093a32bc77bd6f8a7e332\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvcr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6wr7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:32Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.488400 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg454" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81411ec7-f865-48e2-b464-ab6213ef5e52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db59b77d323079e0b588a6adc0f247dda84b5497de17264390366124934cdca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a366b314408eb8453ab7f970a243c511a2716b5472d92b03c81061135e4e59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a366b314408eb8453ab7f970a243c511a2716b5472d92b03c81061135e4e59c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f841cf4a886c5e17d9919787305ce297d7910b2692be53169324a56bd545be5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f841cf4a886c5e17d9919787305ce297d7910b2692be53169324a56bd545be5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e7825cb1d51e854b5218b09bca98a38c75b356ac9d3d6c206147e98b050dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21e7825cb1d51e854b5218b09bca98a38c75b356ac9d3d6c206147e98b050dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5881047b9fa64b7504f1d62b697a6f352b05d14eedd61e6b64d5ff99d91edda2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5881047b9fa64b7504f1d62b697a6f352b05d14eedd61e6b64d5ff99d91edda2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg454\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:32Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.502386 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4txd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4168b721ad650800e545d62906f69c441da51f6d8f11dd2537c8f2b2301afc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4322dbd78237bf443a985ce3c79bca4cba2103862d81177f5474a76f4905eed\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T15:03:30Z\\\",\\\"message\\\":\\\"2025-11-25T15:02:44+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_bd112228-266e-450f-a5e1-285339caf6b6\\\\n2025-11-25T15:02:44+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_bd112228-266e-450f-a5e1-285339caf6b6 to /host/opt/cni/bin/\\\\n2025-11-25T15:02:45Z [verbose] multus-daemon started\\\\n2025-11-25T15:02:45Z [verbose] Readiness Indicator file check\\\\n2025-11-25T15:03:30Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfcq7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4txd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:32Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.518328 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a299b8e0-37bc-48d9-a7e8-64ced2d92d40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5bf7fec98590eb53dd0ecce057c364d4b943bdc0fd152f49a6dfc9586b056f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvcr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90eb11dd969896d6769a18a39ac3c87633b1d42e110093a32bc77bd6f8a7e332\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvcr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6wr7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:32Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.538336 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86e031a7-b4c8-4030-a078-cdca7ccc9dde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4c4921f0973b03f6dade29c2ef32684a481cd1c3e1e23b64ed636943c9bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f9285acc7ac3cbb503a1e546728748c42971c8c9019cc8ca837076f3e55558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa240816a6e9d2821006ce240eccf27d8f57a7f0c5394314fa7177488da4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://659004bd47948def6b33498d8d1474f28c2e98986754f61af012267f3984269e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6d2dcd8df64e44df8cdb4c3a022f03fccb39451631ca35b30d4c7c35271ff0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:32Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.550854 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:32Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.562233 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7g5s8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a7dc232-835b-491c-afc9-9eb14f257af3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9a7ac317fbb9c589573e80babe484c21bef4da3d9ab79dfb7ebf5b8300c472c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4cs2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7g5s8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:32Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.563421 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.563450 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.563458 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.563473 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.563483 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:32Z","lastTransitionTime":"2025-11-25T15:03:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.585532 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a404958-dffa-4a6a-99fc-3e4950614bd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc71e729066adb3b4b41639d6028364bc27ae5a938d23ff9b00f7c813f207d8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2f3af2cdb804a47cddb121d619a7545a54af0888b120babca0aa2217ff01d7b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"message\\\":\\\"42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-authentication/oauth-openshift]} name:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 15:02:55.414271 6305 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1125 15:02:55.414314 6305 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 15:02:55.414334 6305 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 15:02:55.414375 6305 factory.go:656] Stopping watch factory\\\\nI1125 15:02:55.414396 6305 ovnkube.go:599] Stopped ovnkube\\\\nI1125 15:02:55.414398 6305 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 15:02:55.414425 6305 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 15:02:55.414443 6305 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 15:02:55.414540 6305 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc71e729066adb3b4b41639d6028364bc27ae5a938d23ff9b00f7c813f207d8d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T15:03:19Z\\\",\\\"message\\\":\\\"egistry/node-ca-8tdrh in node crc\\\\nI1125 15:03:19.917932 6586 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1125 15:03:19.917938 6586 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-8tdrh after 0 failed attempt(s)\\\\nI1125 15:03:19.917940 6586 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1125 15:03:19.917942 6586 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-8tdrh\\\\nI1125 15:03:19.917945 6586 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1125 15:03:19.917892 6586 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s in node crc\\\\nI1125 15:03:19.917955 6586 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s after 0 failed attempt(s)\\\\nI1125 15:03:19.917960 6586 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s\\\\nI1125 15:03:19.917799 6586 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-zg454\\\\nI1125 15:03:19.917970\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T15:03:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zdk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:32Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.598721 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e4f849d-f239-4727-a73e-18327856929a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://490f1595bc77030d2ee789197fa3748324e1b7a0c9d9180ba08383bbac43adc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0902d6f8e56aecc5a2a13a7148f94a8a00ad139027b3c036aba0215ba0a67f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xx2tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:32Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.612376 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583964bf462fa0005d471f82f6827552a60659ab97984eb1faf074c7b82c2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:32Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.625547 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d5a9ccb-13ba-4dd5-9904-42aac5eca374\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d19bc48acb392051f0f6c5a6820135bf29ebe6d2afd098697553298933cf280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93325a6407389c4a7db0ba3c1a9c84a6980db3c3277cdf3860583a564b562bbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20505e91fc60158eeb3a8cfa0aafc5b4e4436d5def5e735a3a58b0f9f3621058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3281314d85a20945d56b75ab7b7aef1e382a73e74dff03773d118d03eb0ee16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:32Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.640991 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2921ec70-5155-49f0-b2d8-a0e3e21b599c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83ee18494b4658ba014b0013673acea38ee70056ab1f6cd6aa4e173fe8531458\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a6ad38c7fe3f9eb0dfc09bbd22aba8c16d4b51b9bd9142f4204f25b914155c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a6ad38c7fe3f9eb0dfc09bbd22aba8c16d4b51b9bd9142f4204f25b914155c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:32Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.657068 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b33b805b76abbcace899e527d5ee3eaeb9e480fb03dc8eacd3f6dd90d2dd6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:32Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.666023 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.666056 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.666065 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.666078 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.666087 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:32Z","lastTransitionTime":"2025-11-25T15:03:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.669969 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2nlkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16f52653-c067-4186-bc40-9c0f7c383b64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmlnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmlnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2nlkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:32Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.683371 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd5e648-6322-404d-bf1b-3b59a8092291\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76d277b4001db9d63b0af88b98afcc9019321101429c9be47b3022d3575d4e06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e3a186b14ea199d1aeba5751290c266ec918cb055120cfc5c8a2ae03b1469e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a860a998c40aee3e94e4cfd748f54391b3096fd3ed75c84c7ea2a5a7cefa33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://081a6d623cd043c0919d04e2ca93314b2b85fe33048634aa2637297d54b52cee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f506ec649ce24f4df6fd33e4aeafd13e42167dd71323524a5a3164517659fcc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:32Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.696016 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:32Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.709781 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:32Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.725825 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3358a356ae5b422d6bba72bc516374e14ccdace80726aa71b10e25102e0f1100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba3acdae961f49f797f66bbebe350e34a4df5882bcf2a4fe8d37f2ca95865c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:32Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.736895 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8tdrh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1097de4-74e4-4178-b9b4-14a22539da9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfa0291da1598e930aeb4a76f9f4e8b0728001a770acd8f10215465539f8d8b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8tdrh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:32Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.748706 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9762ab15-c9d3-40ac-a789-7b885e5f779f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1154cddf4cb46b8a9db26e7f982ed9dd382d85b5ceb95e50325a475eed408f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca4d48ac675ee73e0946cc802994bb407b6770695320ce2380ddfce26f1b5b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://539bab27d80f8d76cc54a60f508a9985e8a3aa3d2f949ea5a5c2f0d92227da30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8092652bfe2886bf2ea5e6e0cbd6c84efc82b21182a4509a624b399943eede7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8092652bfe2886bf2ea5e6e0cbd6c84efc82b21182a4509a624b399943eede7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:32Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.768195 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.768225 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.768236 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.768253 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.768263 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:32Z","lastTransitionTime":"2025-11-25T15:03:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.871031 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.871624 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.871700 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.871827 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.871886 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:32Z","lastTransitionTime":"2025-11-25T15:03:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.974127 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.974212 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.974223 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.974240 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:32 crc kubenswrapper[4890]: I1125 15:03:32.974251 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:32Z","lastTransitionTime":"2025-11-25T15:03:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:33 crc kubenswrapper[4890]: I1125 15:03:33.077510 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:33 crc kubenswrapper[4890]: I1125 15:03:33.077547 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:33 crc kubenswrapper[4890]: I1125 15:03:33.077559 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:33 crc kubenswrapper[4890]: I1125 15:03:33.077578 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:33 crc kubenswrapper[4890]: I1125 15:03:33.077589 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:33Z","lastTransitionTime":"2025-11-25T15:03:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:33 crc kubenswrapper[4890]: I1125 15:03:33.172070 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nlkp" Nov 25 15:03:33 crc kubenswrapper[4890]: E1125 15:03:33.172507 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nlkp" podUID="16f52653-c067-4186-bc40-9c0f7c383b64" Nov 25 15:03:33 crc kubenswrapper[4890]: I1125 15:03:33.179510 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:33 crc kubenswrapper[4890]: I1125 15:03:33.179549 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:33 crc kubenswrapper[4890]: I1125 15:03:33.179559 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:33 crc kubenswrapper[4890]: I1125 15:03:33.179574 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:33 crc kubenswrapper[4890]: I1125 15:03:33.179584 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:33Z","lastTransitionTime":"2025-11-25T15:03:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:33 crc kubenswrapper[4890]: I1125 15:03:33.281742 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:33 crc kubenswrapper[4890]: I1125 15:03:33.281776 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:33 crc kubenswrapper[4890]: I1125 15:03:33.281784 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:33 crc kubenswrapper[4890]: I1125 15:03:33.281797 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:33 crc kubenswrapper[4890]: I1125 15:03:33.281805 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:33Z","lastTransitionTime":"2025-11-25T15:03:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:33 crc kubenswrapper[4890]: I1125 15:03:33.385764 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:33 crc kubenswrapper[4890]: I1125 15:03:33.385859 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:33 crc kubenswrapper[4890]: I1125 15:03:33.385876 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:33 crc kubenswrapper[4890]: I1125 15:03:33.385895 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:33 crc kubenswrapper[4890]: I1125 15:03:33.385908 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:33Z","lastTransitionTime":"2025-11-25T15:03:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:33 crc kubenswrapper[4890]: I1125 15:03:33.488560 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:33 crc kubenswrapper[4890]: I1125 15:03:33.488862 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:33 crc kubenswrapper[4890]: I1125 15:03:33.488927 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:33 crc kubenswrapper[4890]: I1125 15:03:33.488998 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:33 crc kubenswrapper[4890]: I1125 15:03:33.489080 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:33Z","lastTransitionTime":"2025-11-25T15:03:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:33 crc kubenswrapper[4890]: I1125 15:03:33.591339 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:33 crc kubenswrapper[4890]: I1125 15:03:33.591379 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:33 crc kubenswrapper[4890]: I1125 15:03:33.591387 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:33 crc kubenswrapper[4890]: I1125 15:03:33.591402 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:33 crc kubenswrapper[4890]: I1125 15:03:33.591413 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:33Z","lastTransitionTime":"2025-11-25T15:03:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:33 crc kubenswrapper[4890]: I1125 15:03:33.694285 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:33 crc kubenswrapper[4890]: I1125 15:03:33.694321 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:33 crc kubenswrapper[4890]: I1125 15:03:33.694329 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:33 crc kubenswrapper[4890]: I1125 15:03:33.694347 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:33 crc kubenswrapper[4890]: I1125 15:03:33.694356 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:33Z","lastTransitionTime":"2025-11-25T15:03:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:33 crc kubenswrapper[4890]: I1125 15:03:33.796833 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:33 crc kubenswrapper[4890]: I1125 15:03:33.796873 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:33 crc kubenswrapper[4890]: I1125 15:03:33.796881 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:33 crc kubenswrapper[4890]: I1125 15:03:33.796898 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:33 crc kubenswrapper[4890]: I1125 15:03:33.796907 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:33Z","lastTransitionTime":"2025-11-25T15:03:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:33 crc kubenswrapper[4890]: I1125 15:03:33.899519 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:33 crc kubenswrapper[4890]: I1125 15:03:33.899554 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:33 crc kubenswrapper[4890]: I1125 15:03:33.899563 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:33 crc kubenswrapper[4890]: I1125 15:03:33.899577 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:33 crc kubenswrapper[4890]: I1125 15:03:33.899587 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:33Z","lastTransitionTime":"2025-11-25T15:03:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.002087 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.002134 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.002145 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.002176 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.002188 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:34Z","lastTransitionTime":"2025-11-25T15:03:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.062667 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.063510 4890 scope.go:117] "RemoveContainer" containerID="fc71e729066adb3b4b41639d6028364bc27ae5a938d23ff9b00f7c813f207d8d" Nov 25 15:03:34 crc kubenswrapper[4890]: E1125 15:03:34.063655 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7zdk8_openshift-ovn-kubernetes(9a404958-dffa-4a6a-99fc-3e4950614bd4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" podUID="9a404958-dffa-4a6a-99fc-3e4950614bd4" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.074718 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8tdrh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1097de4-74e4-4178-b9b4-14a22539da9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfa0291da1598e930aeb4a76f9f4e8b0728001a770acd8f10215465539f8d8b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8tdrh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:34Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.086288 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9762ab15-c9d3-40ac-a789-7b885e5f779f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1154cddf4cb46b8a9db26e7f982ed9dd382d85b5ceb95e50325a475eed408f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca4d48ac675ee73e0946cc802994bb407b6770695320ce2380ddfce26f1b5b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://539bab27d80f8d76cc54a60f508a9985e8a3aa3d2f949ea5a5c2f0d92227da30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8092652bfe2886bf2ea5e6e0cbd6c84efc82b21182a4509a624b399943eede7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8092652bfe2886bf2ea5e6e0cbd6c84efc82b21182a4509a624b399943eede7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:34Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.098989 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:34Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.108656 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.108702 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.108714 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.108730 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.108742 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:34Z","lastTransitionTime":"2025-11-25T15:03:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.112089 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:34Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.125994 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3358a356ae5b422d6bba72bc516374e14ccdace80726aa71b10e25102e0f1100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba3acdae961f49f797f66bbebe350e34a4df5882bcf2a4fe8d37f2ca95865c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:34Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.148562 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86e031a7-b4c8-4030-a078-cdca7ccc9dde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4c4921f0973b03f6dade29c2ef32684a481cd1c3e1e23b64ed636943c9bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f9285acc7ac3cbb503a1e546728748c42971c8c9019cc8ca837076f3e55558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa240816a6e9d2821006ce240eccf27d8f57a7f0c5394314fa7177488da4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://659004bd47948def6b33498d8d1474f28c2e98986754f61af012267f3984269e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6d2dcd8df64e44df8cdb4c3a022f03fccb39451631ca35b30d4c7c35271ff0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:34Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.164436 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg454" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81411ec7-f865-48e2-b464-ab6213ef5e52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db59b77d323079e0b588a6adc0f247dda84b5497de17264390366124934cdca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a366b314408eb8453ab7f970a243c511a2716b5472d92b03c81061135e4e59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a366b314408eb8453ab7f970a243c511a2716b5472d92b03c81061135e4e59c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f841cf4a886c5e17d9919787305ce297d7910b2692be53169324a56bd545be5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f841cf4a886c5e17d9919787305ce297d7910b2692be53169324a56bd545be5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e7825cb1d51e854b5218b09bca98a38c75b356ac9d3d6c206147e98b050dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21e7825cb1d51e854b5218b09bca98a38c75b356ac9d3d6c206147e98b050dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5881047b9fa64b7504f1d62b697a6f352b05d14eedd61e6b64d5ff99d91edda2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5881047b9fa64b7504f1d62b697a6f352b05d14eedd61e6b64d5ff99d91edda2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg454\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:34Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.172593 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.172636 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.172616 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 15:03:34 crc kubenswrapper[4890]: E1125 15:03:34.172754 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 15:03:34 crc kubenswrapper[4890]: E1125 15:03:34.172857 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 15:03:34 crc kubenswrapper[4890]: E1125 15:03:34.172923 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.178273 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4txd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4168b721ad650800e545d62906f69c441da51f6d8f11dd2537c8f2b2301afc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4322dbd78237bf443a985ce3c79bca4cba2103862d81177f5474a76f4905eed\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T15:03:30Z\\\",\\\"message\\\":\\\"2025-11-25T15:02:44+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_bd112228-266e-450f-a5e1-285339caf6b6\\\\n2025-11-25T15:02:44+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_bd112228-266e-450f-a5e1-285339caf6b6 to /host/opt/cni/bin/\\\\n2025-11-25T15:02:45Z [verbose] multus-daemon started\\\\n2025-11-25T15:02:45Z [verbose] Readiness Indicator file check\\\\n2025-11-25T15:03:30Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfcq7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4txd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:34Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.190232 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a299b8e0-37bc-48d9-a7e8-64ced2d92d40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5bf7fec98590eb53dd0ecce057c364d4b943bdc0fd152f49a6dfc9586b056f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvcr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90eb11dd969896d6769a18a39ac3c87633b1d42e110093a32bc77bd6f8a7e332\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvcr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6wr7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:34Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.201257 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e4f849d-f239-4727-a73e-18327856929a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://490f1595bc77030d2ee789197fa3748324e1b7a0c9d9180ba08383bbac43adc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0902d6f8e56aecc5a2a13a7148f94a8a00ad139027b3c036aba0215ba0a67f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xx2tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:34Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.210962 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.211001 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.211011 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.211024 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.211033 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:34Z","lastTransitionTime":"2025-11-25T15:03:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.213832 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583964bf462fa0005d471f82f6827552a60659ab97984eb1faf074c7b82c2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:34Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.225154 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:34Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.235497 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7g5s8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a7dc232-835b-491c-afc9-9eb14f257af3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9a7ac317fbb9c589573e80babe484c21bef4da3d9ab79dfb7ebf5b8300c472c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4cs2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7g5s8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:34Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.262718 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a404958-dffa-4a6a-99fc-3e4950614bd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc71e729066adb3b4b41639d6028364bc27ae5a938d23ff9b00f7c813f207d8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc71e729066adb3b4b41639d6028364bc27ae5a938d23ff9b00f7c813f207d8d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T15:03:19Z\\\",\\\"message\\\":\\\"egistry/node-ca-8tdrh in node crc\\\\nI1125 15:03:19.917932 6586 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1125 15:03:19.917938 6586 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-8tdrh after 0 failed attempt(s)\\\\nI1125 15:03:19.917940 6586 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1125 15:03:19.917942 6586 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-8tdrh\\\\nI1125 15:03:19.917945 6586 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1125 15:03:19.917892 6586 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s in node crc\\\\nI1125 15:03:19.917955 6586 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s after 0 failed attempt(s)\\\\nI1125 15:03:19.917960 6586 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s\\\\nI1125 15:03:19.917799 6586 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-zg454\\\\nI1125 15:03:19.917970\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T15:03:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7zdk8_openshift-ovn-kubernetes(9a404958-dffa-4a6a-99fc-3e4950614bd4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zdk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:34Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.274759 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2nlkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16f52653-c067-4186-bc40-9c0f7c383b64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmlnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmlnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2nlkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:34Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.288363 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd5e648-6322-404d-bf1b-3b59a8092291\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76d277b4001db9d63b0af88b98afcc9019321101429c9be47b3022d3575d4e06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e3a186b14ea199d1aeba5751290c266ec918cb055120cfc5c8a2ae03b1469e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a860a998c40aee3e94e4cfd748f54391b3096fd3ed75c84c7ea2a5a7cefa33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://081a6d623cd043c0919d04e2ca93314b2b85fe33048634aa2637297d54b52cee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f506ec649ce24f4df6fd33e4aeafd13e42167dd71323524a5a3164517659fcc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:34Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.301564 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d5a9ccb-13ba-4dd5-9904-42aac5eca374\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d19bc48acb392051f0f6c5a6820135bf29ebe6d2afd098697553298933cf280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93325a6407389c4a7db0ba3c1a9c84a6980db3c3277cdf3860583a564b562bbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20505e91fc60158eeb3a8cfa0aafc5b4e4436d5def5e735a3a58b0f9f3621058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3281314d85a20945d56b75ab7b7aef1e382a73e74dff03773d118d03eb0ee16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:34Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.311539 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2921ec70-5155-49f0-b2d8-a0e3e21b599c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83ee18494b4658ba014b0013673acea38ee70056ab1f6cd6aa4e173fe8531458\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a6ad38c7fe3f9eb0dfc09bbd22aba8c16d4b51b9bd9142f4204f25b914155c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a6ad38c7fe3f9eb0dfc09bbd22aba8c16d4b51b9bd9142f4204f25b914155c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:34Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.313893 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.314092 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.314263 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.314338 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.314406 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:34Z","lastTransitionTime":"2025-11-25T15:03:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.322870 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b33b805b76abbcace899e527d5ee3eaeb9e480fb03dc8eacd3f6dd90d2dd6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:34Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.417009 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.417056 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.417069 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.417085 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.417097 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:34Z","lastTransitionTime":"2025-11-25T15:03:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.519063 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.519099 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.519108 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.519124 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.519134 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:34Z","lastTransitionTime":"2025-11-25T15:03:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.621096 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.621144 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.621152 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.621187 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.621200 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:34Z","lastTransitionTime":"2025-11-25T15:03:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.723735 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.723799 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.723815 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.723839 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.723864 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:34Z","lastTransitionTime":"2025-11-25T15:03:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.826241 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.826273 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.826282 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.826295 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.826302 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:34Z","lastTransitionTime":"2025-11-25T15:03:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.932715 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.932808 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.932827 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.932855 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:34 crc kubenswrapper[4890]: I1125 15:03:34.932878 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:34Z","lastTransitionTime":"2025-11-25T15:03:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.036085 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.036138 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.036179 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.036200 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.036212 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:35Z","lastTransitionTime":"2025-11-25T15:03:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.138922 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.138956 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.138964 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.138979 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.138988 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:35Z","lastTransitionTime":"2025-11-25T15:03:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.172782 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nlkp" Nov 25 15:03:35 crc kubenswrapper[4890]: E1125 15:03:35.172939 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nlkp" podUID="16f52653-c067-4186-bc40-9c0f7c383b64" Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.241152 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.241209 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.241225 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.241243 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.241255 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:35Z","lastTransitionTime":"2025-11-25T15:03:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.343618 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.343668 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.343679 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.343692 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.343702 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:35Z","lastTransitionTime":"2025-11-25T15:03:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.445927 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.445978 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.445990 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.446007 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.446022 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:35Z","lastTransitionTime":"2025-11-25T15:03:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.548643 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.548696 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.548714 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.548739 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.548756 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:35Z","lastTransitionTime":"2025-11-25T15:03:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.651518 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.651579 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.651592 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.651611 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.651624 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:35Z","lastTransitionTime":"2025-11-25T15:03:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.754233 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.754282 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.754292 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.754308 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.754320 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:35Z","lastTransitionTime":"2025-11-25T15:03:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.856444 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.856475 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.856492 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.856510 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.856520 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:35Z","lastTransitionTime":"2025-11-25T15:03:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.958779 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.958838 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.958856 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.958877 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:35 crc kubenswrapper[4890]: I1125 15:03:35.958889 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:35Z","lastTransitionTime":"2025-11-25T15:03:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.061817 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.061867 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.061880 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.061896 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.061907 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:36Z","lastTransitionTime":"2025-11-25T15:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.164536 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.164592 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.164603 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.164624 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.164636 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:36Z","lastTransitionTime":"2025-11-25T15:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.173201 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.173308 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 15:03:36 crc kubenswrapper[4890]: E1125 15:03:36.173414 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 15:03:36 crc kubenswrapper[4890]: E1125 15:03:36.173518 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.173554 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:03:36 crc kubenswrapper[4890]: E1125 15:03:36.173801 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.267579 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.267650 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.267669 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.267691 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.267704 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:36Z","lastTransitionTime":"2025-11-25T15:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.371557 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.371617 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.371627 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.371652 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.371666 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:36Z","lastTransitionTime":"2025-11-25T15:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.474829 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.474892 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.474903 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.474922 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.474939 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:36Z","lastTransitionTime":"2025-11-25T15:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.578323 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.578397 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.578417 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.578449 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.578472 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:36Z","lastTransitionTime":"2025-11-25T15:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.682227 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.682291 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.682304 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.682328 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.682342 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:36Z","lastTransitionTime":"2025-11-25T15:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.785704 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.785757 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.785766 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.785784 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.785794 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:36Z","lastTransitionTime":"2025-11-25T15:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.888241 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.888298 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.888310 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.888328 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.888341 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:36Z","lastTransitionTime":"2025-11-25T15:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.990761 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.990802 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.990810 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.990825 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:36 crc kubenswrapper[4890]: I1125 15:03:36.990834 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:36Z","lastTransitionTime":"2025-11-25T15:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:37 crc kubenswrapper[4890]: I1125 15:03:37.094314 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:37 crc kubenswrapper[4890]: I1125 15:03:37.094385 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:37 crc kubenswrapper[4890]: I1125 15:03:37.094405 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:37 crc kubenswrapper[4890]: I1125 15:03:37.094431 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:37 crc kubenswrapper[4890]: I1125 15:03:37.094448 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:37Z","lastTransitionTime":"2025-11-25T15:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:37 crc kubenswrapper[4890]: I1125 15:03:37.173016 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nlkp" Nov 25 15:03:37 crc kubenswrapper[4890]: E1125 15:03:37.173285 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nlkp" podUID="16f52653-c067-4186-bc40-9c0f7c383b64" Nov 25 15:03:37 crc kubenswrapper[4890]: I1125 15:03:37.197378 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:37 crc kubenswrapper[4890]: I1125 15:03:37.197436 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:37 crc kubenswrapper[4890]: I1125 15:03:37.197454 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:37 crc kubenswrapper[4890]: I1125 15:03:37.197476 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:37 crc kubenswrapper[4890]: I1125 15:03:37.197494 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:37Z","lastTransitionTime":"2025-11-25T15:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:37 crc kubenswrapper[4890]: I1125 15:03:37.300971 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:37 crc kubenswrapper[4890]: I1125 15:03:37.301030 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:37 crc kubenswrapper[4890]: I1125 15:03:37.301042 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:37 crc kubenswrapper[4890]: I1125 15:03:37.301059 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:37 crc kubenswrapper[4890]: I1125 15:03:37.301070 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:37Z","lastTransitionTime":"2025-11-25T15:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:37 crc kubenswrapper[4890]: I1125 15:03:37.404851 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:37 crc kubenswrapper[4890]: I1125 15:03:37.404902 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:37 crc kubenswrapper[4890]: I1125 15:03:37.404910 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:37 crc kubenswrapper[4890]: I1125 15:03:37.404927 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:37 crc kubenswrapper[4890]: I1125 15:03:37.404938 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:37Z","lastTransitionTime":"2025-11-25T15:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:37 crc kubenswrapper[4890]: I1125 15:03:37.508571 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:37 crc kubenswrapper[4890]: I1125 15:03:37.508618 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:37 crc kubenswrapper[4890]: I1125 15:03:37.508632 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:37 crc kubenswrapper[4890]: I1125 15:03:37.508655 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:37 crc kubenswrapper[4890]: I1125 15:03:37.508667 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:37Z","lastTransitionTime":"2025-11-25T15:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:37 crc kubenswrapper[4890]: I1125 15:03:37.612420 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:37 crc kubenswrapper[4890]: I1125 15:03:37.612476 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:37 crc kubenswrapper[4890]: I1125 15:03:37.612488 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:37 crc kubenswrapper[4890]: I1125 15:03:37.612509 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:37 crc kubenswrapper[4890]: I1125 15:03:37.612523 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:37Z","lastTransitionTime":"2025-11-25T15:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:37 crc kubenswrapper[4890]: I1125 15:03:37.715359 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:37 crc kubenswrapper[4890]: I1125 15:03:37.715420 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:37 crc kubenswrapper[4890]: I1125 15:03:37.715431 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:37 crc kubenswrapper[4890]: I1125 15:03:37.715445 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:37 crc kubenswrapper[4890]: I1125 15:03:37.715454 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:37Z","lastTransitionTime":"2025-11-25T15:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:37 crc kubenswrapper[4890]: I1125 15:03:37.819130 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:37 crc kubenswrapper[4890]: I1125 15:03:37.819210 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:37 crc kubenswrapper[4890]: I1125 15:03:37.819223 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:37 crc kubenswrapper[4890]: I1125 15:03:37.819244 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:37 crc kubenswrapper[4890]: I1125 15:03:37.819256 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:37Z","lastTransitionTime":"2025-11-25T15:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:37 crc kubenswrapper[4890]: I1125 15:03:37.922476 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:37 crc kubenswrapper[4890]: I1125 15:03:37.922645 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:37 crc kubenswrapper[4890]: I1125 15:03:37.922665 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:37 crc kubenswrapper[4890]: I1125 15:03:37.922684 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:37 crc kubenswrapper[4890]: I1125 15:03:37.922699 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:37Z","lastTransitionTime":"2025-11-25T15:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.026505 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.026566 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.026576 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.026599 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.026613 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:38Z","lastTransitionTime":"2025-11-25T15:03:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.128996 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.129032 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.129040 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.129055 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.129065 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:38Z","lastTransitionTime":"2025-11-25T15:03:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.172277 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.172278 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 15:03:38 crc kubenswrapper[4890]: E1125 15:03:38.172497 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.172430 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:03:38 crc kubenswrapper[4890]: E1125 15:03:38.172760 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 15:03:38 crc kubenswrapper[4890]: E1125 15:03:38.172807 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.233003 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.233062 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.233077 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.233098 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.233116 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:38Z","lastTransitionTime":"2025-11-25T15:03:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.336487 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.336548 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.336558 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.336577 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.336591 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:38Z","lastTransitionTime":"2025-11-25T15:03:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.439427 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.439502 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.439521 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.439547 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.439565 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:38Z","lastTransitionTime":"2025-11-25T15:03:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.542725 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.542781 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.542810 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.542836 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.542858 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:38Z","lastTransitionTime":"2025-11-25T15:03:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.645710 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.645769 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.645828 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.645863 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.645886 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:38Z","lastTransitionTime":"2025-11-25T15:03:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.748716 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.748787 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.748803 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.748828 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.748843 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:38Z","lastTransitionTime":"2025-11-25T15:03:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.851766 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.851808 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.851819 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.851834 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.851846 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:38Z","lastTransitionTime":"2025-11-25T15:03:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.954668 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.954772 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.954798 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.954834 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:38 crc kubenswrapper[4890]: I1125 15:03:38.954877 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:38Z","lastTransitionTime":"2025-11-25T15:03:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.057682 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.057741 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.057758 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.057781 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.057800 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:39Z","lastTransitionTime":"2025-11-25T15:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.161625 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.161712 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.161728 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.162086 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.162126 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:39Z","lastTransitionTime":"2025-11-25T15:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.172623 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nlkp" Nov 25 15:03:39 crc kubenswrapper[4890]: E1125 15:03:39.172816 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nlkp" podUID="16f52653-c067-4186-bc40-9c0f7c383b64" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.265553 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.265603 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.265618 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.265638 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.265653 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:39Z","lastTransitionTime":"2025-11-25T15:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.358511 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.358581 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.358597 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.358623 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.358642 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:39Z","lastTransitionTime":"2025-11-25T15:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:39 crc kubenswrapper[4890]: E1125 15:03:39.379032 4890 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7c7b4592-d206-4ee7-bf93-ada4b8bc755a\\\",\\\"systemUUID\\\":\\\"5350de30-707e-4bde-b990-1fa65ae3a01e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:39Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.384154 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.384258 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.384277 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.384306 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.384324 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:39Z","lastTransitionTime":"2025-11-25T15:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:39 crc kubenswrapper[4890]: E1125 15:03:39.399783 4890 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7c7b4592-d206-4ee7-bf93-ada4b8bc755a\\\",\\\"systemUUID\\\":\\\"5350de30-707e-4bde-b990-1fa65ae3a01e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:39Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.405400 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.405476 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.405500 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.405531 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.405556 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:39Z","lastTransitionTime":"2025-11-25T15:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:39 crc kubenswrapper[4890]: E1125 15:03:39.423964 4890 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7c7b4592-d206-4ee7-bf93-ada4b8bc755a\\\",\\\"systemUUID\\\":\\\"5350de30-707e-4bde-b990-1fa65ae3a01e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:39Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.429095 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.429145 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.429173 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.429200 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.429213 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:39Z","lastTransitionTime":"2025-11-25T15:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:39 crc kubenswrapper[4890]: E1125 15:03:39.441932 4890 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7c7b4592-d206-4ee7-bf93-ada4b8bc755a\\\",\\\"systemUUID\\\":\\\"5350de30-707e-4bde-b990-1fa65ae3a01e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:39Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.446225 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.446296 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.446310 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.446332 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.446345 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:39Z","lastTransitionTime":"2025-11-25T15:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:39 crc kubenswrapper[4890]: E1125 15:03:39.459546 4890 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T15:03:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7c7b4592-d206-4ee7-bf93-ada4b8bc755a\\\",\\\"systemUUID\\\":\\\"5350de30-707e-4bde-b990-1fa65ae3a01e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:39Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:39 crc kubenswrapper[4890]: E1125 15:03:39.459708 4890 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.461944 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.461994 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.462011 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.462035 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.462054 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:39Z","lastTransitionTime":"2025-11-25T15:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.564304 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.564344 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.564354 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.564373 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.564385 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:39Z","lastTransitionTime":"2025-11-25T15:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.667713 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.667776 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.667789 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.667814 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.667828 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:39Z","lastTransitionTime":"2025-11-25T15:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.770847 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.770909 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.770920 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.770941 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.770956 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:39Z","lastTransitionTime":"2025-11-25T15:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.875592 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.875651 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.875663 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.875691 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.875706 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:39Z","lastTransitionTime":"2025-11-25T15:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.978776 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.978845 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.978857 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.978882 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:39 crc kubenswrapper[4890]: I1125 15:03:39.978897 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:39Z","lastTransitionTime":"2025-11-25T15:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:40 crc kubenswrapper[4890]: I1125 15:03:40.081767 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:40 crc kubenswrapper[4890]: I1125 15:03:40.081840 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:40 crc kubenswrapper[4890]: I1125 15:03:40.081852 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:40 crc kubenswrapper[4890]: I1125 15:03:40.081873 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:40 crc kubenswrapper[4890]: I1125 15:03:40.081887 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:40Z","lastTransitionTime":"2025-11-25T15:03:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:40 crc kubenswrapper[4890]: I1125 15:03:40.172515 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 15:03:40 crc kubenswrapper[4890]: I1125 15:03:40.172607 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 15:03:40 crc kubenswrapper[4890]: I1125 15:03:40.172631 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:03:40 crc kubenswrapper[4890]: E1125 15:03:40.172769 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 15:03:40 crc kubenswrapper[4890]: E1125 15:03:40.172847 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 15:03:40 crc kubenswrapper[4890]: E1125 15:03:40.173039 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 15:03:40 crc kubenswrapper[4890]: I1125 15:03:40.184471 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:40 crc kubenswrapper[4890]: I1125 15:03:40.184544 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:40 crc kubenswrapper[4890]: I1125 15:03:40.184558 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:40 crc kubenswrapper[4890]: I1125 15:03:40.184584 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:40 crc kubenswrapper[4890]: I1125 15:03:40.184598 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:40Z","lastTransitionTime":"2025-11-25T15:03:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:40 crc kubenswrapper[4890]: I1125 15:03:40.287603 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:40 crc kubenswrapper[4890]: I1125 15:03:40.287654 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:40 crc kubenswrapper[4890]: I1125 15:03:40.287664 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:40 crc kubenswrapper[4890]: I1125 15:03:40.287682 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:40 crc kubenswrapper[4890]: I1125 15:03:40.287694 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:40Z","lastTransitionTime":"2025-11-25T15:03:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:40 crc kubenswrapper[4890]: I1125 15:03:40.390964 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:40 crc kubenswrapper[4890]: I1125 15:03:40.391028 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:40 crc kubenswrapper[4890]: I1125 15:03:40.391043 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:40 crc kubenswrapper[4890]: I1125 15:03:40.391069 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:40 crc kubenswrapper[4890]: I1125 15:03:40.391085 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:40Z","lastTransitionTime":"2025-11-25T15:03:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:40 crc kubenswrapper[4890]: I1125 15:03:40.493619 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:40 crc kubenswrapper[4890]: I1125 15:03:40.493655 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:40 crc kubenswrapper[4890]: I1125 15:03:40.493664 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:40 crc kubenswrapper[4890]: I1125 15:03:40.493676 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:40 crc kubenswrapper[4890]: I1125 15:03:40.493685 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:40Z","lastTransitionTime":"2025-11-25T15:03:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:40 crc kubenswrapper[4890]: I1125 15:03:40.597526 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:40 crc kubenswrapper[4890]: I1125 15:03:40.597635 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:40 crc kubenswrapper[4890]: I1125 15:03:40.597652 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:40 crc kubenswrapper[4890]: I1125 15:03:40.597692 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:40 crc kubenswrapper[4890]: I1125 15:03:40.597706 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:40Z","lastTransitionTime":"2025-11-25T15:03:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:40 crc kubenswrapper[4890]: I1125 15:03:40.701630 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:40 crc kubenswrapper[4890]: I1125 15:03:40.701720 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:40 crc kubenswrapper[4890]: I1125 15:03:40.701737 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:40 crc kubenswrapper[4890]: I1125 15:03:40.701784 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:40 crc kubenswrapper[4890]: I1125 15:03:40.701803 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:40Z","lastTransitionTime":"2025-11-25T15:03:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:40 crc kubenswrapper[4890]: I1125 15:03:40.806969 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:40 crc kubenswrapper[4890]: I1125 15:03:40.807053 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:40 crc kubenswrapper[4890]: I1125 15:03:40.807075 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:40 crc kubenswrapper[4890]: I1125 15:03:40.807105 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:40 crc kubenswrapper[4890]: I1125 15:03:40.807139 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:40Z","lastTransitionTime":"2025-11-25T15:03:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:40 crc kubenswrapper[4890]: I1125 15:03:40.910858 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:40 crc kubenswrapper[4890]: I1125 15:03:40.910912 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:40 crc kubenswrapper[4890]: I1125 15:03:40.910921 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:40 crc kubenswrapper[4890]: I1125 15:03:40.910940 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:40 crc kubenswrapper[4890]: I1125 15:03:40.910949 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:40Z","lastTransitionTime":"2025-11-25T15:03:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.014446 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.014509 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.014529 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.014556 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.014574 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:41Z","lastTransitionTime":"2025-11-25T15:03:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.119292 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.119390 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.119414 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.119449 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.119473 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:41Z","lastTransitionTime":"2025-11-25T15:03:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.172768 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nlkp" Nov 25 15:03:41 crc kubenswrapper[4890]: E1125 15:03:41.173105 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nlkp" podUID="16f52653-c067-4186-bc40-9c0f7c383b64" Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.222564 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.222630 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.222642 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.222660 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.222675 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:41Z","lastTransitionTime":"2025-11-25T15:03:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.325655 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.325696 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.325708 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.325723 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.325734 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:41Z","lastTransitionTime":"2025-11-25T15:03:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.428997 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.429087 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.429097 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.429119 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.429139 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:41Z","lastTransitionTime":"2025-11-25T15:03:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.532223 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.532306 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.532329 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.532366 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.532391 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:41Z","lastTransitionTime":"2025-11-25T15:03:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.635229 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.635299 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.635322 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.635361 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.635386 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:41Z","lastTransitionTime":"2025-11-25T15:03:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.739908 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.740018 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.740037 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.740071 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.740102 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:41Z","lastTransitionTime":"2025-11-25T15:03:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.843229 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.843264 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.843273 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.843286 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.843297 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:41Z","lastTransitionTime":"2025-11-25T15:03:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.946601 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.946668 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.946686 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.946719 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:41 crc kubenswrapper[4890]: I1125 15:03:41.946735 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:41Z","lastTransitionTime":"2025-11-25T15:03:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.051090 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.051180 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.051193 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.051216 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.051229 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:42Z","lastTransitionTime":"2025-11-25T15:03:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.153584 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.153661 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.153672 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.153693 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.153707 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:42Z","lastTransitionTime":"2025-11-25T15:03:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.172589 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.172699 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.172589 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 15:03:42 crc kubenswrapper[4890]: E1125 15:03:42.172878 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 15:03:42 crc kubenswrapper[4890]: E1125 15:03:42.173018 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 15:03:42 crc kubenswrapper[4890]: E1125 15:03:42.173306 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.192416 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zg454" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81411ec7-f865-48e2-b464-ab6213ef5e52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db59b77d323079e0b588a6adc0f247dda84b5497de17264390366124934cdca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ad72107e6fc557c38e1287caa640cf6280f15b8cec1d65fb6cfdea4a29dae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e502530ace8ae2c0cff8f459599185c062b77b953eccfbf2c68d3cbefc3821b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a366b314408eb8453ab7f970a243c511a2716b5472d92b03c81061135e4e59c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a366b314408eb8453ab7f970a243c511a2716b5472d92b03c81061135e4e59c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f841cf4a886c5e17d9919787305ce297d7910b2692be53169324a56bd545be5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f841cf4a886c5e17d9919787305ce297d7910b2692be53169324a56bd545be5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21e7825cb1d51e854b5218b09bca98a38c75b356ac9d3d6c206147e98b050dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21e7825cb1d51e854b5218b09bca98a38c75b356ac9d3d6c206147e98b050dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5881047b9fa64b7504f1d62b697a6f352b05d14eedd61e6b64d5ff99d91edda2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5881047b9fa64b7504f1d62b697a6f352b05d14eedd61e6b64d5ff99d91edda2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4g7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zg454\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:42Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.212306 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4txd2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcdeaed-8701-42ad-8534-a77b7f3f93f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4168b721ad650800e545d62906f69c441da51f6d8f11dd2537c8f2b2301afc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4322dbd78237bf443a985ce3c79bca4cba2103862d81177f5474a76f4905eed\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T15:03:30Z\\\",\\\"message\\\":\\\"2025-11-25T15:02:44+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_bd112228-266e-450f-a5e1-285339caf6b6\\\\n2025-11-25T15:02:44+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_bd112228-266e-450f-a5e1-285339caf6b6 to /host/opt/cni/bin/\\\\n2025-11-25T15:02:45Z [verbose] multus-daemon started\\\\n2025-11-25T15:02:45Z [verbose] Readiness Indicator file check\\\\n2025-11-25T15:03:30Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfcq7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4txd2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:42Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.229843 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a299b8e0-37bc-48d9-a7e8-64ced2d92d40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5bf7fec98590eb53dd0ecce057c364d4b943bdc0fd152f49a6dfc9586b056f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvcr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90eb11dd969896d6769a18a39ac3c87633b1d42e110093a32bc77bd6f8a7e332\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvcr5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6wr7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:42Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.252563 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86e031a7-b4c8-4030-a078-cdca7ccc9dde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4c4921f0973b03f6dade29c2ef32684a481cd1c3e1e23b64ed636943c9bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f9285acc7ac3cbb503a1e546728748c42971c8c9019cc8ca837076f3e55558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa240816a6e9d2821006ce240eccf27d8f57a7f0c5394314fa7177488da4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://659004bd47948def6b33498d8d1474f28c2e98986754f61af012267f3984269e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6d2dcd8df64e44df8cdb4c3a022f03fccb39451631ca35b30d4c7c35271ff0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c56e1f116d9e32d7966c4805162a371c9d0a025ba2702dc49ad6c4600276f325\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d390debfab62a403c573ed7154fe47463d9090566cc768e9ce0e51b1001c787\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b16c153eba9cd92e4b9edd924e213bda864d3968f41b1e4996f07986f635f17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:42Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.257723 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.257776 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.257791 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.257817 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.257834 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:42Z","lastTransitionTime":"2025-11-25T15:03:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.268897 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:42Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.280813 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7g5s8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a7dc232-835b-491c-afc9-9eb14f257af3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9a7ac317fbb9c589573e80babe484c21bef4da3d9ab79dfb7ebf5b8300c472c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4cs2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7g5s8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:42Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.311030 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a404958-dffa-4a6a-99fc-3e4950614bd4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc71e729066adb3b4b41639d6028364bc27ae5a938d23ff9b00f7c813f207d8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc71e729066adb3b4b41639d6028364bc27ae5a938d23ff9b00f7c813f207d8d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T15:03:19Z\\\",\\\"message\\\":\\\"egistry/node-ca-8tdrh in node crc\\\\nI1125 15:03:19.917932 6586 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1125 15:03:19.917938 6586 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-8tdrh after 0 failed attempt(s)\\\\nI1125 15:03:19.917940 6586 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1125 15:03:19.917942 6586 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-8tdrh\\\\nI1125 15:03:19.917945 6586 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1125 15:03:19.917892 6586 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s in node crc\\\\nI1125 15:03:19.917955 6586 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s after 0 failed attempt(s)\\\\nI1125 15:03:19.917960 6586 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s\\\\nI1125 15:03:19.917799 6586 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-zg454\\\\nI1125 15:03:19.917970\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T15:03:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7zdk8_openshift-ovn-kubernetes(9a404958-dffa-4a6a-99fc-3e4950614bd4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-92hnd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7zdk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:42Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.324650 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e4f849d-f239-4727-a73e-18327856929a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://490f1595bc77030d2ee789197fa3748324e1b7a0c9d9180ba08383bbac43adc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0902d6f8e56aecc5a2a13a7148f94a8a00ad139027b3c036aba0215ba0a67f59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x828g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xx2tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:42Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.340589 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583964bf462fa0005d471f82f6827552a60659ab97984eb1faf074c7b82c2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:42Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.357883 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d5a9ccb-13ba-4dd5-9904-42aac5eca374\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d19bc48acb392051f0f6c5a6820135bf29ebe6d2afd098697553298933cf280d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93325a6407389c4a7db0ba3c1a9c84a6980db3c3277cdf3860583a564b562bbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20505e91fc60158eeb3a8cfa0aafc5b4e4436d5def5e735a3a58b0f9f3621058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3281314d85a20945d56b75ab7b7aef1e382a73e74dff03773d118d03eb0ee16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:42Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.361200 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.361244 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.361286 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.361307 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.361318 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:42Z","lastTransitionTime":"2025-11-25T15:03:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.370304 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2921ec70-5155-49f0-b2d8-a0e3e21b599c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83ee18494b4658ba014b0013673acea38ee70056ab1f6cd6aa4e173fe8531458\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a6ad38c7fe3f9eb0dfc09bbd22aba8c16d4b51b9bd9142f4204f25b914155c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a6ad38c7fe3f9eb0dfc09bbd22aba8c16d4b51b9bd9142f4204f25b914155c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:42Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.382487 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b33b805b76abbcace899e527d5ee3eaeb9e480fb03dc8eacd3f6dd90d2dd6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:42Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.393034 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2nlkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16f52653-c067-4186-bc40-9c0f7c383b64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmlnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmlnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2nlkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:42Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.407318 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bd5e648-6322-404d-bf1b-3b59a8092291\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76d277b4001db9d63b0af88b98afcc9019321101429c9be47b3022d3575d4e06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e3a186b14ea199d1aeba5751290c266ec918cb055120cfc5c8a2ae03b1469e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a860a998c40aee3e94e4cfd748f54391b3096fd3ed75c84c7ea2a5a7cefa33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://081a6d623cd043c0919d04e2ca93314b2b85fe33048634aa2637297d54b52cee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f506ec649ce24f4df6fd33e4aeafd13e42167dd71323524a5a3164517659fcc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:42Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.424848 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:42Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.439523 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:42Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.455264 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3358a356ae5b422d6bba72bc516374e14ccdace80726aa71b10e25102e0f1100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba3acdae961f49f797f66bbebe350e34a4df5882bcf2a4fe8d37f2ca95865c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:42Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.464474 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.464509 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.464518 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.464535 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.464546 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:42Z","lastTransitionTime":"2025-11-25T15:03:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.466823 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8tdrh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1097de4-74e4-4178-b9b4-14a22539da9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfa0291da1598e930aeb4a76f9f4e8b0728001a770acd8f10215465539f8d8b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbz6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8tdrh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:42Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.479356 4890 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9762ab15-c9d3-40ac-a789-7b885e5f779f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:03:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T15:02:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1154cddf4cb46b8a9db26e7f982ed9dd382d85b5ceb95e50325a475eed408f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca4d48ac675ee73e0946cc802994bb407b6770695320ce2380ddfce26f1b5b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://539bab27d80f8d76cc54a60f508a9985e8a3aa3d2f949ea5a5c2f0d92227da30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T15:02:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8092652bfe2886bf2ea5e6e0cbd6c84efc82b21182a4509a624b399943eede7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8092652bfe2886bf2ea5e6e0cbd6c84efc82b21182a4509a624b399943eede7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T15:02:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T15:02:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T15:02:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T15:03:42Z is after 2025-08-24T17:21:41Z" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.567296 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.567350 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.567368 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.567395 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.567413 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:42Z","lastTransitionTime":"2025-11-25T15:03:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.670757 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.670825 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.670838 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.670858 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.670874 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:42Z","lastTransitionTime":"2025-11-25T15:03:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.774224 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.774566 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.774923 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.775047 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.775128 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:42Z","lastTransitionTime":"2025-11-25T15:03:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.878000 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.878082 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.878104 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.878134 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.878181 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:42Z","lastTransitionTime":"2025-11-25T15:03:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.981703 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.981777 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.981794 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.981822 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:42 crc kubenswrapper[4890]: I1125 15:03:42.981840 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:42Z","lastTransitionTime":"2025-11-25T15:03:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:43 crc kubenswrapper[4890]: I1125 15:03:43.084480 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:43 crc kubenswrapper[4890]: I1125 15:03:43.084523 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:43 crc kubenswrapper[4890]: I1125 15:03:43.084534 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:43 crc kubenswrapper[4890]: I1125 15:03:43.084551 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:43 crc kubenswrapper[4890]: I1125 15:03:43.084561 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:43Z","lastTransitionTime":"2025-11-25T15:03:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:43 crc kubenswrapper[4890]: I1125 15:03:43.172407 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nlkp" Nov 25 15:03:43 crc kubenswrapper[4890]: E1125 15:03:43.172528 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nlkp" podUID="16f52653-c067-4186-bc40-9c0f7c383b64" Nov 25 15:03:43 crc kubenswrapper[4890]: I1125 15:03:43.187975 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:43 crc kubenswrapper[4890]: I1125 15:03:43.188012 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:43 crc kubenswrapper[4890]: I1125 15:03:43.188024 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:43 crc kubenswrapper[4890]: I1125 15:03:43.188039 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:43 crc kubenswrapper[4890]: I1125 15:03:43.188049 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:43Z","lastTransitionTime":"2025-11-25T15:03:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:43 crc kubenswrapper[4890]: I1125 15:03:43.291097 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:43 crc kubenswrapper[4890]: I1125 15:03:43.291138 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:43 crc kubenswrapper[4890]: I1125 15:03:43.291147 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:43 crc kubenswrapper[4890]: I1125 15:03:43.291179 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:43 crc kubenswrapper[4890]: I1125 15:03:43.291189 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:43Z","lastTransitionTime":"2025-11-25T15:03:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:43 crc kubenswrapper[4890]: I1125 15:03:43.395934 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:43 crc kubenswrapper[4890]: I1125 15:03:43.396019 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:43 crc kubenswrapper[4890]: I1125 15:03:43.396031 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:43 crc kubenswrapper[4890]: I1125 15:03:43.396054 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:43 crc kubenswrapper[4890]: I1125 15:03:43.396070 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:43Z","lastTransitionTime":"2025-11-25T15:03:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:43 crc kubenswrapper[4890]: I1125 15:03:43.498674 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:43 crc kubenswrapper[4890]: I1125 15:03:43.498720 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:43 crc kubenswrapper[4890]: I1125 15:03:43.498731 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:43 crc kubenswrapper[4890]: I1125 15:03:43.498751 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:43 crc kubenswrapper[4890]: I1125 15:03:43.498763 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:43Z","lastTransitionTime":"2025-11-25T15:03:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:43 crc kubenswrapper[4890]: I1125 15:03:43.601288 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:43 crc kubenswrapper[4890]: I1125 15:03:43.601330 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:43 crc kubenswrapper[4890]: I1125 15:03:43.601341 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:43 crc kubenswrapper[4890]: I1125 15:03:43.601356 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:43 crc kubenswrapper[4890]: I1125 15:03:43.601366 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:43Z","lastTransitionTime":"2025-11-25T15:03:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:43 crc kubenswrapper[4890]: I1125 15:03:43.704373 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:43 crc kubenswrapper[4890]: I1125 15:03:43.704484 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:43 crc kubenswrapper[4890]: I1125 15:03:43.704533 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:43 crc kubenswrapper[4890]: I1125 15:03:43.704565 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:43 crc kubenswrapper[4890]: I1125 15:03:43.704585 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:43Z","lastTransitionTime":"2025-11-25T15:03:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:43 crc kubenswrapper[4890]: I1125 15:03:43.807700 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:43 crc kubenswrapper[4890]: I1125 15:03:43.807773 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:43 crc kubenswrapper[4890]: I1125 15:03:43.807785 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:43 crc kubenswrapper[4890]: I1125 15:03:43.807805 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:43 crc kubenswrapper[4890]: I1125 15:03:43.807820 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:43Z","lastTransitionTime":"2025-11-25T15:03:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:43 crc kubenswrapper[4890]: I1125 15:03:43.911309 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:43 crc kubenswrapper[4890]: I1125 15:03:43.911359 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:43 crc kubenswrapper[4890]: I1125 15:03:43.911378 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:43 crc kubenswrapper[4890]: I1125 15:03:43.911398 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:43 crc kubenswrapper[4890]: I1125 15:03:43.911411 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:43Z","lastTransitionTime":"2025-11-25T15:03:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.014723 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.014772 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.014783 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.014800 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.014811 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:44Z","lastTransitionTime":"2025-11-25T15:03:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.117857 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.117907 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.117919 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.117941 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.117954 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:44Z","lastTransitionTime":"2025-11-25T15:03:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.172289 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.172367 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.172394 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:03:44 crc kubenswrapper[4890]: E1125 15:03:44.172448 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 15:03:44 crc kubenswrapper[4890]: E1125 15:03:44.172789 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 15:03:44 crc kubenswrapper[4890]: E1125 15:03:44.172915 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.220698 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.220749 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.220759 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.220775 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.220787 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:44Z","lastTransitionTime":"2025-11-25T15:03:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.323137 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.323200 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.323208 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.323222 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.323231 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:44Z","lastTransitionTime":"2025-11-25T15:03:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.425116 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.425181 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.425195 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.425212 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.425222 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:44Z","lastTransitionTime":"2025-11-25T15:03:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.527799 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.527842 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.527851 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.527867 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.527878 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:44Z","lastTransitionTime":"2025-11-25T15:03:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.630336 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.630396 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.630412 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.630440 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.630457 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:44Z","lastTransitionTime":"2025-11-25T15:03:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.732954 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.733012 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.733021 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.733035 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.733046 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:44Z","lastTransitionTime":"2025-11-25T15:03:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.834797 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.834844 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.834869 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.834885 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.834895 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:44Z","lastTransitionTime":"2025-11-25T15:03:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.937612 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.937646 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.937656 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.937669 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:44 crc kubenswrapper[4890]: I1125 15:03:44.937678 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:44Z","lastTransitionTime":"2025-11-25T15:03:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.039504 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.039537 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.039546 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.039559 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.039567 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:45Z","lastTransitionTime":"2025-11-25T15:03:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.142555 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.142638 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.142650 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.142667 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.142694 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:45Z","lastTransitionTime":"2025-11-25T15:03:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.173005 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nlkp" Nov 25 15:03:45 crc kubenswrapper[4890]: E1125 15:03:45.173136 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nlkp" podUID="16f52653-c067-4186-bc40-9c0f7c383b64" Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.246028 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.246122 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.246195 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.246226 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.246270 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:45Z","lastTransitionTime":"2025-11-25T15:03:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.309283 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.309612 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.309666 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:03:45 crc kubenswrapper[4890]: E1125 15:03:45.309763 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:04:49.309677511 +0000 UTC m=+147.752140151 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:45 crc kubenswrapper[4890]: E1125 15:03:45.309796 4890 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 15:03:45 crc kubenswrapper[4890]: E1125 15:03:45.309917 4890 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.309927 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:03:45 crc kubenswrapper[4890]: E1125 15:03:45.309954 4890 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 15:03:45 crc kubenswrapper[4890]: E1125 15:03:45.309973 4890 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 15:03:45 crc kubenswrapper[4890]: E1125 15:03:45.310043 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 15:04:49.309978579 +0000 UTC m=+147.752441379 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.310075 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 15:03:45 crc kubenswrapper[4890]: E1125 15:03:45.310100 4890 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 15:03:45 crc kubenswrapper[4890]: E1125 15:03:45.310206 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 15:04:49.310185515 +0000 UTC m=+147.752648315 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 15:03:45 crc kubenswrapper[4890]: E1125 15:03:45.310301 4890 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 15:03:45 crc kubenswrapper[4890]: E1125 15:03:45.310339 4890 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 15:03:45 crc kubenswrapper[4890]: E1125 15:03:45.310364 4890 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 15:03:45 crc kubenswrapper[4890]: E1125 15:03:45.310415 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 15:04:49.310319358 +0000 UTC m=+147.752781968 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 15:03:45 crc kubenswrapper[4890]: E1125 15:03:45.310477 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 15:04:49.310468022 +0000 UTC m=+147.752930632 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.349806 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.349893 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.349920 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.349960 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.349986 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:45Z","lastTransitionTime":"2025-11-25T15:03:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.453782 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.453846 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.453862 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.453884 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.453903 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:45Z","lastTransitionTime":"2025-11-25T15:03:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.556901 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.556938 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.556949 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.556964 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.556978 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:45Z","lastTransitionTime":"2025-11-25T15:03:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.659973 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.660018 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.660031 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.660052 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.660064 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:45Z","lastTransitionTime":"2025-11-25T15:03:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.764195 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.764261 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.764275 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.764301 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.764316 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:45Z","lastTransitionTime":"2025-11-25T15:03:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.867487 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.867552 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.867571 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.867599 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.867623 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:45Z","lastTransitionTime":"2025-11-25T15:03:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.971131 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.971219 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.971235 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.971260 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:45 crc kubenswrapper[4890]: I1125 15:03:45.971276 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:45Z","lastTransitionTime":"2025-11-25T15:03:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:46 crc kubenswrapper[4890]: I1125 15:03:46.074876 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:46 crc kubenswrapper[4890]: I1125 15:03:46.074957 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:46 crc kubenswrapper[4890]: I1125 15:03:46.074981 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:46 crc kubenswrapper[4890]: I1125 15:03:46.075007 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:46 crc kubenswrapper[4890]: I1125 15:03:46.075030 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:46Z","lastTransitionTime":"2025-11-25T15:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:46 crc kubenswrapper[4890]: I1125 15:03:46.172411 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 15:03:46 crc kubenswrapper[4890]: I1125 15:03:46.172411 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 15:03:46 crc kubenswrapper[4890]: I1125 15:03:46.172751 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:03:46 crc kubenswrapper[4890]: E1125 15:03:46.172880 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 15:03:46 crc kubenswrapper[4890]: E1125 15:03:46.173548 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 15:03:46 crc kubenswrapper[4890]: E1125 15:03:46.173666 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 15:03:46 crc kubenswrapper[4890]: I1125 15:03:46.177534 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:46 crc kubenswrapper[4890]: I1125 15:03:46.177582 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:46 crc kubenswrapper[4890]: I1125 15:03:46.177600 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:46 crc kubenswrapper[4890]: I1125 15:03:46.177621 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:46 crc kubenswrapper[4890]: I1125 15:03:46.177639 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:46Z","lastTransitionTime":"2025-11-25T15:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:46 crc kubenswrapper[4890]: I1125 15:03:46.280930 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:46 crc kubenswrapper[4890]: I1125 15:03:46.281004 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:46 crc kubenswrapper[4890]: I1125 15:03:46.281016 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:46 crc kubenswrapper[4890]: I1125 15:03:46.281099 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:46 crc kubenswrapper[4890]: I1125 15:03:46.281115 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:46Z","lastTransitionTime":"2025-11-25T15:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:46 crc kubenswrapper[4890]: I1125 15:03:46.387782 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:46 crc kubenswrapper[4890]: I1125 15:03:46.388214 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:46 crc kubenswrapper[4890]: I1125 15:03:46.388224 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:46 crc kubenswrapper[4890]: I1125 15:03:46.388240 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:46 crc kubenswrapper[4890]: I1125 15:03:46.388250 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:46Z","lastTransitionTime":"2025-11-25T15:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:46 crc kubenswrapper[4890]: I1125 15:03:46.490191 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:46 crc kubenswrapper[4890]: I1125 15:03:46.490226 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:46 crc kubenswrapper[4890]: I1125 15:03:46.490234 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:46 crc kubenswrapper[4890]: I1125 15:03:46.490275 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:46 crc kubenswrapper[4890]: I1125 15:03:46.490290 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:46Z","lastTransitionTime":"2025-11-25T15:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:46 crc kubenswrapper[4890]: I1125 15:03:46.593130 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:46 crc kubenswrapper[4890]: I1125 15:03:46.593185 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:46 crc kubenswrapper[4890]: I1125 15:03:46.593194 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:46 crc kubenswrapper[4890]: I1125 15:03:46.593208 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:46 crc kubenswrapper[4890]: I1125 15:03:46.593217 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:46Z","lastTransitionTime":"2025-11-25T15:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:46 crc kubenswrapper[4890]: I1125 15:03:46.695744 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:46 crc kubenswrapper[4890]: I1125 15:03:46.695784 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:46 crc kubenswrapper[4890]: I1125 15:03:46.695793 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:46 crc kubenswrapper[4890]: I1125 15:03:46.695805 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:46 crc kubenswrapper[4890]: I1125 15:03:46.695814 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:46Z","lastTransitionTime":"2025-11-25T15:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:46 crc kubenswrapper[4890]: I1125 15:03:46.798761 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:46 crc kubenswrapper[4890]: I1125 15:03:46.798812 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:46 crc kubenswrapper[4890]: I1125 15:03:46.798824 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:46 crc kubenswrapper[4890]: I1125 15:03:46.798843 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:46 crc kubenswrapper[4890]: I1125 15:03:46.798854 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:46Z","lastTransitionTime":"2025-11-25T15:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:46 crc kubenswrapper[4890]: I1125 15:03:46.901988 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:46 crc kubenswrapper[4890]: I1125 15:03:46.902026 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:46 crc kubenswrapper[4890]: I1125 15:03:46.902035 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:46 crc kubenswrapper[4890]: I1125 15:03:46.902050 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:46 crc kubenswrapper[4890]: I1125 15:03:46.902063 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:46Z","lastTransitionTime":"2025-11-25T15:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.005110 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.005215 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.005231 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.005259 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.005274 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:47Z","lastTransitionTime":"2025-11-25T15:03:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.109581 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.109625 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.109639 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.109655 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.109667 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:47Z","lastTransitionTime":"2025-11-25T15:03:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.172862 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nlkp" Nov 25 15:03:47 crc kubenswrapper[4890]: E1125 15:03:47.173449 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nlkp" podUID="16f52653-c067-4186-bc40-9c0f7c383b64" Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.212652 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.212702 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.212713 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.212732 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.212742 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:47Z","lastTransitionTime":"2025-11-25T15:03:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.314969 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.315007 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.315017 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.315034 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.315045 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:47Z","lastTransitionTime":"2025-11-25T15:03:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.417424 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.417480 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.417496 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.417516 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.417532 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:47Z","lastTransitionTime":"2025-11-25T15:03:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.519348 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.519396 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.519411 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.519431 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.519446 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:47Z","lastTransitionTime":"2025-11-25T15:03:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.621200 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.621251 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.621262 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.621276 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.621285 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:47Z","lastTransitionTime":"2025-11-25T15:03:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.723591 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.723635 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.723644 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.723658 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.723669 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:47Z","lastTransitionTime":"2025-11-25T15:03:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.825717 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.825775 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.825793 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.825811 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.825823 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:47Z","lastTransitionTime":"2025-11-25T15:03:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.928613 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.928654 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.928664 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.928679 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:47 crc kubenswrapper[4890]: I1125 15:03:47.928689 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:47Z","lastTransitionTime":"2025-11-25T15:03:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.030965 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.031018 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.031034 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.031052 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.031063 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:48Z","lastTransitionTime":"2025-11-25T15:03:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.133725 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.133757 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.133765 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.133779 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.133788 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:48Z","lastTransitionTime":"2025-11-25T15:03:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.172762 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.172786 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.173000 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:03:48 crc kubenswrapper[4890]: E1125 15:03:48.172994 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 15:03:48 crc kubenswrapper[4890]: E1125 15:03:48.173109 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 15:03:48 crc kubenswrapper[4890]: E1125 15:03:48.173737 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.174021 4890 scope.go:117] "RemoveContainer" containerID="fc71e729066adb3b4b41639d6028364bc27ae5a938d23ff9b00f7c813f207d8d" Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.236453 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.236496 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.236506 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.236524 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.236537 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:48Z","lastTransitionTime":"2025-11-25T15:03:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.338983 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.339022 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.339034 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.339050 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.339061 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:48Z","lastTransitionTime":"2025-11-25T15:03:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.441774 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.442362 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.442380 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.442401 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.442415 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:48Z","lastTransitionTime":"2025-11-25T15:03:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.549542 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.549583 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.549592 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.549607 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.549617 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:48Z","lastTransitionTime":"2025-11-25T15:03:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.651928 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.651970 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.651979 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.651993 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.652003 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:48Z","lastTransitionTime":"2025-11-25T15:03:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.753637 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.753669 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.753677 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.753689 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.753698 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:48Z","lastTransitionTime":"2025-11-25T15:03:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.855817 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.855854 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.855865 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.855881 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.855892 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:48Z","lastTransitionTime":"2025-11-25T15:03:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.957828 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.957866 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.957879 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.957897 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:48 crc kubenswrapper[4890]: I1125 15:03:48.957909 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:48Z","lastTransitionTime":"2025-11-25T15:03:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.003104 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-2nlkp"] Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.003243 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nlkp" Nov 25 15:03:49 crc kubenswrapper[4890]: E1125 15:03:49.003332 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nlkp" podUID="16f52653-c067-4186-bc40-9c0f7c383b64" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.059743 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.059792 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.059802 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.059816 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.059827 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:49Z","lastTransitionTime":"2025-11-25T15:03:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.162492 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.162523 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.162533 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.162545 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.162553 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:49Z","lastTransitionTime":"2025-11-25T15:03:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.227218 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7zdk8_9a404958-dffa-4a6a-99fc-3e4950614bd4/ovnkube-controller/2.log" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.230241 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" event={"ID":"9a404958-dffa-4a6a-99fc-3e4950614bd4","Type":"ContainerStarted","Data":"10d43df056314c8846655bab7a1e130886141cb09aee7b55ffd6845e3e982a19"} Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.230786 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.265142 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.265203 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.265215 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.265234 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.265248 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:49Z","lastTransitionTime":"2025-11-25T15:03:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.306085 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-7g5s8" podStartSLOduration=68.306067122 podStartE2EDuration="1m8.306067122s" podCreationTimestamp="2025-11-25 15:02:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:49.283867515 +0000 UTC m=+87.726330125" watchObservedRunningTime="2025-11-25 15:03:49.306067122 +0000 UTC m=+87.748529732" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.306459 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" podStartSLOduration=68.306452823 podStartE2EDuration="1m8.306452823s" podCreationTimestamp="2025-11-25 15:02:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:49.304832299 +0000 UTC m=+87.747294909" watchObservedRunningTime="2025-11-25 15:03:49.306452823 +0000 UTC m=+87.748915433" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.318487 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podStartSLOduration=68.318470952 podStartE2EDuration="1m8.318470952s" podCreationTimestamp="2025-11-25 15:02:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:49.317380022 +0000 UTC m=+87.759842652" watchObservedRunningTime="2025-11-25 15:03:49.318470952 +0000 UTC m=+87.760933562" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.367296 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.367343 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.367356 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.367374 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.367385 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:49Z","lastTransitionTime":"2025-11-25T15:03:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.369022 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=69.369008224 podStartE2EDuration="1m9.369008224s" podCreationTimestamp="2025-11-25 15:02:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:49.347071574 +0000 UTC m=+87.789534184" watchObservedRunningTime="2025-11-25 15:03:49.369008224 +0000 UTC m=+87.811470864" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.369150 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=68.369144938 podStartE2EDuration="1m8.369144938s" podCreationTimestamp="2025-11-25 15:02:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:49.368719207 +0000 UTC m=+87.811181817" watchObservedRunningTime="2025-11-25 15:03:49.369144938 +0000 UTC m=+87.811607548" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.386152 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=18.386129333 podStartE2EDuration="18.386129333s" podCreationTimestamp="2025-11-25 15:03:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:49.380572061 +0000 UTC m=+87.823034671" watchObservedRunningTime="2025-11-25 15:03:49.386129333 +0000 UTC m=+87.828591943" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.424095 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=35.424077181 podStartE2EDuration="35.424077181s" podCreationTimestamp="2025-11-25 15:03:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:49.423971168 +0000 UTC m=+87.866433788" watchObservedRunningTime="2025-11-25 15:03:49.424077181 +0000 UTC m=+87.866539791" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.470369 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.470406 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.470415 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.470430 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.470440 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:49Z","lastTransitionTime":"2025-11-25T15:03:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.482322 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-8tdrh" podStartSLOduration=68.482300234 podStartE2EDuration="1m8.482300234s" podCreationTimestamp="2025-11-25 15:02:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:49.482210472 +0000 UTC m=+87.924673092" watchObservedRunningTime="2025-11-25 15:03:49.482300234 +0000 UTC m=+87.924762844" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.507885 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=68.507868094 podStartE2EDuration="1m8.507868094s" podCreationTimestamp="2025-11-25 15:02:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:49.506709102 +0000 UTC m=+87.949171712" watchObservedRunningTime="2025-11-25 15:03:49.507868094 +0000 UTC m=+87.950330704" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.525434 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-zg454" podStartSLOduration=68.525417974 podStartE2EDuration="1m8.525417974s" podCreationTimestamp="2025-11-25 15:02:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:49.523846981 +0000 UTC m=+87.966309591" watchObservedRunningTime="2025-11-25 15:03:49.525417974 +0000 UTC m=+87.967880584" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.537392 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-4txd2" podStartSLOduration=68.537372161 podStartE2EDuration="1m8.537372161s" podCreationTimestamp="2025-11-25 15:02:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:49.536856457 +0000 UTC m=+87.979319067" watchObservedRunningTime="2025-11-25 15:03:49.537372161 +0000 UTC m=+87.979834771" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.573121 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.573201 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.573215 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.573234 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.573247 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:49Z","lastTransitionTime":"2025-11-25T15:03:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.675438 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.675464 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.675473 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.675486 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.675511 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:49Z","lastTransitionTime":"2025-11-25T15:03:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.777410 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.777449 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.777459 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.777473 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.777484 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:49Z","lastTransitionTime":"2025-11-25T15:03:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.840937 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.840999 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.841010 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.841028 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.841038 4890 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T15:03:49Z","lastTransitionTime":"2025-11-25T15:03:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.882869 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wr7s" podStartSLOduration=67.882854843 podStartE2EDuration="1m7.882854843s" podCreationTimestamp="2025-11-25 15:02:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:49.549473992 +0000 UTC m=+87.991936612" watchObservedRunningTime="2025-11-25 15:03:49.882854843 +0000 UTC m=+88.325317453" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.883434 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-jrdzr"] Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.883887 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jrdzr" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.885720 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.886356 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.886528 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 25 15:03:49 crc kubenswrapper[4890]: I1125 15:03:49.886711 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 25 15:03:50 crc kubenswrapper[4890]: I1125 15:03:50.065679 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/bd997913-00a3-4947-8aa1-c39442a4503e-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-jrdzr\" (UID: \"bd997913-00a3-4947-8aa1-c39442a4503e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jrdzr" Nov 25 15:03:50 crc kubenswrapper[4890]: I1125 15:03:50.065718 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/bd997913-00a3-4947-8aa1-c39442a4503e-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-jrdzr\" (UID: \"bd997913-00a3-4947-8aa1-c39442a4503e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jrdzr" Nov 25 15:03:50 crc kubenswrapper[4890]: I1125 15:03:50.065754 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bd997913-00a3-4947-8aa1-c39442a4503e-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-jrdzr\" (UID: \"bd997913-00a3-4947-8aa1-c39442a4503e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jrdzr" Nov 25 15:03:50 crc kubenswrapper[4890]: I1125 15:03:50.065772 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bd997913-00a3-4947-8aa1-c39442a4503e-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-jrdzr\" (UID: \"bd997913-00a3-4947-8aa1-c39442a4503e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jrdzr" Nov 25 15:03:50 crc kubenswrapper[4890]: I1125 15:03:50.065804 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/bd997913-00a3-4947-8aa1-c39442a4503e-service-ca\") pod \"cluster-version-operator-5c965bbfc6-jrdzr\" (UID: \"bd997913-00a3-4947-8aa1-c39442a4503e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jrdzr" Nov 25 15:03:50 crc kubenswrapper[4890]: I1125 15:03:50.167320 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bd997913-00a3-4947-8aa1-c39442a4503e-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-jrdzr\" (UID: \"bd997913-00a3-4947-8aa1-c39442a4503e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jrdzr" Nov 25 15:03:50 crc kubenswrapper[4890]: I1125 15:03:50.167388 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/bd997913-00a3-4947-8aa1-c39442a4503e-service-ca\") pod \"cluster-version-operator-5c965bbfc6-jrdzr\" (UID: \"bd997913-00a3-4947-8aa1-c39442a4503e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jrdzr" Nov 25 15:03:50 crc kubenswrapper[4890]: I1125 15:03:50.167466 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/bd997913-00a3-4947-8aa1-c39442a4503e-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-jrdzr\" (UID: \"bd997913-00a3-4947-8aa1-c39442a4503e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jrdzr" Nov 25 15:03:50 crc kubenswrapper[4890]: I1125 15:03:50.167495 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/bd997913-00a3-4947-8aa1-c39442a4503e-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-jrdzr\" (UID: \"bd997913-00a3-4947-8aa1-c39442a4503e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jrdzr" Nov 25 15:03:50 crc kubenswrapper[4890]: I1125 15:03:50.167554 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bd997913-00a3-4947-8aa1-c39442a4503e-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-jrdzr\" (UID: \"bd997913-00a3-4947-8aa1-c39442a4503e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jrdzr" Nov 25 15:03:50 crc kubenswrapper[4890]: I1125 15:03:50.167953 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/bd997913-00a3-4947-8aa1-c39442a4503e-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-jrdzr\" (UID: \"bd997913-00a3-4947-8aa1-c39442a4503e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jrdzr" Nov 25 15:03:50 crc kubenswrapper[4890]: I1125 15:03:50.167964 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/bd997913-00a3-4947-8aa1-c39442a4503e-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-jrdzr\" (UID: \"bd997913-00a3-4947-8aa1-c39442a4503e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jrdzr" Nov 25 15:03:50 crc kubenswrapper[4890]: I1125 15:03:50.168374 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/bd997913-00a3-4947-8aa1-c39442a4503e-service-ca\") pod \"cluster-version-operator-5c965bbfc6-jrdzr\" (UID: \"bd997913-00a3-4947-8aa1-c39442a4503e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jrdzr" Nov 25 15:03:50 crc kubenswrapper[4890]: I1125 15:03:50.173078 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:03:50 crc kubenswrapper[4890]: E1125 15:03:50.173400 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 15:03:50 crc kubenswrapper[4890]: I1125 15:03:50.173868 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 15:03:50 crc kubenswrapper[4890]: E1125 15:03:50.173985 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 15:03:50 crc kubenswrapper[4890]: I1125 15:03:50.174233 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 15:03:50 crc kubenswrapper[4890]: E1125 15:03:50.174417 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 15:03:50 crc kubenswrapper[4890]: I1125 15:03:50.174836 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nlkp" Nov 25 15:03:50 crc kubenswrapper[4890]: E1125 15:03:50.174967 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2nlkp" podUID="16f52653-c067-4186-bc40-9c0f7c383b64" Nov 25 15:03:50 crc kubenswrapper[4890]: I1125 15:03:50.177510 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bd997913-00a3-4947-8aa1-c39442a4503e-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-jrdzr\" (UID: \"bd997913-00a3-4947-8aa1-c39442a4503e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jrdzr" Nov 25 15:03:50 crc kubenswrapper[4890]: I1125 15:03:50.184581 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bd997913-00a3-4947-8aa1-c39442a4503e-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-jrdzr\" (UID: \"bd997913-00a3-4947-8aa1-c39442a4503e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jrdzr" Nov 25 15:03:50 crc kubenswrapper[4890]: I1125 15:03:50.196282 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jrdzr" Nov 25 15:03:50 crc kubenswrapper[4890]: I1125 15:03:50.233662 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jrdzr" event={"ID":"bd997913-00a3-4947-8aa1-c39442a4503e","Type":"ContainerStarted","Data":"4f23d8a89ff009d3f071869f2473c88f6880f5daa1da598f2fa531d95d225751"} Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.006401 4890 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.006565 4890 kubelet_node_status.go:538] "Fast updating node status as it just became ready" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.040248 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xldhl"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.040706 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-6dmh8"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.040906 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xldhl" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.041216 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-6dmh8" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.043419 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v6ps9"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.044085 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v6ps9" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.044972 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-t29fd"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.045223 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.045580 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.045590 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-t29fd" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.045846 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.045851 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.046553 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-m95j6"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.047194 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-m95j6" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.048759 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nhq4p"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.049270 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nhq4p" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.052430 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-8bln9"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.053313 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-8bln9" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.059517 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.060018 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.060786 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.061114 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.061323 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.068901 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.069109 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.069203 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-9l4z5"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.069206 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.069233 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.069268 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.069315 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.069659 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-9l4z5" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.069706 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.069784 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.070221 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.070376 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.070800 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.070921 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.070943 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.071131 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.071254 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.071457 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.071636 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.071765 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.071887 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.072314 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.073316 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.073784 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.074060 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.074995 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mmhbh"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.075696 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.078758 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.079008 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.084374 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.090485 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-k4hj2"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.091020 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-k4hj2" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.092101 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-476wv"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.092485 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-476wv" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.092887 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-pshcp"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.093792 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.094032 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.094185 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.094204 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.094321 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.094373 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.094582 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.094718 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.094863 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.094997 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.096467 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-vkrx2"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.096789 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-mqvqx"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.097112 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-d7vvw"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.097354 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-pshcp" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.097642 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-d7vvw" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.097724 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.097951 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mqvqx" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.099435 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.099526 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.099821 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.100111 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.100126 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.099456 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.100328 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.100503 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.100556 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.100696 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.100803 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.101080 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.101204 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.103272 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-b8zql"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.104298 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-46rkq"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.104879 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-46rkq" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.111444 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8zql" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.113987 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.114143 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-m95j6"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.114479 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-n7q4s"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.116751 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n7q4s" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.118372 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h7rbb"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.118461 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.119003 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.119385 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.129890 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.130941 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.131149 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.132139 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.132831 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.132968 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.133123 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.133216 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.133280 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.133501 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.133799 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.134062 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.134213 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.134785 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.135261 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-7qxf9"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.135337 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.135465 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.135532 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.135574 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-7qxf9" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.135889 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.135936 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.135958 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h7rbb" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.136054 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.136126 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.136257 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.136284 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.136427 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.136540 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.136652 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.136894 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.136999 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.137173 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.137188 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.137315 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.137603 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.138101 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.138398 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pgn88"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.139218 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-glmvm"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.139770 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pgn88" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.139937 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hd47j"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.140330 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.140447 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-glmvm" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.140529 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hd47j" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.144071 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.146433 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-kxfxn"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.147533 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-ms5f2"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.147994 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-ms5f2" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.148102 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.148228 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-kxfxn" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.148608 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-ffl2m"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.149128 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-ffl2m" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.149958 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.151482 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.154449 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.154577 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.164782 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.165310 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.165478 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-v2tt6"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.167419 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-v2tt6" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.171808 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-c7vh9"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.172432 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-c7vh9" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.175334 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zs6w7"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.175990 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zs6w7" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.176646 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v6ps9"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.178423 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-t29fd"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.178987 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bc80d112-21a3-4721-9585-7b5c0a41d14d-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-v6ps9\" (UID: \"bc80d112-21a3-4721-9585-7b5c0a41d14d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v6ps9" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.179029 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttw72\" (UniqueName: \"kubernetes.io/projected/49586ef2-b2a8-4deb-8438-1e23cc0609df-kube-api-access-ttw72\") pod \"cluster-samples-operator-665b6dd947-m95j6\" (UID: \"49586ef2-b2a8-4deb-8438-1e23cc0609df\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-m95j6" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.179059 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-mmhbh\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.179077 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-874xv\" (UniqueName: \"kubernetes.io/projected/8b5c8177-1153-453a-9110-6023d01e3fb7-kube-api-access-874xv\") pod \"machine-api-operator-5694c8668f-6dmh8\" (UID: \"8b5c8177-1153-453a-9110-6023d01e3fb7\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6dmh8" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.179093 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfl4f\" (UniqueName: \"kubernetes.io/projected/6e313ec3-70c2-4220-b08a-ef79b8a25847-kube-api-access-qfl4f\") pod \"controller-manager-879f6c89f-xldhl\" (UID: \"6e313ec3-70c2-4220-b08a-ef79b8a25847\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xldhl" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.179111 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bc80d112-21a3-4721-9585-7b5c0a41d14d-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-v6ps9\" (UID: \"bc80d112-21a3-4721-9585-7b5c0a41d14d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v6ps9" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.179137 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c32eac2f-1dd7-4de9-97cf-32dc872da26b-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-nhq4p\" (UID: \"c32eac2f-1dd7-4de9-97cf-32dc872da26b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nhq4p" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.179154 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d6ec53ec-533b-49d7-966b-79b8c4f13e5f-trusted-ca\") pod \"console-operator-58897d9998-9l4z5\" (UID: \"d6ec53ec-533b-49d7-966b-79b8c4f13e5f\") " pod="openshift-console-operator/console-operator-58897d9998-9l4z5" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.179193 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b5c8177-1153-453a-9110-6023d01e3fb7-config\") pod \"machine-api-operator-5694c8668f-6dmh8\" (UID: \"8b5c8177-1153-453a-9110-6023d01e3fb7\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6dmh8" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.179209 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/8b5c8177-1153-453a-9110-6023d01e3fb7-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-6dmh8\" (UID: \"8b5c8177-1153-453a-9110-6023d01e3fb7\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6dmh8" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.179223 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6ec53ec-533b-49d7-966b-79b8c4f13e5f-config\") pod \"console-operator-58897d9998-9l4z5\" (UID: \"d6ec53ec-533b-49d7-966b-79b8c4f13e5f\") " pod="openshift-console-operator/console-operator-58897d9998-9l4z5" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.179243 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bl8cq\" (UniqueName: \"kubernetes.io/projected/bc80d112-21a3-4721-9585-7b5c0a41d14d-kube-api-access-bl8cq\") pod \"cluster-image-registry-operator-dc59b4c8b-v6ps9\" (UID: \"bc80d112-21a3-4721-9585-7b5c0a41d14d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v6ps9" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.179260 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6e313ec3-70c2-4220-b08a-ef79b8a25847-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-xldhl\" (UID: \"6e313ec3-70c2-4220-b08a-ef79b8a25847\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xldhl" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.179275 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-mmhbh\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.179293 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f748296c-d668-46bd-bcb7-4eb21822a244-console-oauth-config\") pod \"console-f9d7485db-t29fd\" (UID: \"f748296c-d668-46bd-bcb7-4eb21822a244\") " pod="openshift-console/console-f9d7485db-t29fd" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.179308 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-audit-policies\") pod \"oauth-openshift-558db77b4-mmhbh\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.179326 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-mmhbh\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.179343 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/8b5c8177-1153-453a-9110-6023d01e3fb7-images\") pod \"machine-api-operator-5694c8668f-6dmh8\" (UID: \"8b5c8177-1153-453a-9110-6023d01e3fb7\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6dmh8" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.179360 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lh5wj\" (UniqueName: \"kubernetes.io/projected/c32eac2f-1dd7-4de9-97cf-32dc872da26b-kube-api-access-lh5wj\") pod \"openshift-controller-manager-operator-756b6f6bc6-nhq4p\" (UID: \"c32eac2f-1dd7-4de9-97cf-32dc872da26b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nhq4p" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.179377 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f748296c-d668-46bd-bcb7-4eb21822a244-console-serving-cert\") pod \"console-f9d7485db-t29fd\" (UID: \"f748296c-d668-46bd-bcb7-4eb21822a244\") " pod="openshift-console/console-f9d7485db-t29fd" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.179391 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f748296c-d668-46bd-bcb7-4eb21822a244-trusted-ca-bundle\") pod \"console-f9d7485db-t29fd\" (UID: \"f748296c-d668-46bd-bcb7-4eb21822a244\") " pod="openshift-console/console-f9d7485db-t29fd" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.179407 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22rsz\" (UniqueName: \"kubernetes.io/projected/89bebf40-d718-4168-8377-4fa334afaa17-kube-api-access-22rsz\") pod \"downloads-7954f5f757-8bln9\" (UID: \"89bebf40-d718-4168-8377-4fa334afaa17\") " pod="openshift-console/downloads-7954f5f757-8bln9" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.179422 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-mmhbh\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.179438 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e313ec3-70c2-4220-b08a-ef79b8a25847-serving-cert\") pod \"controller-manager-879f6c89f-xldhl\" (UID: \"6e313ec3-70c2-4220-b08a-ef79b8a25847\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xldhl" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.179454 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e313ec3-70c2-4220-b08a-ef79b8a25847-config\") pod \"controller-manager-879f6c89f-xldhl\" (UID: \"6e313ec3-70c2-4220-b08a-ef79b8a25847\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xldhl" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.179470 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f748296c-d668-46bd-bcb7-4eb21822a244-console-config\") pod \"console-f9d7485db-t29fd\" (UID: \"f748296c-d668-46bd-bcb7-4eb21822a244\") " pod="openshift-console/console-f9d7485db-t29fd" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.179485 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-mmhbh\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.179500 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/49586ef2-b2a8-4deb-8438-1e23cc0609df-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-m95j6\" (UID: \"49586ef2-b2a8-4deb-8438-1e23cc0609df\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-m95j6" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.179517 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-mmhbh\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.179533 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-mmhbh\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.179548 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f748296c-d668-46bd-bcb7-4eb21822a244-oauth-serving-cert\") pod \"console-f9d7485db-t29fd\" (UID: \"f748296c-d668-46bd-bcb7-4eb21822a244\") " pod="openshift-console/console-f9d7485db-t29fd" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.179567 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bc80d112-21a3-4721-9585-7b5c0a41d14d-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-v6ps9\" (UID: \"bc80d112-21a3-4721-9585-7b5c0a41d14d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v6ps9" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.179587 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-mmhbh\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.179607 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-mmhbh\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.179615 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xldhl"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.179627 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-mmhbh\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.179649 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6e313ec3-70c2-4220-b08a-ef79b8a25847-client-ca\") pod \"controller-manager-879f6c89f-xldhl\" (UID: \"6e313ec3-70c2-4220-b08a-ef79b8a25847\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xldhl" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.179690 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-mmhbh\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.179725 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2p5mv\" (UniqueName: \"kubernetes.io/projected/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-kube-api-access-2p5mv\") pod \"oauth-openshift-558db77b4-mmhbh\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.179893 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c32eac2f-1dd7-4de9-97cf-32dc872da26b-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-nhq4p\" (UID: \"c32eac2f-1dd7-4de9-97cf-32dc872da26b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nhq4p" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.179962 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8kw4\" (UniqueName: \"kubernetes.io/projected/f748296c-d668-46bd-bcb7-4eb21822a244-kube-api-access-c8kw4\") pod \"console-f9d7485db-t29fd\" (UID: \"f748296c-d668-46bd-bcb7-4eb21822a244\") " pod="openshift-console/console-f9d7485db-t29fd" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.180000 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f748296c-d668-46bd-bcb7-4eb21822a244-service-ca\") pod \"console-f9d7485db-t29fd\" (UID: \"f748296c-d668-46bd-bcb7-4eb21822a244\") " pod="openshift-console/console-f9d7485db-t29fd" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.180026 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-audit-dir\") pod \"oauth-openshift-558db77b4-mmhbh\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.180049 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99bnn\" (UniqueName: \"kubernetes.io/projected/d6ec53ec-533b-49d7-966b-79b8c4f13e5f-kube-api-access-99bnn\") pod \"console-operator-58897d9998-9l4z5\" (UID: \"d6ec53ec-533b-49d7-966b-79b8c4f13e5f\") " pod="openshift-console-operator/console-operator-58897d9998-9l4z5" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.180111 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d6ec53ec-533b-49d7-966b-79b8c4f13e5f-serving-cert\") pod \"console-operator-58897d9998-9l4z5\" (UID: \"d6ec53ec-533b-49d7-966b-79b8c4f13e5f\") " pod="openshift-console-operator/console-operator-58897d9998-9l4z5" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.180305 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ktld7"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.180782 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ktld7" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.181361 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fc2g9"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.182447 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fc2g9" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.182620 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-z5tfd"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.183776 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-t5v4f"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.184685 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-z5tfd" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.184989 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t5v4f" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.185765 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nfp7k"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.186520 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nfp7k" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.187996 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.188441 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xmhq9"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.189095 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xmhq9" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.191030 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-jkwk5"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.198758 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-b8zql"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.198811 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-klnk6"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.199270 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-6dmh8"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.199295 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401380-96zpb"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.199672 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-99n87"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.200040 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-99n87" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.200210 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-klnk6" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.200524 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401380-96zpb" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.200552 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jkwk5" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.202811 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nhq4p"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.203227 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-k4hj2"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.206484 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.208146 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h7rbb"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.208198 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-46rkq"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.208216 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ktld7"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.209079 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-8bln9"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.211405 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-mqvqx"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.212825 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-vkrx2"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.217003 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-c7vh9"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.220770 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-ms5f2"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.223540 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-glmvm"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.226059 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-ffl2m"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.227343 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.227848 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-v2tt6"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.230587 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pgn88"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.231742 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-z5tfd"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.232952 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-n7q4s"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.234936 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mmhbh"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.239701 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-jkwk5"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.239769 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-476wv"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.246054 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.248823 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-pshcp"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.252895 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jrdzr" event={"ID":"bd997913-00a3-4947-8aa1-c39442a4503e","Type":"ContainerStarted","Data":"b8b03f5526640bb362a3789b8541e40896e16a88b14eb34301e4a373d5e423cd"} Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.257302 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-7qxf9"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.261417 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zs6w7"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.263015 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hd47j"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.264672 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.265179 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-rx9dj"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.267811 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-rx9dj" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.269433 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fc2g9"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.273670 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-9l4z5"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.274815 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-sm46z"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.276059 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-sm46z" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.280606 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401380-96zpb"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.281154 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f748296c-d668-46bd-bcb7-4eb21822a244-trusted-ca-bundle\") pod \"console-f9d7485db-t29fd\" (UID: \"f748296c-d668-46bd-bcb7-4eb21822a244\") " pod="openshift-console/console-f9d7485db-t29fd" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.281233 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22rsz\" (UniqueName: \"kubernetes.io/projected/89bebf40-d718-4168-8377-4fa334afaa17-kube-api-access-22rsz\") pod \"downloads-7954f5f757-8bln9\" (UID: \"89bebf40-d718-4168-8377-4fa334afaa17\") " pod="openshift-console/downloads-7954f5f757-8bln9" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.281258 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f748296c-d668-46bd-bcb7-4eb21822a244-console-serving-cert\") pod \"console-f9d7485db-t29fd\" (UID: \"f748296c-d668-46bd-bcb7-4eb21822a244\") " pod="openshift-console/console-f9d7485db-t29fd" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.281285 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e313ec3-70c2-4220-b08a-ef79b8a25847-serving-cert\") pod \"controller-manager-879f6c89f-xldhl\" (UID: \"6e313ec3-70c2-4220-b08a-ef79b8a25847\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xldhl" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.281306 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-mmhbh\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.281336 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/a6795dc3-313a-4599-a07c-588365bd8965-encryption-config\") pod \"apiserver-76f77b778f-glmvm\" (UID: \"a6795dc3-313a-4599-a07c-588365bd8965\") " pod="openshift-apiserver/apiserver-76f77b778f-glmvm" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.281361 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f748296c-d668-46bd-bcb7-4eb21822a244-console-config\") pod \"console-f9d7485db-t29fd\" (UID: \"f748296c-d668-46bd-bcb7-4eb21822a244\") " pod="openshift-console/console-f9d7485db-t29fd" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.281391 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-mmhbh\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.281411 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/49586ef2-b2a8-4deb-8438-1e23cc0609df-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-m95j6\" (UID: \"49586ef2-b2a8-4deb-8438-1e23cc0609df\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-m95j6" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.281458 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a6795dc3-313a-4599-a07c-588365bd8965-audit-dir\") pod \"apiserver-76f77b778f-glmvm\" (UID: \"a6795dc3-313a-4599-a07c-588365bd8965\") " pod="openshift-apiserver/apiserver-76f77b778f-glmvm" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.281481 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e313ec3-70c2-4220-b08a-ef79b8a25847-config\") pod \"controller-manager-879f6c89f-xldhl\" (UID: \"6e313ec3-70c2-4220-b08a-ef79b8a25847\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xldhl" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.281503 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-mmhbh\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.281527 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-mmhbh\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.281551 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6795dc3-313a-4599-a07c-588365bd8965-config\") pod \"apiserver-76f77b778f-glmvm\" (UID: \"a6795dc3-313a-4599-a07c-588365bd8965\") " pod="openshift-apiserver/apiserver-76f77b778f-glmvm" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.281575 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f748296c-d668-46bd-bcb7-4eb21822a244-oauth-serving-cert\") pod \"console-f9d7485db-t29fd\" (UID: \"f748296c-d668-46bd-bcb7-4eb21822a244\") " pod="openshift-console/console-f9d7485db-t29fd" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.281595 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bc80d112-21a3-4721-9585-7b5c0a41d14d-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-v6ps9\" (UID: \"bc80d112-21a3-4721-9585-7b5c0a41d14d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v6ps9" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.281615 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-mmhbh\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.281639 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-mmhbh\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.281663 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-mmhbh\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.281690 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6e313ec3-70c2-4220-b08a-ef79b8a25847-client-ca\") pod \"controller-manager-879f6c89f-xldhl\" (UID: \"6e313ec3-70c2-4220-b08a-ef79b8a25847\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xldhl" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.281713 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-mmhbh\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.281770 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2p5mv\" (UniqueName: \"kubernetes.io/projected/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-kube-api-access-2p5mv\") pod \"oauth-openshift-558db77b4-mmhbh\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.281801 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c32eac2f-1dd7-4de9-97cf-32dc872da26b-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-nhq4p\" (UID: \"c32eac2f-1dd7-4de9-97cf-32dc872da26b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nhq4p" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.281822 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8kw4\" (UniqueName: \"kubernetes.io/projected/f748296c-d668-46bd-bcb7-4eb21822a244-kube-api-access-c8kw4\") pod \"console-f9d7485db-t29fd\" (UID: \"f748296c-d668-46bd-bcb7-4eb21822a244\") " pod="openshift-console/console-f9d7485db-t29fd" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.281843 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99bnn\" (UniqueName: \"kubernetes.io/projected/d6ec53ec-533b-49d7-966b-79b8c4f13e5f-kube-api-access-99bnn\") pod \"console-operator-58897d9998-9l4z5\" (UID: \"d6ec53ec-533b-49d7-966b-79b8c4f13e5f\") " pod="openshift-console-operator/console-operator-58897d9998-9l4z5" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.281868 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a6795dc3-313a-4599-a07c-588365bd8965-etcd-client\") pod \"apiserver-76f77b778f-glmvm\" (UID: \"a6795dc3-313a-4599-a07c-588365bd8965\") " pod="openshift-apiserver/apiserver-76f77b778f-glmvm" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.281886 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/a6795dc3-313a-4599-a07c-588365bd8965-etcd-serving-ca\") pod \"apiserver-76f77b778f-glmvm\" (UID: \"a6795dc3-313a-4599-a07c-588365bd8965\") " pod="openshift-apiserver/apiserver-76f77b778f-glmvm" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.281910 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a6795dc3-313a-4599-a07c-588365bd8965-serving-cert\") pod \"apiserver-76f77b778f-glmvm\" (UID: \"a6795dc3-313a-4599-a07c-588365bd8965\") " pod="openshift-apiserver/apiserver-76f77b778f-glmvm" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.281931 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f748296c-d668-46bd-bcb7-4eb21822a244-service-ca\") pod \"console-f9d7485db-t29fd\" (UID: \"f748296c-d668-46bd-bcb7-4eb21822a244\") " pod="openshift-console/console-f9d7485db-t29fd" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.281948 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-audit-dir\") pod \"oauth-openshift-558db77b4-mmhbh\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.281970 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d6ec53ec-533b-49d7-966b-79b8c4f13e5f-serving-cert\") pod \"console-operator-58897d9998-9l4z5\" (UID: \"d6ec53ec-533b-49d7-966b-79b8c4f13e5f\") " pod="openshift-console-operator/console-operator-58897d9998-9l4z5" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.281998 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bc80d112-21a3-4721-9585-7b5c0a41d14d-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-v6ps9\" (UID: \"bc80d112-21a3-4721-9585-7b5c0a41d14d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v6ps9" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.282016 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttw72\" (UniqueName: \"kubernetes.io/projected/49586ef2-b2a8-4deb-8438-1e23cc0609df-kube-api-access-ttw72\") pod \"cluster-samples-operator-665b6dd947-m95j6\" (UID: \"49586ef2-b2a8-4deb-8438-1e23cc0609df\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-m95j6" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.282046 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-mmhbh\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.282068 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a6795dc3-313a-4599-a07c-588365bd8965-trusted-ca-bundle\") pod \"apiserver-76f77b778f-glmvm\" (UID: \"a6795dc3-313a-4599-a07c-588365bd8965\") " pod="openshift-apiserver/apiserver-76f77b778f-glmvm" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.282093 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-874xv\" (UniqueName: \"kubernetes.io/projected/8b5c8177-1153-453a-9110-6023d01e3fb7-kube-api-access-874xv\") pod \"machine-api-operator-5694c8668f-6dmh8\" (UID: \"8b5c8177-1153-453a-9110-6023d01e3fb7\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6dmh8" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.282117 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfl4f\" (UniqueName: \"kubernetes.io/projected/6e313ec3-70c2-4220-b08a-ef79b8a25847-kube-api-access-qfl4f\") pod \"controller-manager-879f6c89f-xldhl\" (UID: \"6e313ec3-70c2-4220-b08a-ef79b8a25847\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xldhl" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.282139 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bc80d112-21a3-4721-9585-7b5c0a41d14d-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-v6ps9\" (UID: \"bc80d112-21a3-4721-9585-7b5c0a41d14d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v6ps9" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.282179 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/a6795dc3-313a-4599-a07c-588365bd8965-image-import-ca\") pod \"apiserver-76f77b778f-glmvm\" (UID: \"a6795dc3-313a-4599-a07c-588365bd8965\") " pod="openshift-apiserver/apiserver-76f77b778f-glmvm" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.282205 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c32eac2f-1dd7-4de9-97cf-32dc872da26b-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-nhq4p\" (UID: \"c32eac2f-1dd7-4de9-97cf-32dc872da26b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nhq4p" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.282225 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d6ec53ec-533b-49d7-966b-79b8c4f13e5f-trusted-ca\") pod \"console-operator-58897d9998-9l4z5\" (UID: \"d6ec53ec-533b-49d7-966b-79b8c4f13e5f\") " pod="openshift-console-operator/console-operator-58897d9998-9l4z5" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.282258 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/a6795dc3-313a-4599-a07c-588365bd8965-audit\") pod \"apiserver-76f77b778f-glmvm\" (UID: \"a6795dc3-313a-4599-a07c-588365bd8965\") " pod="openshift-apiserver/apiserver-76f77b778f-glmvm" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.282277 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4w9hj\" (UniqueName: \"kubernetes.io/projected/a6795dc3-313a-4599-a07c-588365bd8965-kube-api-access-4w9hj\") pod \"apiserver-76f77b778f-glmvm\" (UID: \"a6795dc3-313a-4599-a07c-588365bd8965\") " pod="openshift-apiserver/apiserver-76f77b778f-glmvm" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.282310 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b5c8177-1153-453a-9110-6023d01e3fb7-config\") pod \"machine-api-operator-5694c8668f-6dmh8\" (UID: \"8b5c8177-1153-453a-9110-6023d01e3fb7\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6dmh8" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.282332 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/8b5c8177-1153-453a-9110-6023d01e3fb7-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-6dmh8\" (UID: \"8b5c8177-1153-453a-9110-6023d01e3fb7\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6dmh8" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.282352 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6ec53ec-533b-49d7-966b-79b8c4f13e5f-config\") pod \"console-operator-58897d9998-9l4z5\" (UID: \"d6ec53ec-533b-49d7-966b-79b8c4f13e5f\") " pod="openshift-console-operator/console-operator-58897d9998-9l4z5" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.282383 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bl8cq\" (UniqueName: \"kubernetes.io/projected/bc80d112-21a3-4721-9585-7b5c0a41d14d-kube-api-access-bl8cq\") pod \"cluster-image-registry-operator-dc59b4c8b-v6ps9\" (UID: \"bc80d112-21a3-4721-9585-7b5c0a41d14d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v6ps9" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.282409 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6e313ec3-70c2-4220-b08a-ef79b8a25847-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-xldhl\" (UID: \"6e313ec3-70c2-4220-b08a-ef79b8a25847\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xldhl" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.282431 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-mmhbh\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.282452 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f748296c-d668-46bd-bcb7-4eb21822a244-console-oauth-config\") pod \"console-f9d7485db-t29fd\" (UID: \"f748296c-d668-46bd-bcb7-4eb21822a244\") " pod="openshift-console/console-f9d7485db-t29fd" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.282474 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-audit-policies\") pod \"oauth-openshift-558db77b4-mmhbh\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.282496 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-mmhbh\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.283616 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-mmhbh\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.284453 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f748296c-d668-46bd-bcb7-4eb21822a244-trusted-ca-bundle\") pod \"console-f9d7485db-t29fd\" (UID: \"f748296c-d668-46bd-bcb7-4eb21822a244\") " pod="openshift-console/console-f9d7485db-t29fd" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.285817 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f748296c-d668-46bd-bcb7-4eb21822a244-oauth-serving-cert\") pod \"console-f9d7485db-t29fd\" (UID: \"f748296c-d668-46bd-bcb7-4eb21822a244\") " pod="openshift-console/console-f9d7485db-t29fd" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.287962 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-mmhbh\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.288612 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/49586ef2-b2a8-4deb-8438-1e23cc0609df-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-m95j6\" (UID: \"49586ef2-b2a8-4deb-8438-1e23cc0609df\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-m95j6" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.288899 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/8b5c8177-1153-453a-9110-6023d01e3fb7-images\") pod \"machine-api-operator-5694c8668f-6dmh8\" (UID: \"8b5c8177-1153-453a-9110-6023d01e3fb7\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6dmh8" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.289169 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lh5wj\" (UniqueName: \"kubernetes.io/projected/c32eac2f-1dd7-4de9-97cf-32dc872da26b-kube-api-access-lh5wj\") pod \"openshift-controller-manager-operator-756b6f6bc6-nhq4p\" (UID: \"c32eac2f-1dd7-4de9-97cf-32dc872da26b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nhq4p" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.289216 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a6795dc3-313a-4599-a07c-588365bd8965-node-pullsecrets\") pod \"apiserver-76f77b778f-glmvm\" (UID: \"a6795dc3-313a-4599-a07c-588365bd8965\") " pod="openshift-apiserver/apiserver-76f77b778f-glmvm" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.289494 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/8b5c8177-1153-453a-9110-6023d01e3fb7-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-6dmh8\" (UID: \"8b5c8177-1153-453a-9110-6023d01e3fb7\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6dmh8" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.289557 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-mmhbh\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.290326 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6e313ec3-70c2-4220-b08a-ef79b8a25847-client-ca\") pod \"controller-manager-879f6c89f-xldhl\" (UID: \"6e313ec3-70c2-4220-b08a-ef79b8a25847\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xldhl" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.291342 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.292113 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6ec53ec-533b-49d7-966b-79b8c4f13e5f-config\") pod \"console-operator-58897d9998-9l4z5\" (UID: \"d6ec53ec-533b-49d7-966b-79b8c4f13e5f\") " pod="openshift-console-operator/console-operator-58897d9998-9l4z5" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.292181 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c32eac2f-1dd7-4de9-97cf-32dc872da26b-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-nhq4p\" (UID: \"c32eac2f-1dd7-4de9-97cf-32dc872da26b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nhq4p" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.292693 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f748296c-d668-46bd-bcb7-4eb21822a244-console-config\") pod \"console-f9d7485db-t29fd\" (UID: \"f748296c-d668-46bd-bcb7-4eb21822a244\") " pod="openshift-console/console-f9d7485db-t29fd" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.292775 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e313ec3-70c2-4220-b08a-ef79b8a25847-serving-cert\") pod \"controller-manager-879f6c89f-xldhl\" (UID: \"6e313ec3-70c2-4220-b08a-ef79b8a25847\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xldhl" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.292981 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b5c8177-1153-453a-9110-6023d01e3fb7-config\") pod \"machine-api-operator-5694c8668f-6dmh8\" (UID: \"8b5c8177-1153-453a-9110-6023d01e3fb7\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6dmh8" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.293594 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-audit-policies\") pod \"oauth-openshift-558db77b4-mmhbh\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.293829 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f748296c-d668-46bd-bcb7-4eb21822a244-service-ca\") pod \"console-f9d7485db-t29fd\" (UID: \"f748296c-d668-46bd-bcb7-4eb21822a244\") " pod="openshift-console/console-f9d7485db-t29fd" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.293904 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-audit-dir\") pod \"oauth-openshift-558db77b4-mmhbh\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.294763 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d6ec53ec-533b-49d7-966b-79b8c4f13e5f-trusted-ca\") pod \"console-operator-58897d9998-9l4z5\" (UID: \"d6ec53ec-533b-49d7-966b-79b8c4f13e5f\") " pod="openshift-console-operator/console-operator-58897d9998-9l4z5" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.295653 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e313ec3-70c2-4220-b08a-ef79b8a25847-config\") pod \"controller-manager-879f6c89f-xldhl\" (UID: \"6e313ec3-70c2-4220-b08a-ef79b8a25847\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xldhl" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.296332 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6e313ec3-70c2-4220-b08a-ef79b8a25847-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-xldhl\" (UID: \"6e313ec3-70c2-4220-b08a-ef79b8a25847\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xldhl" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.296917 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/8b5c8177-1153-453a-9110-6023d01e3fb7-images\") pod \"machine-api-operator-5694c8668f-6dmh8\" (UID: \"8b5c8177-1153-453a-9110-6023d01e3fb7\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6dmh8" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.297648 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c32eac2f-1dd7-4de9-97cf-32dc872da26b-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-nhq4p\" (UID: \"c32eac2f-1dd7-4de9-97cf-32dc872da26b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nhq4p" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.298767 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bc80d112-21a3-4721-9585-7b5c0a41d14d-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-v6ps9\" (UID: \"bc80d112-21a3-4721-9585-7b5c0a41d14d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v6ps9" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.298984 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f748296c-d668-46bd-bcb7-4eb21822a244-console-serving-cert\") pod \"console-f9d7485db-t29fd\" (UID: \"f748296c-d668-46bd-bcb7-4eb21822a244\") " pod="openshift-console/console-f9d7485db-t29fd" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.301502 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f748296c-d668-46bd-bcb7-4eb21822a244-console-oauth-config\") pod \"console-f9d7485db-t29fd\" (UID: \"f748296c-d668-46bd-bcb7-4eb21822a244\") " pod="openshift-console/console-f9d7485db-t29fd" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.301560 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-99n87"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.302225 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-mmhbh\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.302713 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-mmhbh\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.304060 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-klnk6"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.304393 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-mmhbh\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.304690 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.304783 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-mmhbh\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.304888 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-mmhbh\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.305702 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bc80d112-21a3-4721-9585-7b5c0a41d14d-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-v6ps9\" (UID: \"bc80d112-21a3-4721-9585-7b5c0a41d14d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v6ps9" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.306588 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-mmhbh\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.306891 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-mmhbh\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.307426 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-mmhbh\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.309996 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xmhq9"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.310042 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nfp7k"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.310813 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-rx9dj"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.310961 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d6ec53ec-533b-49d7-966b-79b8c4f13e5f-serving-cert\") pod \"console-operator-58897d9998-9l4z5\" (UID: \"d6ec53ec-533b-49d7-966b-79b8c4f13e5f\") " pod="openshift-console-operator/console-operator-58897d9998-9l4z5" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.311992 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-t5v4f"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.313076 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-cmpbm"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.314790 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-cmpbm"] Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.314947 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-cmpbm" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.323366 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.344138 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.363612 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.382970 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.390342 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a6795dc3-313a-4599-a07c-588365bd8965-etcd-client\") pod \"apiserver-76f77b778f-glmvm\" (UID: \"a6795dc3-313a-4599-a07c-588365bd8965\") " pod="openshift-apiserver/apiserver-76f77b778f-glmvm" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.390385 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/a6795dc3-313a-4599-a07c-588365bd8965-etcd-serving-ca\") pod \"apiserver-76f77b778f-glmvm\" (UID: \"a6795dc3-313a-4599-a07c-588365bd8965\") " pod="openshift-apiserver/apiserver-76f77b778f-glmvm" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.390412 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a6795dc3-313a-4599-a07c-588365bd8965-serving-cert\") pod \"apiserver-76f77b778f-glmvm\" (UID: \"a6795dc3-313a-4599-a07c-588365bd8965\") " pod="openshift-apiserver/apiserver-76f77b778f-glmvm" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.390480 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a6795dc3-313a-4599-a07c-588365bd8965-trusted-ca-bundle\") pod \"apiserver-76f77b778f-glmvm\" (UID: \"a6795dc3-313a-4599-a07c-588365bd8965\") " pod="openshift-apiserver/apiserver-76f77b778f-glmvm" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.390519 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/a6795dc3-313a-4599-a07c-588365bd8965-image-import-ca\") pod \"apiserver-76f77b778f-glmvm\" (UID: \"a6795dc3-313a-4599-a07c-588365bd8965\") " pod="openshift-apiserver/apiserver-76f77b778f-glmvm" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.390542 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4w9hj\" (UniqueName: \"kubernetes.io/projected/a6795dc3-313a-4599-a07c-588365bd8965-kube-api-access-4w9hj\") pod \"apiserver-76f77b778f-glmvm\" (UID: \"a6795dc3-313a-4599-a07c-588365bd8965\") " pod="openshift-apiserver/apiserver-76f77b778f-glmvm" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.390576 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/a6795dc3-313a-4599-a07c-588365bd8965-audit\") pod \"apiserver-76f77b778f-glmvm\" (UID: \"a6795dc3-313a-4599-a07c-588365bd8965\") " pod="openshift-apiserver/apiserver-76f77b778f-glmvm" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.390632 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a6795dc3-313a-4599-a07c-588365bd8965-node-pullsecrets\") pod \"apiserver-76f77b778f-glmvm\" (UID: \"a6795dc3-313a-4599-a07c-588365bd8965\") " pod="openshift-apiserver/apiserver-76f77b778f-glmvm" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.390674 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/a6795dc3-313a-4599-a07c-588365bd8965-encryption-config\") pod \"apiserver-76f77b778f-glmvm\" (UID: \"a6795dc3-313a-4599-a07c-588365bd8965\") " pod="openshift-apiserver/apiserver-76f77b778f-glmvm" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.390698 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a6795dc3-313a-4599-a07c-588365bd8965-audit-dir\") pod \"apiserver-76f77b778f-glmvm\" (UID: \"a6795dc3-313a-4599-a07c-588365bd8965\") " pod="openshift-apiserver/apiserver-76f77b778f-glmvm" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.390735 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6795dc3-313a-4599-a07c-588365bd8965-config\") pod \"apiserver-76f77b778f-glmvm\" (UID: \"a6795dc3-313a-4599-a07c-588365bd8965\") " pod="openshift-apiserver/apiserver-76f77b778f-glmvm" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.390819 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a6795dc3-313a-4599-a07c-588365bd8965-node-pullsecrets\") pod \"apiserver-76f77b778f-glmvm\" (UID: \"a6795dc3-313a-4599-a07c-588365bd8965\") " pod="openshift-apiserver/apiserver-76f77b778f-glmvm" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.390883 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a6795dc3-313a-4599-a07c-588365bd8965-audit-dir\") pod \"apiserver-76f77b778f-glmvm\" (UID: \"a6795dc3-313a-4599-a07c-588365bd8965\") " pod="openshift-apiserver/apiserver-76f77b778f-glmvm" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.403539 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.423769 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.443094 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.463550 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.483290 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.503117 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.523259 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.531230 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/a6795dc3-313a-4599-a07c-588365bd8965-etcd-serving-ca\") pod \"apiserver-76f77b778f-glmvm\" (UID: \"a6795dc3-313a-4599-a07c-588365bd8965\") " pod="openshift-apiserver/apiserver-76f77b778f-glmvm" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.543176 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.552126 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/a6795dc3-313a-4599-a07c-588365bd8965-audit\") pod \"apiserver-76f77b778f-glmvm\" (UID: \"a6795dc3-313a-4599-a07c-588365bd8965\") " pod="openshift-apiserver/apiserver-76f77b778f-glmvm" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.563626 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.583309 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.591767 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6795dc3-313a-4599-a07c-588365bd8965-config\") pod \"apiserver-76f77b778f-glmvm\" (UID: \"a6795dc3-313a-4599-a07c-588365bd8965\") " pod="openshift-apiserver/apiserver-76f77b778f-glmvm" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.603612 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.611589 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/a6795dc3-313a-4599-a07c-588365bd8965-image-import-ca\") pod \"apiserver-76f77b778f-glmvm\" (UID: \"a6795dc3-313a-4599-a07c-588365bd8965\") " pod="openshift-apiserver/apiserver-76f77b778f-glmvm" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.630873 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.631765 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a6795dc3-313a-4599-a07c-588365bd8965-trusted-ca-bundle\") pod \"apiserver-76f77b778f-glmvm\" (UID: \"a6795dc3-313a-4599-a07c-588365bd8965\") " pod="openshift-apiserver/apiserver-76f77b778f-glmvm" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.643032 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.664139 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.674633 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a6795dc3-313a-4599-a07c-588365bd8965-etcd-client\") pod \"apiserver-76f77b778f-glmvm\" (UID: \"a6795dc3-313a-4599-a07c-588365bd8965\") " pod="openshift-apiserver/apiserver-76f77b778f-glmvm" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.683379 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.695768 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a6795dc3-313a-4599-a07c-588365bd8965-serving-cert\") pod \"apiserver-76f77b778f-glmvm\" (UID: \"a6795dc3-313a-4599-a07c-588365bd8965\") " pod="openshift-apiserver/apiserver-76f77b778f-glmvm" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.704374 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.714319 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/a6795dc3-313a-4599-a07c-588365bd8965-encryption-config\") pod \"apiserver-76f77b778f-glmvm\" (UID: \"a6795dc3-313a-4599-a07c-588365bd8965\") " pod="openshift-apiserver/apiserver-76f77b778f-glmvm" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.723970 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.743814 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.763906 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.784686 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.803559 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.824928 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.863809 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.883550 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.903293 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.923925 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.943264 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.963467 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 25 15:03:51 crc kubenswrapper[4890]: I1125 15:03:51.984246 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.004129 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.023218 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.043619 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.063816 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.083300 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.102559 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.123504 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.143194 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.167860 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.172597 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.172650 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nlkp" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.172836 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.172597 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.182082 4890 request.go:700] Waited for 1.009351897s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&limit=500&resourceVersion=0 Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.183972 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.202614 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.223957 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.244761 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.263492 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.282688 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.303546 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.323246 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.342516 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.363192 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.403262 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.423365 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.443634 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.463341 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.484089 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.503992 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.523433 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.551610 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.563033 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.584033 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.604549 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.622910 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.642787 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.667523 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.683563 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.703787 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.723998 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.743515 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.764031 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.784113 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.803845 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.824029 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.843526 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.864400 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.883510 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.903302 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.922935 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.943096 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.963584 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 25 15:03:52 crc kubenswrapper[4890]: I1125 15:03:52.983784 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.003627 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.023576 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.060319 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bc80d112-21a3-4721-9585-7b5c0a41d14d-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-v6ps9\" (UID: \"bc80d112-21a3-4721-9585-7b5c0a41d14d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v6ps9" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.087397 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22rsz\" (UniqueName: \"kubernetes.io/projected/89bebf40-d718-4168-8377-4fa334afaa17-kube-api-access-22rsz\") pod \"downloads-7954f5f757-8bln9\" (UID: \"89bebf40-d718-4168-8377-4fa334afaa17\") " pod="openshift-console/downloads-7954f5f757-8bln9" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.099774 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2p5mv\" (UniqueName: \"kubernetes.io/projected/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-kube-api-access-2p5mv\") pod \"oauth-openshift-558db77b4-mmhbh\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.124194 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bl8cq\" (UniqueName: \"kubernetes.io/projected/bc80d112-21a3-4721-9585-7b5c0a41d14d-kube-api-access-bl8cq\") pod \"cluster-image-registry-operator-dc59b4c8b-v6ps9\" (UID: \"bc80d112-21a3-4721-9585-7b5c0a41d14d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v6ps9" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.143375 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lh5wj\" (UniqueName: \"kubernetes.io/projected/c32eac2f-1dd7-4de9-97cf-32dc872da26b-kube-api-access-lh5wj\") pod \"openshift-controller-manager-operator-756b6f6bc6-nhq4p\" (UID: \"c32eac2f-1dd7-4de9-97cf-32dc872da26b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nhq4p" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.158732 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8kw4\" (UniqueName: \"kubernetes.io/projected/f748296c-d668-46bd-bcb7-4eb21822a244-kube-api-access-c8kw4\") pod \"console-f9d7485db-t29fd\" (UID: \"f748296c-d668-46bd-bcb7-4eb21822a244\") " pod="openshift-console/console-f9d7485db-t29fd" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.181132 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttw72\" (UniqueName: \"kubernetes.io/projected/49586ef2-b2a8-4deb-8438-1e23cc0609df-kube-api-access-ttw72\") pod \"cluster-samples-operator-665b6dd947-m95j6\" (UID: \"49586ef2-b2a8-4deb-8438-1e23cc0609df\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-m95j6" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.193407 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v6ps9" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.200349 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99bnn\" (UniqueName: \"kubernetes.io/projected/d6ec53ec-533b-49d7-966b-79b8c4f13e5f-kube-api-access-99bnn\") pod \"console-operator-58897d9998-9l4z5\" (UID: \"d6ec53ec-533b-49d7-966b-79b8c4f13e5f\") " pod="openshift-console-operator/console-operator-58897d9998-9l4z5" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.201780 4890 request.go:700] Waited for 1.900111917s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/serviceaccounts/openshift-controller-manager-sa/token Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.214234 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-t29fd" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.221974 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfl4f\" (UniqueName: \"kubernetes.io/projected/6e313ec3-70c2-4220-b08a-ef79b8a25847-kube-api-access-qfl4f\") pod \"controller-manager-879f6c89f-xldhl\" (UID: \"6e313ec3-70c2-4220-b08a-ef79b8a25847\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xldhl" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.240142 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-874xv\" (UniqueName: \"kubernetes.io/projected/8b5c8177-1153-453a-9110-6023d01e3fb7-kube-api-access-874xv\") pod \"machine-api-operator-5694c8668f-6dmh8\" (UID: \"8b5c8177-1153-453a-9110-6023d01e3fb7\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6dmh8" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.242112 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-m95j6" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.245057 4890 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.251453 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nhq4p" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.264210 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.267512 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-8bln9" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.273968 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-9l4z5" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.282840 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.287633 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.326186 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4w9hj\" (UniqueName: \"kubernetes.io/projected/a6795dc3-313a-4599-a07c-588365bd8965-kube-api-access-4w9hj\") pod \"apiserver-76f77b778f-glmvm\" (UID: \"a6795dc3-313a-4599-a07c-588365bd8965\") " pod="openshift-apiserver/apiserver-76f77b778f-glmvm" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.344908 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.367968 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.383507 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.405016 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.423628 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.444895 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.470423 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xldhl" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.471271 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v6ps9"] Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.484295 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-6dmh8" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.488221 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-glmvm" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.519302 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34cf8610-cc4e-46de-a00c-ff41d7030e1f-config\") pod \"kube-apiserver-operator-766d6c64bb-h7rbb\" (UID: \"34cf8610-cc4e-46de-a00c-ff41d7030e1f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h7rbb" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.519354 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/91dd7b10-99ef-486d-a624-21cb245226c7-encryption-config\") pod \"apiserver-7bbb656c7d-b8zql\" (UID: \"91dd7b10-99ef-486d-a624-21cb245226c7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8zql" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.519382 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e23712e1-8705-4d4b-bd1a-1bea23f765df-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-k4hj2\" (UID: \"e23712e1-8705-4d4b-bd1a-1bea23f765df\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-k4hj2" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.519402 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgh2l\" (UniqueName: \"kubernetes.io/projected/62ff675c-ff12-4c14-9c6b-ce3fb25ea214-kube-api-access-jgh2l\") pod \"openshift-config-operator-7777fb866f-pshcp\" (UID: \"62ff675c-ff12-4c14-9c6b-ce3fb25ea214\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-pshcp" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.519442 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/82ece3ff-72cb-48b3-92ad-251d7bf980c0-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-pgn88\" (UID: \"82ece3ff-72cb-48b3-92ad-251d7bf980c0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pgn88" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.519462 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxr96\" (UniqueName: \"kubernetes.io/projected/91dd7b10-99ef-486d-a624-21cb245226c7-kube-api-access-xxr96\") pod \"apiserver-7bbb656c7d-b8zql\" (UID: \"91dd7b10-99ef-486d-a624-21cb245226c7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8zql" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.519482 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/62ff675c-ff12-4c14-9c6b-ce3fb25ea214-serving-cert\") pod \"openshift-config-operator-7777fb866f-pshcp\" (UID: \"62ff675c-ff12-4c14-9c6b-ce3fb25ea214\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-pshcp" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.519525 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a0b0911d-7b7c-4746-bcfc-d15069434bb1-proxy-tls\") pod \"machine-config-controller-84d6567774-v2tt6\" (UID: \"a0b0911d-7b7c-4746-bcfc-d15069434bb1\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-v2tt6" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.519542 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5863644d-bf1b-4dbf-9c89-99ec5bff14e3-config\") pod \"openshift-apiserver-operator-796bbdcf4f-476wv\" (UID: \"5863644d-bf1b-4dbf-9c89-99ec5bff14e3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-476wv" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.519576 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/91dd7b10-99ef-486d-a624-21cb245226c7-audit-dir\") pod \"apiserver-7bbb656c7d-b8zql\" (UID: \"91dd7b10-99ef-486d-a624-21cb245226c7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8zql" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.519598 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/acb1ddd4-f551-4d06-9021-28d7579bad7f-srv-cert\") pod \"catalog-operator-68c6474976-c7vh9\" (UID: \"acb1ddd4-f551-4d06-9021-28d7579bad7f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-c7vh9" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.519635 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7e0b1cd-5c13-436d-8c35-c0a659f767a0-config\") pod \"etcd-operator-b45778765-7qxf9\" (UID: \"b7e0b1cd-5c13-436d-8c35-c0a659f767a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7qxf9" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.519662 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5cf1f97f-ff78-4047-b839-6e8ffb6b92b9-trusted-ca\") pod \"ingress-operator-5b745b69d9-n7q4s\" (UID: \"5cf1f97f-ff78-4047-b839-6e8ffb6b92b9\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n7q4s" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.519693 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kk9j6\" (UniqueName: \"kubernetes.io/projected/5cf1f97f-ff78-4047-b839-6e8ffb6b92b9-kube-api-access-kk9j6\") pod \"ingress-operator-5b745b69d9-n7q4s\" (UID: \"5cf1f97f-ff78-4047-b839-6e8ffb6b92b9\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n7q4s" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.519709 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/91dd7b10-99ef-486d-a624-21cb245226c7-etcd-client\") pod \"apiserver-7bbb656c7d-b8zql\" (UID: \"91dd7b10-99ef-486d-a624-21cb245226c7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8zql" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.519725 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6vvk\" (UniqueName: \"kubernetes.io/projected/ac102463-bf2a-466d-b24a-e3249661d1ee-kube-api-access-q6vvk\") pod \"machine-approver-56656f9798-d7vvw\" (UID: \"ac102463-bf2a-466d-b24a-e3249661d1ee\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-d7vvw" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.519752 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e09b8f75-2286-4a12-9571-5342b681690f-client-ca\") pod \"route-controller-manager-6576b87f9c-mqvqx\" (UID: \"e09b8f75-2286-4a12-9571-5342b681690f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mqvqx" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.519782 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82ece3ff-72cb-48b3-92ad-251d7bf980c0-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-pgn88\" (UID: \"82ece3ff-72cb-48b3-92ad-251d7bf980c0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pgn88" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.519829 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvshl\" (UniqueName: \"kubernetes.io/projected/1a324db0-9f3e-417e-931f-5b74a44775ed-kube-api-access-zvshl\") pod \"migrator-59844c95c7-t5v4f\" (UID: \"1a324db0-9f3e-417e-931f-5b74a44775ed\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t5v4f" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.519893 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0826463c-2c8a-4b52-8727-4b07de3d86c3-config\") pod \"kube-controller-manager-operator-78b949d7b-ktld7\" (UID: \"0826463c-2c8a-4b52-8727-4b07de3d86c3\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ktld7" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.519924 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/91dd7b10-99ef-486d-a624-21cb245226c7-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-b8zql\" (UID: \"91dd7b10-99ef-486d-a624-21cb245226c7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8zql" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.519942 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5863644d-bf1b-4dbf-9c89-99ec5bff14e3-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-476wv\" (UID: \"5863644d-bf1b-4dbf-9c89-99ec5bff14e3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-476wv" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.519963 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e09b8f75-2286-4a12-9571-5342b681690f-serving-cert\") pod \"route-controller-manager-6576b87f9c-mqvqx\" (UID: \"e09b8f75-2286-4a12-9571-5342b681690f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mqvqx" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.519980 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e23712e1-8705-4d4b-bd1a-1bea23f765df-service-ca-bundle\") pod \"authentication-operator-69f744f599-k4hj2\" (UID: \"e23712e1-8705-4d4b-bd1a-1bea23f765df\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-k4hj2" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.520000 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2w99\" (UniqueName: \"kubernetes.io/projected/10bbf031-b282-4a4e-8fb9-e757afa8c022-kube-api-access-q2w99\") pod \"packageserver-d55dfcdfc-fc2g9\" (UID: \"10bbf031-b282-4a4e-8fb9-e757afa8c022\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fc2g9" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.520024 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n56hr\" (UniqueName: \"kubernetes.io/projected/5863644d-bf1b-4dbf-9c89-99ec5bff14e3-kube-api-access-n56hr\") pod \"openshift-apiserver-operator-796bbdcf4f-476wv\" (UID: \"5863644d-bf1b-4dbf-9c89-99ec5bff14e3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-476wv" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.520086 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5cf1f97f-ff78-4047-b839-6e8ffb6b92b9-metrics-tls\") pod \"ingress-operator-5b745b69d9-n7q4s\" (UID: \"5cf1f97f-ff78-4047-b839-6e8ffb6b92b9\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n7q4s" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.520118 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e2656d73-dfaf-4d40-a1b8-0d08297f9dc4-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-ffl2m\" (UID: \"e2656d73-dfaf-4d40-a1b8-0d08297f9dc4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ffl2m" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.520138 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67bjm\" (UniqueName: \"kubernetes.io/projected/97fc0814-92d4-42a4-89c0-acb55bf47b4d-kube-api-access-67bjm\") pod \"kube-storage-version-migrator-operator-b67b599dd-hd47j\" (UID: \"97fc0814-92d4-42a4-89c0-acb55bf47b4d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hd47j" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.520203 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/34cf8610-cc4e-46de-a00c-ff41d7030e1f-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-h7rbb\" (UID: \"34cf8610-cc4e-46de-a00c-ff41d7030e1f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h7rbb" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.520224 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwrgg\" (UniqueName: \"kubernetes.io/projected/cb6875f9-0bd5-4076-848f-fd734d4f04df-kube-api-access-gwrgg\") pod \"package-server-manager-789f6589d5-z5tfd\" (UID: \"cb6875f9-0bd5-4076-848f-fd734d4f04df\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-z5tfd" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.520286 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/96800ba3-0ee4-42e0-b8f7-316ef57a6173-installation-pull-secrets\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.520303 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/ac102463-bf2a-466d-b24a-e3249661d1ee-machine-approver-tls\") pod \"machine-approver-56656f9798-d7vvw\" (UID: \"ac102463-bf2a-466d-b24a-e3249661d1ee\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-d7vvw" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.520324 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhs55\" (UniqueName: \"kubernetes.io/projected/e09b8f75-2286-4a12-9571-5342b681690f-kube-api-access-lhs55\") pod \"route-controller-manager-6576b87f9c-mqvqx\" (UID: \"e09b8f75-2286-4a12-9571-5342b681690f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mqvqx" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.521925 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/a3f62807-e422-4d74-82c5-81756d963b77-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-zs6w7\" (UID: \"a3f62807-e422-4d74-82c5-81756d963b77\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zs6w7" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.521948 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a0b0911d-7b7c-4746-bcfc-d15069434bb1-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-v2tt6\" (UID: \"a0b0911d-7b7c-4746-bcfc-d15069434bb1\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-v2tt6" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.521966 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0826463c-2c8a-4b52-8727-4b07de3d86c3-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-ktld7\" (UID: \"0826463c-2c8a-4b52-8727-4b07de3d86c3\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ktld7" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.521986 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/96800ba3-0ee4-42e0-b8f7-316ef57a6173-registry-tls\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.522002 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82ece3ff-72cb-48b3-92ad-251d7bf980c0-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-pgn88\" (UID: \"82ece3ff-72cb-48b3-92ad-251d7bf980c0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pgn88" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.522042 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gp9h8\" (UniqueName: \"kubernetes.io/projected/a0b0911d-7b7c-4746-bcfc-d15069434bb1-kube-api-access-gp9h8\") pod \"machine-config-controller-84d6567774-v2tt6\" (UID: \"a0b0911d-7b7c-4746-bcfc-d15069434bb1\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-v2tt6" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.522062 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/10bbf031-b282-4a4e-8fb9-e757afa8c022-tmpfs\") pod \"packageserver-d55dfcdfc-fc2g9\" (UID: \"10bbf031-b282-4a4e-8fb9-e757afa8c022\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fc2g9" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.522087 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/21edd566-f5c3-425d-b719-d68d440884d8-metrics-tls\") pod \"dns-operator-744455d44c-46rkq\" (UID: \"21edd566-f5c3-425d-b719-d68d440884d8\") " pod="openshift-dns-operator/dns-operator-744455d44c-46rkq" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.522797 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/96800ba3-0ee4-42e0-b8f7-316ef57a6173-trusted-ca\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.522861 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97fc0814-92d4-42a4-89c0-acb55bf47b4d-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-hd47j\" (UID: \"97fc0814-92d4-42a4-89c0-acb55bf47b4d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hd47j" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.522962 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/96800ba3-0ee4-42e0-b8f7-316ef57a6173-bound-sa-token\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.523002 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ac102463-bf2a-466d-b24a-e3249661d1ee-auth-proxy-config\") pod \"machine-approver-56656f9798-d7vvw\" (UID: \"ac102463-bf2a-466d-b24a-e3249661d1ee\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-d7vvw" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.523033 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d153ec32-0e4e-41d6-b105-38f91ab40c14-metrics-certs\") pod \"router-default-5444994796-kxfxn\" (UID: \"d153ec32-0e4e-41d6-b105-38f91ab40c14\") " pod="openshift-ingress/router-default-5444994796-kxfxn" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.523077 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/34cf8610-cc4e-46de-a00c-ff41d7030e1f-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-h7rbb\" (UID: \"34cf8610-cc4e-46de-a00c-ff41d7030e1f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h7rbb" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.523121 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5ssx\" (UniqueName: \"kubernetes.io/projected/96800ba3-0ee4-42e0-b8f7-316ef57a6173-kube-api-access-p5ssx\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.523141 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/d153ec32-0e4e-41d6-b105-38f91ab40c14-stats-auth\") pod \"router-default-5444994796-kxfxn\" (UID: \"d153ec32-0e4e-41d6-b105-38f91ab40c14\") " pod="openshift-ingress/router-default-5444994796-kxfxn" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.523175 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/96800ba3-0ee4-42e0-b8f7-316ef57a6173-ca-trust-extracted\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.523193 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wq4rw\" (UniqueName: \"kubernetes.io/projected/21edd566-f5c3-425d-b719-d68d440884d8-kube-api-access-wq4rw\") pod \"dns-operator-744455d44c-46rkq\" (UID: \"21edd566-f5c3-425d-b719-d68d440884d8\") " pod="openshift-dns-operator/dns-operator-744455d44c-46rkq" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.523210 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/96800ba3-0ee4-42e0-b8f7-316ef57a6173-registry-certificates\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.523229 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/10bbf031-b282-4a4e-8fb9-e757afa8c022-apiservice-cert\") pod \"packageserver-d55dfcdfc-fc2g9\" (UID: \"10bbf031-b282-4a4e-8fb9-e757afa8c022\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fc2g9" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.523279 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7gsd\" (UniqueName: \"kubernetes.io/projected/e2656d73-dfaf-4d40-a1b8-0d08297f9dc4-kube-api-access-z7gsd\") pod \"multus-admission-controller-857f4d67dd-ffl2m\" (UID: \"e2656d73-dfaf-4d40-a1b8-0d08297f9dc4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ffl2m" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.523327 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/d153ec32-0e4e-41d6-b105-38f91ab40c14-default-certificate\") pod \"router-default-5444994796-kxfxn\" (UID: \"d153ec32-0e4e-41d6-b105-38f91ab40c14\") " pod="openshift-ingress/router-default-5444994796-kxfxn" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.523345 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d153ec32-0e4e-41d6-b105-38f91ab40c14-service-ca-bundle\") pod \"router-default-5444994796-kxfxn\" (UID: \"d153ec32-0e4e-41d6-b105-38f91ab40c14\") " pod="openshift-ingress/router-default-5444994796-kxfxn" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.523366 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvcwm\" (UniqueName: \"kubernetes.io/projected/acb1ddd4-f551-4d06-9021-28d7579bad7f-kube-api-access-fvcwm\") pod \"catalog-operator-68c6474976-c7vh9\" (UID: \"acb1ddd4-f551-4d06-9021-28d7579bad7f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-c7vh9" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.523441 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/cb6875f9-0bd5-4076-848f-fd734d4f04df-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-z5tfd\" (UID: \"cb6875f9-0bd5-4076-848f-fd734d4f04df\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-z5tfd" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.523460 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/b7e0b1cd-5c13-436d-8c35-c0a659f767a0-etcd-service-ca\") pod \"etcd-operator-b45778765-7qxf9\" (UID: \"b7e0b1cd-5c13-436d-8c35-c0a659f767a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7qxf9" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.523489 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e23712e1-8705-4d4b-bd1a-1bea23f765df-config\") pod \"authentication-operator-69f744f599-k4hj2\" (UID: \"e23712e1-8705-4d4b-bd1a-1bea23f765df\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-k4hj2" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.523521 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/91dd7b10-99ef-486d-a624-21cb245226c7-serving-cert\") pod \"apiserver-7bbb656c7d-b8zql\" (UID: \"91dd7b10-99ef-486d-a624-21cb245226c7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8zql" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.523544 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e09b8f75-2286-4a12-9571-5342b681690f-config\") pod \"route-controller-manager-6576b87f9c-mqvqx\" (UID: \"e09b8f75-2286-4a12-9571-5342b681690f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mqvqx" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.523586 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/79972ac1-e98f-4cd9-ba62-dd5d3b9e81e6-config-volume\") pod \"dns-default-ms5f2\" (UID: \"79972ac1-e98f-4cd9-ba62-dd5d3b9e81e6\") " pod="openshift-dns/dns-default-ms5f2" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.523605 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac102463-bf2a-466d-b24a-e3249661d1ee-config\") pod \"machine-approver-56656f9798-d7vvw\" (UID: \"ac102463-bf2a-466d-b24a-e3249661d1ee\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-d7vvw" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.523660 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0826463c-2c8a-4b52-8727-4b07de3d86c3-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-ktld7\" (UID: \"0826463c-2c8a-4b52-8727-4b07de3d86c3\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ktld7" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.523676 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b7e0b1cd-5c13-436d-8c35-c0a659f767a0-serving-cert\") pod \"etcd-operator-b45778765-7qxf9\" (UID: \"b7e0b1cd-5c13-436d-8c35-c0a659f767a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7qxf9" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.523819 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4mk6\" (UniqueName: \"kubernetes.io/projected/d153ec32-0e4e-41d6-b105-38f91ab40c14-kube-api-access-n4mk6\") pod \"router-default-5444994796-kxfxn\" (UID: \"d153ec32-0e4e-41d6-b105-38f91ab40c14\") " pod="openshift-ingress/router-default-5444994796-kxfxn" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.523837 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lc8sc\" (UniqueName: \"kubernetes.io/projected/a3f62807-e422-4d74-82c5-81756d963b77-kube-api-access-lc8sc\") pod \"control-plane-machine-set-operator-78cbb6b69f-zs6w7\" (UID: \"a3f62807-e422-4d74-82c5-81756d963b77\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zs6w7" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.523855 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/10bbf031-b282-4a4e-8fb9-e757afa8c022-webhook-cert\") pod \"packageserver-d55dfcdfc-fc2g9\" (UID: \"10bbf031-b282-4a4e-8fb9-e757afa8c022\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fc2g9" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.523871 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrft7\" (UniqueName: \"kubernetes.io/projected/e23712e1-8705-4d4b-bd1a-1bea23f765df-kube-api-access-xrft7\") pod \"authentication-operator-69f744f599-k4hj2\" (UID: \"e23712e1-8705-4d4b-bd1a-1bea23f765df\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-k4hj2" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.523898 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e23712e1-8705-4d4b-bd1a-1bea23f765df-serving-cert\") pod \"authentication-operator-69f744f599-k4hj2\" (UID: \"e23712e1-8705-4d4b-bd1a-1bea23f765df\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-k4hj2" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.523914 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97fc0814-92d4-42a4-89c0-acb55bf47b4d-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-hd47j\" (UID: \"97fc0814-92d4-42a4-89c0-acb55bf47b4d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hd47j" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.523941 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/91dd7b10-99ef-486d-a624-21cb245226c7-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-b8zql\" (UID: \"91dd7b10-99ef-486d-a624-21cb245226c7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8zql" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.523966 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmx77\" (UniqueName: \"kubernetes.io/projected/b7e0b1cd-5c13-436d-8c35-c0a659f767a0-kube-api-access-hmx77\") pod \"etcd-operator-b45778765-7qxf9\" (UID: \"b7e0b1cd-5c13-436d-8c35-c0a659f767a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7qxf9" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.523984 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/62ff675c-ff12-4c14-9c6b-ce3fb25ea214-available-featuregates\") pod \"openshift-config-operator-7777fb866f-pshcp\" (UID: \"62ff675c-ff12-4c14-9c6b-ce3fb25ea214\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-pshcp" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.524005 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/acb1ddd4-f551-4d06-9021-28d7579bad7f-profile-collector-cert\") pod \"catalog-operator-68c6474976-c7vh9\" (UID: \"acb1ddd4-f551-4d06-9021-28d7579bad7f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-c7vh9" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.524025 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/79972ac1-e98f-4cd9-ba62-dd5d3b9e81e6-metrics-tls\") pod \"dns-default-ms5f2\" (UID: \"79972ac1-e98f-4cd9-ba62-dd5d3b9e81e6\") " pod="openshift-dns/dns-default-ms5f2" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.524059 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sz4xc\" (UniqueName: \"kubernetes.io/projected/79972ac1-e98f-4cd9-ba62-dd5d3b9e81e6-kube-api-access-sz4xc\") pod \"dns-default-ms5f2\" (UID: \"79972ac1-e98f-4cd9-ba62-dd5d3b9e81e6\") " pod="openshift-dns/dns-default-ms5f2" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.524075 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/91dd7b10-99ef-486d-a624-21cb245226c7-audit-policies\") pod \"apiserver-7bbb656c7d-b8zql\" (UID: \"91dd7b10-99ef-486d-a624-21cb245226c7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8zql" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.524134 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.524153 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5cf1f97f-ff78-4047-b839-6e8ffb6b92b9-bound-sa-token\") pod \"ingress-operator-5b745b69d9-n7q4s\" (UID: \"5cf1f97f-ff78-4047-b839-6e8ffb6b92b9\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n7q4s" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.524194 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/b7e0b1cd-5c13-436d-8c35-c0a659f767a0-etcd-ca\") pod \"etcd-operator-b45778765-7qxf9\" (UID: \"b7e0b1cd-5c13-436d-8c35-c0a659f767a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7qxf9" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.524210 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b7e0b1cd-5c13-436d-8c35-c0a659f767a0-etcd-client\") pod \"etcd-operator-b45778765-7qxf9\" (UID: \"b7e0b1cd-5c13-436d-8c35-c0a659f767a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7qxf9" Nov 25 15:03:53 crc kubenswrapper[4890]: E1125 15:03:53.534856 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:03:54.034815279 +0000 UTC m=+92.477277889 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.625035 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.625185 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gp9h8\" (UniqueName: \"kubernetes.io/projected/a0b0911d-7b7c-4746-bcfc-d15069434bb1-kube-api-access-gp9h8\") pod \"machine-config-controller-84d6567774-v2tt6\" (UID: \"a0b0911d-7b7c-4746-bcfc-d15069434bb1\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-v2tt6" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.625220 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/10bbf031-b282-4a4e-8fb9-e757afa8c022-tmpfs\") pod \"packageserver-d55dfcdfc-fc2g9\" (UID: \"10bbf031-b282-4a4e-8fb9-e757afa8c022\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fc2g9" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.625240 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/21edd566-f5c3-425d-b719-d68d440884d8-metrics-tls\") pod \"dns-operator-744455d44c-46rkq\" (UID: \"21edd566-f5c3-425d-b719-d68d440884d8\") " pod="openshift-dns-operator/dns-operator-744455d44c-46rkq" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.625293 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/96800ba3-0ee4-42e0-b8f7-316ef57a6173-trusted-ca\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.625314 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbx85\" (UniqueName: \"kubernetes.io/projected/429acdcb-8b70-482c-abad-82f4867a9266-kube-api-access-rbx85\") pod \"machine-config-server-sm46z\" (UID: \"429acdcb-8b70-482c-abad-82f4867a9266\") " pod="openshift-machine-config-operator/machine-config-server-sm46z" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.625332 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/429acdcb-8b70-482c-abad-82f4867a9266-certs\") pod \"machine-config-server-sm46z\" (UID: \"429acdcb-8b70-482c-abad-82f4867a9266\") " pod="openshift-machine-config-operator/machine-config-server-sm46z" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.625351 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97fc0814-92d4-42a4-89c0-acb55bf47b4d-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-hd47j\" (UID: \"97fc0814-92d4-42a4-89c0-acb55bf47b4d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hd47j" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.625369 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/96800ba3-0ee4-42e0-b8f7-316ef57a6173-bound-sa-token\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.625384 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ac102463-bf2a-466d-b24a-e3249661d1ee-auth-proxy-config\") pod \"machine-approver-56656f9798-d7vvw\" (UID: \"ac102463-bf2a-466d-b24a-e3249661d1ee\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-d7vvw" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.625410 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d153ec32-0e4e-41d6-b105-38f91ab40c14-metrics-certs\") pod \"router-default-5444994796-kxfxn\" (UID: \"d153ec32-0e4e-41d6-b105-38f91ab40c14\") " pod="openshift-ingress/router-default-5444994796-kxfxn" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.625432 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/34cf8610-cc4e-46de-a00c-ff41d7030e1f-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-h7rbb\" (UID: \"34cf8610-cc4e-46de-a00c-ff41d7030e1f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h7rbb" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.625450 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/ba947c9c-cd84-4c9b-a431-9276412058d1-csi-data-dir\") pod \"csi-hostpathplugin-cmpbm\" (UID: \"ba947c9c-cd84-4c9b-a431-9276412058d1\") " pod="hostpath-provisioner/csi-hostpathplugin-cmpbm" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.625466 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5ssx\" (UniqueName: \"kubernetes.io/projected/96800ba3-0ee4-42e0-b8f7-316ef57a6173-kube-api-access-p5ssx\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.625483 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/d153ec32-0e4e-41d6-b105-38f91ab40c14-stats-auth\") pod \"router-default-5444994796-kxfxn\" (UID: \"d153ec32-0e4e-41d6-b105-38f91ab40c14\") " pod="openshift-ingress/router-default-5444994796-kxfxn" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.625503 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/96800ba3-0ee4-42e0-b8f7-316ef57a6173-ca-trust-extracted\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.625520 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wq4rw\" (UniqueName: \"kubernetes.io/projected/21edd566-f5c3-425d-b719-d68d440884d8-kube-api-access-wq4rw\") pod \"dns-operator-744455d44c-46rkq\" (UID: \"21edd566-f5c3-425d-b719-d68d440884d8\") " pod="openshift-dns-operator/dns-operator-744455d44c-46rkq" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.625536 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/96800ba3-0ee4-42e0-b8f7-316ef57a6173-registry-certificates\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.625551 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/10bbf031-b282-4a4e-8fb9-e757afa8c022-apiservice-cert\") pod \"packageserver-d55dfcdfc-fc2g9\" (UID: \"10bbf031-b282-4a4e-8fb9-e757afa8c022\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fc2g9" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.625566 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7gsd\" (UniqueName: \"kubernetes.io/projected/e2656d73-dfaf-4d40-a1b8-0d08297f9dc4-kube-api-access-z7gsd\") pod \"multus-admission-controller-857f4d67dd-ffl2m\" (UID: \"e2656d73-dfaf-4d40-a1b8-0d08297f9dc4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ffl2m" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.625584 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8ef9a0ab-2131-42ff-823d-17d517929e77-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xmhq9\" (UID: \"8ef9a0ab-2131-42ff-823d-17d517929e77\") " pod="openshift-marketplace/marketplace-operator-79b997595-xmhq9" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.625610 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4622b25d-fb03-4e9f-a66c-7f90def9b5f7-secret-volume\") pod \"collect-profiles-29401380-96zpb\" (UID: \"4622b25d-fb03-4e9f-a66c-7f90def9b5f7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401380-96zpb" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.625628 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/d153ec32-0e4e-41d6-b105-38f91ab40c14-default-certificate\") pod \"router-default-5444994796-kxfxn\" (UID: \"d153ec32-0e4e-41d6-b105-38f91ab40c14\") " pod="openshift-ingress/router-default-5444994796-kxfxn" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.625644 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d153ec32-0e4e-41d6-b105-38f91ab40c14-service-ca-bundle\") pod \"router-default-5444994796-kxfxn\" (UID: \"d153ec32-0e4e-41d6-b105-38f91ab40c14\") " pod="openshift-ingress/router-default-5444994796-kxfxn" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.625663 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvcwm\" (UniqueName: \"kubernetes.io/projected/acb1ddd4-f551-4d06-9021-28d7579bad7f-kube-api-access-fvcwm\") pod \"catalog-operator-68c6474976-c7vh9\" (UID: \"acb1ddd4-f551-4d06-9021-28d7579bad7f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-c7vh9" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.625680 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjjcv\" (UniqueName: \"kubernetes.io/projected/4622b25d-fb03-4e9f-a66c-7f90def9b5f7-kube-api-access-qjjcv\") pod \"collect-profiles-29401380-96zpb\" (UID: \"4622b25d-fb03-4e9f-a66c-7f90def9b5f7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401380-96zpb" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.625700 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/cb6875f9-0bd5-4076-848f-fd734d4f04df-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-z5tfd\" (UID: \"cb6875f9-0bd5-4076-848f-fd734d4f04df\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-z5tfd" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.625720 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/b7e0b1cd-5c13-436d-8c35-c0a659f767a0-etcd-service-ca\") pod \"etcd-operator-b45778765-7qxf9\" (UID: \"b7e0b1cd-5c13-436d-8c35-c0a659f767a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7qxf9" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.625737 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xn8v6\" (UniqueName: \"kubernetes.io/projected/815dd5c7-dda6-42c5-b63f-8e354c5d1143-kube-api-access-xn8v6\") pod \"ingress-canary-rx9dj\" (UID: \"815dd5c7-dda6-42c5-b63f-8e354c5d1143\") " pod="openshift-ingress-canary/ingress-canary-rx9dj" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.625754 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/429acdcb-8b70-482c-abad-82f4867a9266-node-bootstrap-token\") pod \"machine-config-server-sm46z\" (UID: \"429acdcb-8b70-482c-abad-82f4867a9266\") " pod="openshift-machine-config-operator/machine-config-server-sm46z" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.625775 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e23712e1-8705-4d4b-bd1a-1bea23f765df-config\") pod \"authentication-operator-69f744f599-k4hj2\" (UID: \"e23712e1-8705-4d4b-bd1a-1bea23f765df\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-k4hj2" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.626021 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vj4n4\" (UniqueName: \"kubernetes.io/projected/d5dd7cd5-c63e-460f-8d90-d91f5ec85f15-kube-api-access-vj4n4\") pod \"machine-config-operator-74547568cd-jkwk5\" (UID: \"d5dd7cd5-c63e-460f-8d90-d91f5ec85f15\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jkwk5" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.626050 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/91dd7b10-99ef-486d-a624-21cb245226c7-serving-cert\") pod \"apiserver-7bbb656c7d-b8zql\" (UID: \"91dd7b10-99ef-486d-a624-21cb245226c7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8zql" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.626068 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvnfh\" (UniqueName: \"kubernetes.io/projected/8ef9a0ab-2131-42ff-823d-17d517929e77-kube-api-access-wvnfh\") pod \"marketplace-operator-79b997595-xmhq9\" (UID: \"8ef9a0ab-2131-42ff-823d-17d517929e77\") " pod="openshift-marketplace/marketplace-operator-79b997595-xmhq9" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.626086 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8ef9a0ab-2131-42ff-823d-17d517929e77-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xmhq9\" (UID: \"8ef9a0ab-2131-42ff-823d-17d517929e77\") " pod="openshift-marketplace/marketplace-operator-79b997595-xmhq9" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.626105 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e09b8f75-2286-4a12-9571-5342b681690f-config\") pod \"route-controller-manager-6576b87f9c-mqvqx\" (UID: \"e09b8f75-2286-4a12-9571-5342b681690f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mqvqx" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.626120 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/ba947c9c-cd84-4c9b-a431-9276412058d1-mountpoint-dir\") pod \"csi-hostpathplugin-cmpbm\" (UID: \"ba947c9c-cd84-4c9b-a431-9276412058d1\") " pod="hostpath-provisioner/csi-hostpathplugin-cmpbm" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.626136 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ba947c9c-cd84-4c9b-a431-9276412058d1-registration-dir\") pod \"csi-hostpathplugin-cmpbm\" (UID: \"ba947c9c-cd84-4c9b-a431-9276412058d1\") " pod="hostpath-provisioner/csi-hostpathplugin-cmpbm" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.626192 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/79972ac1-e98f-4cd9-ba62-dd5d3b9e81e6-config-volume\") pod \"dns-default-ms5f2\" (UID: \"79972ac1-e98f-4cd9-ba62-dd5d3b9e81e6\") " pod="openshift-dns/dns-default-ms5f2" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.626218 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/e682b6af-7f62-4efb-917e-7453e78138bc-signing-cabundle\") pod \"service-ca-9c57cc56f-klnk6\" (UID: \"e682b6af-7f62-4efb-917e-7453e78138bc\") " pod="openshift-service-ca/service-ca-9c57cc56f-klnk6" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.626242 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac102463-bf2a-466d-b24a-e3249661d1ee-config\") pod \"machine-approver-56656f9798-d7vvw\" (UID: \"ac102463-bf2a-466d-b24a-e3249661d1ee\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-d7vvw" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.626270 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0826463c-2c8a-4b52-8727-4b07de3d86c3-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-ktld7\" (UID: \"0826463c-2c8a-4b52-8727-4b07de3d86c3\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ktld7" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.626293 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b7e0b1cd-5c13-436d-8c35-c0a659f767a0-serving-cert\") pod \"etcd-operator-b45778765-7qxf9\" (UID: \"b7e0b1cd-5c13-436d-8c35-c0a659f767a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7qxf9" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.626315 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/e682b6af-7f62-4efb-917e-7453e78138bc-signing-key\") pod \"service-ca-9c57cc56f-klnk6\" (UID: \"e682b6af-7f62-4efb-917e-7453e78138bc\") " pod="openshift-service-ca/service-ca-9c57cc56f-klnk6" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.626502 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57qks\" (UniqueName: \"kubernetes.io/projected/e682b6af-7f62-4efb-917e-7453e78138bc-kube-api-access-57qks\") pod \"service-ca-9c57cc56f-klnk6\" (UID: \"e682b6af-7f62-4efb-917e-7453e78138bc\") " pod="openshift-service-ca/service-ca-9c57cc56f-klnk6" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.626530 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4mk6\" (UniqueName: \"kubernetes.io/projected/d153ec32-0e4e-41d6-b105-38f91ab40c14-kube-api-access-n4mk6\") pod \"router-default-5444994796-kxfxn\" (UID: \"d153ec32-0e4e-41d6-b105-38f91ab40c14\") " pod="openshift-ingress/router-default-5444994796-kxfxn" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.626548 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lc8sc\" (UniqueName: \"kubernetes.io/projected/a3f62807-e422-4d74-82c5-81756d963b77-kube-api-access-lc8sc\") pod \"control-plane-machine-set-operator-78cbb6b69f-zs6w7\" (UID: \"a3f62807-e422-4d74-82c5-81756d963b77\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zs6w7" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.626562 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/10bbf031-b282-4a4e-8fb9-e757afa8c022-webhook-cert\") pod \"packageserver-d55dfcdfc-fc2g9\" (UID: \"10bbf031-b282-4a4e-8fb9-e757afa8c022\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fc2g9" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.626589 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrft7\" (UniqueName: \"kubernetes.io/projected/e23712e1-8705-4d4b-bd1a-1bea23f765df-kube-api-access-xrft7\") pod \"authentication-operator-69f744f599-k4hj2\" (UID: \"e23712e1-8705-4d4b-bd1a-1bea23f765df\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-k4hj2" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.626606 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e3475b27-002a-4087-ab65-9cdbad01ca92-profile-collector-cert\") pod \"olm-operator-6b444d44fb-nfp7k\" (UID: \"e3475b27-002a-4087-ab65-9cdbad01ca92\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nfp7k" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.626622 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e23712e1-8705-4d4b-bd1a-1bea23f765df-serving-cert\") pod \"authentication-operator-69f744f599-k4hj2\" (UID: \"e23712e1-8705-4d4b-bd1a-1bea23f765df\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-k4hj2" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.626638 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97fc0814-92d4-42a4-89c0-acb55bf47b4d-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-hd47j\" (UID: \"97fc0814-92d4-42a4-89c0-acb55bf47b4d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hd47j" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.626654 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4622b25d-fb03-4e9f-a66c-7f90def9b5f7-config-volume\") pod \"collect-profiles-29401380-96zpb\" (UID: \"4622b25d-fb03-4e9f-a66c-7f90def9b5f7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401380-96zpb" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.626675 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/91dd7b10-99ef-486d-a624-21cb245226c7-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-b8zql\" (UID: \"91dd7b10-99ef-486d-a624-21cb245226c7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8zql" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.626690 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmx77\" (UniqueName: \"kubernetes.io/projected/b7e0b1cd-5c13-436d-8c35-c0a659f767a0-kube-api-access-hmx77\") pod \"etcd-operator-b45778765-7qxf9\" (UID: \"b7e0b1cd-5c13-436d-8c35-c0a659f767a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7qxf9" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.626709 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/62ff675c-ff12-4c14-9c6b-ce3fb25ea214-available-featuregates\") pod \"openshift-config-operator-7777fb866f-pshcp\" (UID: \"62ff675c-ff12-4c14-9c6b-ce3fb25ea214\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-pshcp" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.626743 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/acb1ddd4-f551-4d06-9021-28d7579bad7f-profile-collector-cert\") pod \"catalog-operator-68c6474976-c7vh9\" (UID: \"acb1ddd4-f551-4d06-9021-28d7579bad7f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-c7vh9" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.626778 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/79972ac1-e98f-4cd9-ba62-dd5d3b9e81e6-metrics-tls\") pod \"dns-default-ms5f2\" (UID: \"79972ac1-e98f-4cd9-ba62-dd5d3b9e81e6\") " pod="openshift-dns/dns-default-ms5f2" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.626803 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9hld\" (UniqueName: \"kubernetes.io/projected/ba947c9c-cd84-4c9b-a431-9276412058d1-kube-api-access-g9hld\") pod \"csi-hostpathplugin-cmpbm\" (UID: \"ba947c9c-cd84-4c9b-a431-9276412058d1\") " pod="hostpath-provisioner/csi-hostpathplugin-cmpbm" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.626827 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sz4xc\" (UniqueName: \"kubernetes.io/projected/79972ac1-e98f-4cd9-ba62-dd5d3b9e81e6-kube-api-access-sz4xc\") pod \"dns-default-ms5f2\" (UID: \"79972ac1-e98f-4cd9-ba62-dd5d3b9e81e6\") " pod="openshift-dns/dns-default-ms5f2" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.626851 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/91dd7b10-99ef-486d-a624-21cb245226c7-audit-policies\") pod \"apiserver-7bbb656c7d-b8zql\" (UID: \"91dd7b10-99ef-486d-a624-21cb245226c7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8zql" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.626887 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5cf1f97f-ff78-4047-b839-6e8ffb6b92b9-bound-sa-token\") pod \"ingress-operator-5b745b69d9-n7q4s\" (UID: \"5cf1f97f-ff78-4047-b839-6e8ffb6b92b9\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n7q4s" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.626921 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/b7e0b1cd-5c13-436d-8c35-c0a659f767a0-etcd-ca\") pod \"etcd-operator-b45778765-7qxf9\" (UID: \"b7e0b1cd-5c13-436d-8c35-c0a659f767a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7qxf9" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.626935 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b7e0b1cd-5c13-436d-8c35-c0a659f767a0-etcd-client\") pod \"etcd-operator-b45778765-7qxf9\" (UID: \"b7e0b1cd-5c13-436d-8c35-c0a659f767a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7qxf9" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.626950 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/815dd5c7-dda6-42c5-b63f-8e354c5d1143-cert\") pod \"ingress-canary-rx9dj\" (UID: \"815dd5c7-dda6-42c5-b63f-8e354c5d1143\") " pod="openshift-ingress-canary/ingress-canary-rx9dj" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.626966 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34cf8610-cc4e-46de-a00c-ff41d7030e1f-config\") pod \"kube-apiserver-operator-766d6c64bb-h7rbb\" (UID: \"34cf8610-cc4e-46de-a00c-ff41d7030e1f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h7rbb" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.626980 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/300178a7-3aa1-4fbb-a604-a37365cc968f-serving-cert\") pod \"service-ca-operator-777779d784-99n87\" (UID: \"300178a7-3aa1-4fbb-a604-a37365cc968f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-99n87" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.626995 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ba947c9c-cd84-4c9b-a431-9276412058d1-socket-dir\") pod \"csi-hostpathplugin-cmpbm\" (UID: \"ba947c9c-cd84-4c9b-a431-9276412058d1\") " pod="hostpath-provisioner/csi-hostpathplugin-cmpbm" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.627011 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/91dd7b10-99ef-486d-a624-21cb245226c7-encryption-config\") pod \"apiserver-7bbb656c7d-b8zql\" (UID: \"91dd7b10-99ef-486d-a624-21cb245226c7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8zql" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.627028 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e23712e1-8705-4d4b-bd1a-1bea23f765df-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-k4hj2\" (UID: \"e23712e1-8705-4d4b-bd1a-1bea23f765df\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-k4hj2" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.627048 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgh2l\" (UniqueName: \"kubernetes.io/projected/62ff675c-ff12-4c14-9c6b-ce3fb25ea214-kube-api-access-jgh2l\") pod \"openshift-config-operator-7777fb866f-pshcp\" (UID: \"62ff675c-ff12-4c14-9c6b-ce3fb25ea214\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-pshcp" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.627065 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/82ece3ff-72cb-48b3-92ad-251d7bf980c0-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-pgn88\" (UID: \"82ece3ff-72cb-48b3-92ad-251d7bf980c0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pgn88" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.627089 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxr96\" (UniqueName: \"kubernetes.io/projected/91dd7b10-99ef-486d-a624-21cb245226c7-kube-api-access-xxr96\") pod \"apiserver-7bbb656c7d-b8zql\" (UID: \"91dd7b10-99ef-486d-a624-21cb245226c7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8zql" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.627104 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/62ff675c-ff12-4c14-9c6b-ce3fb25ea214-serving-cert\") pod \"openshift-config-operator-7777fb866f-pshcp\" (UID: \"62ff675c-ff12-4c14-9c6b-ce3fb25ea214\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-pshcp" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.627131 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a0b0911d-7b7c-4746-bcfc-d15069434bb1-proxy-tls\") pod \"machine-config-controller-84d6567774-v2tt6\" (UID: \"a0b0911d-7b7c-4746-bcfc-d15069434bb1\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-v2tt6" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.627149 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5863644d-bf1b-4dbf-9c89-99ec5bff14e3-config\") pod \"openshift-apiserver-operator-796bbdcf4f-476wv\" (UID: \"5863644d-bf1b-4dbf-9c89-99ec5bff14e3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-476wv" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.627181 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/91dd7b10-99ef-486d-a624-21cb245226c7-audit-dir\") pod \"apiserver-7bbb656c7d-b8zql\" (UID: \"91dd7b10-99ef-486d-a624-21cb245226c7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8zql" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.627198 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/acb1ddd4-f551-4d06-9021-28d7579bad7f-srv-cert\") pod \"catalog-operator-68c6474976-c7vh9\" (UID: \"acb1ddd4-f551-4d06-9021-28d7579bad7f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-c7vh9" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.627213 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7e0b1cd-5c13-436d-8c35-c0a659f767a0-config\") pod \"etcd-operator-b45778765-7qxf9\" (UID: \"b7e0b1cd-5c13-436d-8c35-c0a659f767a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7qxf9" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.627257 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5cf1f97f-ff78-4047-b839-6e8ffb6b92b9-trusted-ca\") pod \"ingress-operator-5b745b69d9-n7q4s\" (UID: \"5cf1f97f-ff78-4047-b839-6e8ffb6b92b9\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n7q4s" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.627278 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e3475b27-002a-4087-ab65-9cdbad01ca92-srv-cert\") pod \"olm-operator-6b444d44fb-nfp7k\" (UID: \"e3475b27-002a-4087-ab65-9cdbad01ca92\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nfp7k" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.627297 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d5dd7cd5-c63e-460f-8d90-d91f5ec85f15-proxy-tls\") pod \"machine-config-operator-74547568cd-jkwk5\" (UID: \"d5dd7cd5-c63e-460f-8d90-d91f5ec85f15\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jkwk5" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.627318 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kk9j6\" (UniqueName: \"kubernetes.io/projected/5cf1f97f-ff78-4047-b839-6e8ffb6b92b9-kube-api-access-kk9j6\") pod \"ingress-operator-5b745b69d9-n7q4s\" (UID: \"5cf1f97f-ff78-4047-b839-6e8ffb6b92b9\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n7q4s" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.627335 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/91dd7b10-99ef-486d-a624-21cb245226c7-etcd-client\") pod \"apiserver-7bbb656c7d-b8zql\" (UID: \"91dd7b10-99ef-486d-a624-21cb245226c7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8zql" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.627353 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6vvk\" (UniqueName: \"kubernetes.io/projected/ac102463-bf2a-466d-b24a-e3249661d1ee-kube-api-access-q6vvk\") pod \"machine-approver-56656f9798-d7vvw\" (UID: \"ac102463-bf2a-466d-b24a-e3249661d1ee\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-d7vvw" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.627369 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtx7v\" (UniqueName: \"kubernetes.io/projected/e3475b27-002a-4087-ab65-9cdbad01ca92-kube-api-access-qtx7v\") pod \"olm-operator-6b444d44fb-nfp7k\" (UID: \"e3475b27-002a-4087-ab65-9cdbad01ca92\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nfp7k" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.627390 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e09b8f75-2286-4a12-9571-5342b681690f-client-ca\") pod \"route-controller-manager-6576b87f9c-mqvqx\" (UID: \"e09b8f75-2286-4a12-9571-5342b681690f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mqvqx" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.627407 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82ece3ff-72cb-48b3-92ad-251d7bf980c0-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-pgn88\" (UID: \"82ece3ff-72cb-48b3-92ad-251d7bf980c0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pgn88" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.627424 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvshl\" (UniqueName: \"kubernetes.io/projected/1a324db0-9f3e-417e-931f-5b74a44775ed-kube-api-access-zvshl\") pod \"migrator-59844c95c7-t5v4f\" (UID: \"1a324db0-9f3e-417e-931f-5b74a44775ed\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t5v4f" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.627440 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0826463c-2c8a-4b52-8727-4b07de3d86c3-config\") pod \"kube-controller-manager-operator-78b949d7b-ktld7\" (UID: \"0826463c-2c8a-4b52-8727-4b07de3d86c3\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ktld7" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.627457 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d5dd7cd5-c63e-460f-8d90-d91f5ec85f15-images\") pod \"machine-config-operator-74547568cd-jkwk5\" (UID: \"d5dd7cd5-c63e-460f-8d90-d91f5ec85f15\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jkwk5" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.627474 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/91dd7b10-99ef-486d-a624-21cb245226c7-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-b8zql\" (UID: \"91dd7b10-99ef-486d-a624-21cb245226c7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8zql" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.627491 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5863644d-bf1b-4dbf-9c89-99ec5bff14e3-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-476wv\" (UID: \"5863644d-bf1b-4dbf-9c89-99ec5bff14e3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-476wv" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.627508 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/300178a7-3aa1-4fbb-a604-a37365cc968f-config\") pod \"service-ca-operator-777779d784-99n87\" (UID: \"300178a7-3aa1-4fbb-a604-a37365cc968f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-99n87" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.627525 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e09b8f75-2286-4a12-9571-5342b681690f-serving-cert\") pod \"route-controller-manager-6576b87f9c-mqvqx\" (UID: \"e09b8f75-2286-4a12-9571-5342b681690f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mqvqx" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.627541 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e23712e1-8705-4d4b-bd1a-1bea23f765df-service-ca-bundle\") pod \"authentication-operator-69f744f599-k4hj2\" (UID: \"e23712e1-8705-4d4b-bd1a-1bea23f765df\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-k4hj2" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.627559 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2w99\" (UniqueName: \"kubernetes.io/projected/10bbf031-b282-4a4e-8fb9-e757afa8c022-kube-api-access-q2w99\") pod \"packageserver-d55dfcdfc-fc2g9\" (UID: \"10bbf031-b282-4a4e-8fb9-e757afa8c022\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fc2g9" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.627578 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d5dd7cd5-c63e-460f-8d90-d91f5ec85f15-auth-proxy-config\") pod \"machine-config-operator-74547568cd-jkwk5\" (UID: \"d5dd7cd5-c63e-460f-8d90-d91f5ec85f15\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jkwk5" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.627603 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n56hr\" (UniqueName: \"kubernetes.io/projected/5863644d-bf1b-4dbf-9c89-99ec5bff14e3-kube-api-access-n56hr\") pod \"openshift-apiserver-operator-796bbdcf4f-476wv\" (UID: \"5863644d-bf1b-4dbf-9c89-99ec5bff14e3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-476wv" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.627688 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5cf1f97f-ff78-4047-b839-6e8ffb6b92b9-metrics-tls\") pod \"ingress-operator-5b745b69d9-n7q4s\" (UID: \"5cf1f97f-ff78-4047-b839-6e8ffb6b92b9\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n7q4s" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.627749 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e2656d73-dfaf-4d40-a1b8-0d08297f9dc4-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-ffl2m\" (UID: \"e2656d73-dfaf-4d40-a1b8-0d08297f9dc4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ffl2m" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.627803 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67bjm\" (UniqueName: \"kubernetes.io/projected/97fc0814-92d4-42a4-89c0-acb55bf47b4d-kube-api-access-67bjm\") pod \"kube-storage-version-migrator-operator-b67b599dd-hd47j\" (UID: \"97fc0814-92d4-42a4-89c0-acb55bf47b4d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hd47j" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.627841 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/34cf8610-cc4e-46de-a00c-ff41d7030e1f-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-h7rbb\" (UID: \"34cf8610-cc4e-46de-a00c-ff41d7030e1f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h7rbb" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.627871 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwrgg\" (UniqueName: \"kubernetes.io/projected/cb6875f9-0bd5-4076-848f-fd734d4f04df-kube-api-access-gwrgg\") pod \"package-server-manager-789f6589d5-z5tfd\" (UID: \"cb6875f9-0bd5-4076-848f-fd734d4f04df\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-z5tfd" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.627900 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/96800ba3-0ee4-42e0-b8f7-316ef57a6173-installation-pull-secrets\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.627924 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/ac102463-bf2a-466d-b24a-e3249661d1ee-machine-approver-tls\") pod \"machine-approver-56656f9798-d7vvw\" (UID: \"ac102463-bf2a-466d-b24a-e3249661d1ee\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-d7vvw" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.627951 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/ba947c9c-cd84-4c9b-a431-9276412058d1-plugins-dir\") pod \"csi-hostpathplugin-cmpbm\" (UID: \"ba947c9c-cd84-4c9b-a431-9276412058d1\") " pod="hostpath-provisioner/csi-hostpathplugin-cmpbm" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.627974 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhs55\" (UniqueName: \"kubernetes.io/projected/e09b8f75-2286-4a12-9571-5342b681690f-kube-api-access-lhs55\") pod \"route-controller-manager-6576b87f9c-mqvqx\" (UID: \"e09b8f75-2286-4a12-9571-5342b681690f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mqvqx" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.627999 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/a3f62807-e422-4d74-82c5-81756d963b77-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-zs6w7\" (UID: \"a3f62807-e422-4d74-82c5-81756d963b77\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zs6w7" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.628030 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a0b0911d-7b7c-4746-bcfc-d15069434bb1-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-v2tt6\" (UID: \"a0b0911d-7b7c-4746-bcfc-d15069434bb1\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-v2tt6" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.628057 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0826463c-2c8a-4b52-8727-4b07de3d86c3-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-ktld7\" (UID: \"0826463c-2c8a-4b52-8727-4b07de3d86c3\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ktld7" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.628090 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/96800ba3-0ee4-42e0-b8f7-316ef57a6173-registry-tls\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.628122 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82ece3ff-72cb-48b3-92ad-251d7bf980c0-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-pgn88\" (UID: \"82ece3ff-72cb-48b3-92ad-251d7bf980c0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pgn88" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.628151 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpgtd\" (UniqueName: \"kubernetes.io/projected/300178a7-3aa1-4fbb-a604-a37365cc968f-kube-api-access-fpgtd\") pod \"service-ca-operator-777779d784-99n87\" (UID: \"300178a7-3aa1-4fbb-a604-a37365cc968f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-99n87" Nov 25 15:03:53 crc kubenswrapper[4890]: E1125 15:03:53.628370 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:03:54.128340628 +0000 UTC m=+92.570803448 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.628934 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/10bbf031-b282-4a4e-8fb9-e757afa8c022-tmpfs\") pod \"packageserver-d55dfcdfc-fc2g9\" (UID: \"10bbf031-b282-4a4e-8fb9-e757afa8c022\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fc2g9" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.632107 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/79972ac1-e98f-4cd9-ba62-dd5d3b9e81e6-config-volume\") pod \"dns-default-ms5f2\" (UID: \"79972ac1-e98f-4cd9-ba62-dd5d3b9e81e6\") " pod="openshift-dns/dns-default-ms5f2" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.633004 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac102463-bf2a-466d-b24a-e3249661d1ee-config\") pod \"machine-approver-56656f9798-d7vvw\" (UID: \"ac102463-bf2a-466d-b24a-e3249661d1ee\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-d7vvw" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.634230 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/b7e0b1cd-5c13-436d-8c35-c0a659f767a0-etcd-service-ca\") pod \"etcd-operator-b45778765-7qxf9\" (UID: \"b7e0b1cd-5c13-436d-8c35-c0a659f767a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7qxf9" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.635362 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e09b8f75-2286-4a12-9571-5342b681690f-config\") pod \"route-controller-manager-6576b87f9c-mqvqx\" (UID: \"e09b8f75-2286-4a12-9571-5342b681690f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mqvqx" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.635515 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a0b0911d-7b7c-4746-bcfc-d15069434bb1-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-v2tt6\" (UID: \"a0b0911d-7b7c-4746-bcfc-d15069434bb1\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-v2tt6" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.636285 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34cf8610-cc4e-46de-a00c-ff41d7030e1f-config\") pod \"kube-apiserver-operator-766d6c64bb-h7rbb\" (UID: \"34cf8610-cc4e-46de-a00c-ff41d7030e1f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h7rbb" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.636549 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/91dd7b10-99ef-486d-a624-21cb245226c7-serving-cert\") pod \"apiserver-7bbb656c7d-b8zql\" (UID: \"91dd7b10-99ef-486d-a624-21cb245226c7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8zql" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.637108 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/96800ba3-0ee4-42e0-b8f7-316ef57a6173-registry-certificates\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.637187 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5cf1f97f-ff78-4047-b839-6e8ffb6b92b9-metrics-tls\") pod \"ingress-operator-5b745b69d9-n7q4s\" (UID: \"5cf1f97f-ff78-4047-b839-6e8ffb6b92b9\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n7q4s" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.638038 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b7e0b1cd-5c13-436d-8c35-c0a659f767a0-etcd-client\") pod \"etcd-operator-b45778765-7qxf9\" (UID: \"b7e0b1cd-5c13-436d-8c35-c0a659f767a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7qxf9" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.638475 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/91dd7b10-99ef-486d-a624-21cb245226c7-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-b8zql\" (UID: \"91dd7b10-99ef-486d-a624-21cb245226c7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8zql" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.638817 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/62ff675c-ff12-4c14-9c6b-ce3fb25ea214-available-featuregates\") pod \"openshift-config-operator-7777fb866f-pshcp\" (UID: \"62ff675c-ff12-4c14-9c6b-ce3fb25ea214\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-pshcp" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.640466 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/21edd566-f5c3-425d-b719-d68d440884d8-metrics-tls\") pod \"dns-operator-744455d44c-46rkq\" (UID: \"21edd566-f5c3-425d-b719-d68d440884d8\") " pod="openshift-dns-operator/dns-operator-744455d44c-46rkq" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.641266 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/cb6875f9-0bd5-4076-848f-fd734d4f04df-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-z5tfd\" (UID: \"cb6875f9-0bd5-4076-848f-fd734d4f04df\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-z5tfd" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.641643 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/ac102463-bf2a-466d-b24a-e3249661d1ee-machine-approver-tls\") pod \"machine-approver-56656f9798-d7vvw\" (UID: \"ac102463-bf2a-466d-b24a-e3249661d1ee\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-d7vvw" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.641640 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7e0b1cd-5c13-436d-8c35-c0a659f767a0-config\") pod \"etcd-operator-b45778765-7qxf9\" (UID: \"b7e0b1cd-5c13-436d-8c35-c0a659f767a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7qxf9" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.642620 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d153ec32-0e4e-41d6-b105-38f91ab40c14-service-ca-bundle\") pod \"router-default-5444994796-kxfxn\" (UID: \"d153ec32-0e4e-41d6-b105-38f91ab40c14\") " pod="openshift-ingress/router-default-5444994796-kxfxn" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.642788 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/96800ba3-0ee4-42e0-b8f7-316ef57a6173-trusted-ca\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.642994 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97fc0814-92d4-42a4-89c0-acb55bf47b4d-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-hd47j\" (UID: \"97fc0814-92d4-42a4-89c0-acb55bf47b4d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hd47j" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.643035 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/acb1ddd4-f551-4d06-9021-28d7579bad7f-profile-collector-cert\") pod \"catalog-operator-68c6474976-c7vh9\" (UID: \"acb1ddd4-f551-4d06-9021-28d7579bad7f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-c7vh9" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.643519 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/91dd7b10-99ef-486d-a624-21cb245226c7-audit-policies\") pod \"apiserver-7bbb656c7d-b8zql\" (UID: \"91dd7b10-99ef-486d-a624-21cb245226c7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8zql" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.644206 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/b7e0b1cd-5c13-436d-8c35-c0a659f767a0-etcd-ca\") pod \"etcd-operator-b45778765-7qxf9\" (UID: \"b7e0b1cd-5c13-436d-8c35-c0a659f767a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7qxf9" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.644517 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/96800ba3-0ee4-42e0-b8f7-316ef57a6173-ca-trust-extracted\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.645679 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/10bbf031-b282-4a4e-8fb9-e757afa8c022-apiservice-cert\") pod \"packageserver-d55dfcdfc-fc2g9\" (UID: \"10bbf031-b282-4a4e-8fb9-e757afa8c022\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fc2g9" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.645947 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/91dd7b10-99ef-486d-a624-21cb245226c7-encryption-config\") pod \"apiserver-7bbb656c7d-b8zql\" (UID: \"91dd7b10-99ef-486d-a624-21cb245226c7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8zql" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.646176 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ac102463-bf2a-466d-b24a-e3249661d1ee-auth-proxy-config\") pod \"machine-approver-56656f9798-d7vvw\" (UID: \"ac102463-bf2a-466d-b24a-e3249661d1ee\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-d7vvw" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.650211 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b7e0b1cd-5c13-436d-8c35-c0a659f767a0-serving-cert\") pod \"etcd-operator-b45778765-7qxf9\" (UID: \"b7e0b1cd-5c13-436d-8c35-c0a659f767a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7qxf9" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.652516 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/d153ec32-0e4e-41d6-b105-38f91ab40c14-stats-auth\") pod \"router-default-5444994796-kxfxn\" (UID: \"d153ec32-0e4e-41d6-b105-38f91ab40c14\") " pod="openshift-ingress/router-default-5444994796-kxfxn" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.653040 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e23712e1-8705-4d4b-bd1a-1bea23f765df-config\") pod \"authentication-operator-69f744f599-k4hj2\" (UID: \"e23712e1-8705-4d4b-bd1a-1bea23f765df\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-k4hj2" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.653453 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0826463c-2c8a-4b52-8727-4b07de3d86c3-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-ktld7\" (UID: \"0826463c-2c8a-4b52-8727-4b07de3d86c3\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ktld7" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.653759 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97fc0814-92d4-42a4-89c0-acb55bf47b4d-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-hd47j\" (UID: \"97fc0814-92d4-42a4-89c0-acb55bf47b4d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hd47j" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.653807 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5cf1f97f-ff78-4047-b839-6e8ffb6b92b9-trusted-ca\") pod \"ingress-operator-5b745b69d9-n7q4s\" (UID: \"5cf1f97f-ff78-4047-b839-6e8ffb6b92b9\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n7q4s" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.654603 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e2656d73-dfaf-4d40-a1b8-0d08297f9dc4-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-ffl2m\" (UID: \"e2656d73-dfaf-4d40-a1b8-0d08297f9dc4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ffl2m" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.655440 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a0b0911d-7b7c-4746-bcfc-d15069434bb1-proxy-tls\") pod \"machine-config-controller-84d6567774-v2tt6\" (UID: \"a0b0911d-7b7c-4746-bcfc-d15069434bb1\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-v2tt6" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.655530 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e23712e1-8705-4d4b-bd1a-1bea23f765df-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-k4hj2\" (UID: \"e23712e1-8705-4d4b-bd1a-1bea23f765df\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-k4hj2" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.655868 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e09b8f75-2286-4a12-9571-5342b681690f-client-ca\") pod \"route-controller-manager-6576b87f9c-mqvqx\" (UID: \"e09b8f75-2286-4a12-9571-5342b681690f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mqvqx" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.656669 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5863644d-bf1b-4dbf-9c89-99ec5bff14e3-config\") pod \"openshift-apiserver-operator-796bbdcf4f-476wv\" (UID: \"5863644d-bf1b-4dbf-9c89-99ec5bff14e3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-476wv" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.656918 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/91dd7b10-99ef-486d-a624-21cb245226c7-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-b8zql\" (UID: \"91dd7b10-99ef-486d-a624-21cb245226c7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8zql" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.657549 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0826463c-2c8a-4b52-8727-4b07de3d86c3-config\") pod \"kube-controller-manager-operator-78b949d7b-ktld7\" (UID: \"0826463c-2c8a-4b52-8727-4b07de3d86c3\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ktld7" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.658112 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e23712e1-8705-4d4b-bd1a-1bea23f765df-service-ca-bundle\") pod \"authentication-operator-69f744f599-k4hj2\" (UID: \"e23712e1-8705-4d4b-bd1a-1bea23f765df\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-k4hj2" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.658510 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/91dd7b10-99ef-486d-a624-21cb245226c7-audit-dir\") pod \"apiserver-7bbb656c7d-b8zql\" (UID: \"91dd7b10-99ef-486d-a624-21cb245226c7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8zql" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.659001 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5863644d-bf1b-4dbf-9c89-99ec5bff14e3-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-476wv\" (UID: \"5863644d-bf1b-4dbf-9c89-99ec5bff14e3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-476wv" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.659140 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/a3f62807-e422-4d74-82c5-81756d963b77-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-zs6w7\" (UID: \"a3f62807-e422-4d74-82c5-81756d963b77\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zs6w7" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.659483 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/96800ba3-0ee4-42e0-b8f7-316ef57a6173-installation-pull-secrets\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.659973 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82ece3ff-72cb-48b3-92ad-251d7bf980c0-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-pgn88\" (UID: \"82ece3ff-72cb-48b3-92ad-251d7bf980c0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pgn88" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.661703 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/10bbf031-b282-4a4e-8fb9-e757afa8c022-webhook-cert\") pod \"packageserver-d55dfcdfc-fc2g9\" (UID: \"10bbf031-b282-4a4e-8fb9-e757afa8c022\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fc2g9" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.662567 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e23712e1-8705-4d4b-bd1a-1bea23f765df-serving-cert\") pod \"authentication-operator-69f744f599-k4hj2\" (UID: \"e23712e1-8705-4d4b-bd1a-1bea23f765df\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-k4hj2" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.663143 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e09b8f75-2286-4a12-9571-5342b681690f-serving-cert\") pod \"route-controller-manager-6576b87f9c-mqvqx\" (UID: \"e09b8f75-2286-4a12-9571-5342b681690f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mqvqx" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.663539 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n56hr\" (UniqueName: \"kubernetes.io/projected/5863644d-bf1b-4dbf-9c89-99ec5bff14e3-kube-api-access-n56hr\") pod \"openshift-apiserver-operator-796bbdcf4f-476wv\" (UID: \"5863644d-bf1b-4dbf-9c89-99ec5bff14e3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-476wv" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.664903 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/62ff675c-ff12-4c14-9c6b-ce3fb25ea214-serving-cert\") pod \"openshift-config-operator-7777fb866f-pshcp\" (UID: \"62ff675c-ff12-4c14-9c6b-ce3fb25ea214\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-pshcp" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.665063 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/d153ec32-0e4e-41d6-b105-38f91ab40c14-default-certificate\") pod \"router-default-5444994796-kxfxn\" (UID: \"d153ec32-0e4e-41d6-b105-38f91ab40c14\") " pod="openshift-ingress/router-default-5444994796-kxfxn" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.665760 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/acb1ddd4-f551-4d06-9021-28d7579bad7f-srv-cert\") pod \"catalog-operator-68c6474976-c7vh9\" (UID: \"acb1ddd4-f551-4d06-9021-28d7579bad7f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-c7vh9" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.671601 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/79972ac1-e98f-4cd9-ba62-dd5d3b9e81e6-metrics-tls\") pod \"dns-default-ms5f2\" (UID: \"79972ac1-e98f-4cd9-ba62-dd5d3b9e81e6\") " pod="openshift-dns/dns-default-ms5f2" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.672122 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82ece3ff-72cb-48b3-92ad-251d7bf980c0-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-pgn88\" (UID: \"82ece3ff-72cb-48b3-92ad-251d7bf980c0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pgn88" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.672425 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/96800ba3-0ee4-42e0-b8f7-316ef57a6173-registry-tls\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.672940 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d153ec32-0e4e-41d6-b105-38f91ab40c14-metrics-certs\") pod \"router-default-5444994796-kxfxn\" (UID: \"d153ec32-0e4e-41d6-b105-38f91ab40c14\") " pod="openshift-ingress/router-default-5444994796-kxfxn" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.683237 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/91dd7b10-99ef-486d-a624-21cb245226c7-etcd-client\") pod \"apiserver-7bbb656c7d-b8zql\" (UID: \"91dd7b10-99ef-486d-a624-21cb245226c7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8zql" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.684658 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/34cf8610-cc4e-46de-a00c-ff41d7030e1f-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-h7rbb\" (UID: \"34cf8610-cc4e-46de-a00c-ff41d7030e1f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h7rbb" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.693973 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gp9h8\" (UniqueName: \"kubernetes.io/projected/a0b0911d-7b7c-4746-bcfc-d15069434bb1-kube-api-access-gp9h8\") pod \"machine-config-controller-84d6567774-v2tt6\" (UID: \"a0b0911d-7b7c-4746-bcfc-d15069434bb1\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-v2tt6" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.703462 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/82ece3ff-72cb-48b3-92ad-251d7bf980c0-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-pgn88\" (UID: \"82ece3ff-72cb-48b3-92ad-251d7bf980c0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pgn88" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.704802 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-8bln9"] Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.729684 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8ef9a0ab-2131-42ff-823d-17d517929e77-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xmhq9\" (UID: \"8ef9a0ab-2131-42ff-823d-17d517929e77\") " pod="openshift-marketplace/marketplace-operator-79b997595-xmhq9" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.729731 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/ba947c9c-cd84-4c9b-a431-9276412058d1-mountpoint-dir\") pod \"csi-hostpathplugin-cmpbm\" (UID: \"ba947c9c-cd84-4c9b-a431-9276412058d1\") " pod="hostpath-provisioner/csi-hostpathplugin-cmpbm" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.729767 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/e682b6af-7f62-4efb-917e-7453e78138bc-signing-cabundle\") pod \"service-ca-9c57cc56f-klnk6\" (UID: \"e682b6af-7f62-4efb-917e-7453e78138bc\") " pod="openshift-service-ca/service-ca-9c57cc56f-klnk6" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.729788 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ba947c9c-cd84-4c9b-a431-9276412058d1-registration-dir\") pod \"csi-hostpathplugin-cmpbm\" (UID: \"ba947c9c-cd84-4c9b-a431-9276412058d1\") " pod="hostpath-provisioner/csi-hostpathplugin-cmpbm" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.729810 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/e682b6af-7f62-4efb-917e-7453e78138bc-signing-key\") pod \"service-ca-9c57cc56f-klnk6\" (UID: \"e682b6af-7f62-4efb-917e-7453e78138bc\") " pod="openshift-service-ca/service-ca-9c57cc56f-klnk6" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.729852 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57qks\" (UniqueName: \"kubernetes.io/projected/e682b6af-7f62-4efb-917e-7453e78138bc-kube-api-access-57qks\") pod \"service-ca-9c57cc56f-klnk6\" (UID: \"e682b6af-7f62-4efb-917e-7453e78138bc\") " pod="openshift-service-ca/service-ca-9c57cc56f-klnk6" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.729881 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e3475b27-002a-4087-ab65-9cdbad01ca92-profile-collector-cert\") pod \"olm-operator-6b444d44fb-nfp7k\" (UID: \"e3475b27-002a-4087-ab65-9cdbad01ca92\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nfp7k" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.729905 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4622b25d-fb03-4e9f-a66c-7f90def9b5f7-config-volume\") pod \"collect-profiles-29401380-96zpb\" (UID: \"4622b25d-fb03-4e9f-a66c-7f90def9b5f7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401380-96zpb" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.729941 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9hld\" (UniqueName: \"kubernetes.io/projected/ba947c9c-cd84-4c9b-a431-9276412058d1-kube-api-access-g9hld\") pod \"csi-hostpathplugin-cmpbm\" (UID: \"ba947c9c-cd84-4c9b-a431-9276412058d1\") " pod="hostpath-provisioner/csi-hostpathplugin-cmpbm" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.729989 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.730024 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/815dd5c7-dda6-42c5-b63f-8e354c5d1143-cert\") pod \"ingress-canary-rx9dj\" (UID: \"815dd5c7-dda6-42c5-b63f-8e354c5d1143\") " pod="openshift-ingress-canary/ingress-canary-rx9dj" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.730044 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/300178a7-3aa1-4fbb-a604-a37365cc968f-serving-cert\") pod \"service-ca-operator-777779d784-99n87\" (UID: \"300178a7-3aa1-4fbb-a604-a37365cc968f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-99n87" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.730065 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ba947c9c-cd84-4c9b-a431-9276412058d1-socket-dir\") pod \"csi-hostpathplugin-cmpbm\" (UID: \"ba947c9c-cd84-4c9b-a431-9276412058d1\") " pod="hostpath-provisioner/csi-hostpathplugin-cmpbm" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.730119 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e3475b27-002a-4087-ab65-9cdbad01ca92-srv-cert\") pod \"olm-operator-6b444d44fb-nfp7k\" (UID: \"e3475b27-002a-4087-ab65-9cdbad01ca92\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nfp7k" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.730144 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d5dd7cd5-c63e-460f-8d90-d91f5ec85f15-proxy-tls\") pod \"machine-config-operator-74547568cd-jkwk5\" (UID: \"d5dd7cd5-c63e-460f-8d90-d91f5ec85f15\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jkwk5" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.730214 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtx7v\" (UniqueName: \"kubernetes.io/projected/e3475b27-002a-4087-ab65-9cdbad01ca92-kube-api-access-qtx7v\") pod \"olm-operator-6b444d44fb-nfp7k\" (UID: \"e3475b27-002a-4087-ab65-9cdbad01ca92\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nfp7k" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.730253 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d5dd7cd5-c63e-460f-8d90-d91f5ec85f15-images\") pod \"machine-config-operator-74547568cd-jkwk5\" (UID: \"d5dd7cd5-c63e-460f-8d90-d91f5ec85f15\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jkwk5" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.730279 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/300178a7-3aa1-4fbb-a604-a37365cc968f-config\") pod \"service-ca-operator-777779d784-99n87\" (UID: \"300178a7-3aa1-4fbb-a604-a37365cc968f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-99n87" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.730317 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d5dd7cd5-c63e-460f-8d90-d91f5ec85f15-auth-proxy-config\") pod \"machine-config-operator-74547568cd-jkwk5\" (UID: \"d5dd7cd5-c63e-460f-8d90-d91f5ec85f15\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jkwk5" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.730372 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/ba947c9c-cd84-4c9b-a431-9276412058d1-plugins-dir\") pod \"csi-hostpathplugin-cmpbm\" (UID: \"ba947c9c-cd84-4c9b-a431-9276412058d1\") " pod="hostpath-provisioner/csi-hostpathplugin-cmpbm" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.730417 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpgtd\" (UniqueName: \"kubernetes.io/projected/300178a7-3aa1-4fbb-a604-a37365cc968f-kube-api-access-fpgtd\") pod \"service-ca-operator-777779d784-99n87\" (UID: \"300178a7-3aa1-4fbb-a604-a37365cc968f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-99n87" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.730464 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbx85\" (UniqueName: \"kubernetes.io/projected/429acdcb-8b70-482c-abad-82f4867a9266-kube-api-access-rbx85\") pod \"machine-config-server-sm46z\" (UID: \"429acdcb-8b70-482c-abad-82f4867a9266\") " pod="openshift-machine-config-operator/machine-config-server-sm46z" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.730502 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/429acdcb-8b70-482c-abad-82f4867a9266-certs\") pod \"machine-config-server-sm46z\" (UID: \"429acdcb-8b70-482c-abad-82f4867a9266\") " pod="openshift-machine-config-operator/machine-config-server-sm46z" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.730866 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/ba947c9c-cd84-4c9b-a431-9276412058d1-mountpoint-dir\") pod \"csi-hostpathplugin-cmpbm\" (UID: \"ba947c9c-cd84-4c9b-a431-9276412058d1\") " pod="hostpath-provisioner/csi-hostpathplugin-cmpbm" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.731579 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d5dd7cd5-c63e-460f-8d90-d91f5ec85f15-images\") pod \"machine-config-operator-74547568cd-jkwk5\" (UID: \"d5dd7cd5-c63e-460f-8d90-d91f5ec85f15\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jkwk5" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.734327 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/815dd5c7-dda6-42c5-b63f-8e354c5d1143-cert\") pod \"ingress-canary-rx9dj\" (UID: \"815dd5c7-dda6-42c5-b63f-8e354c5d1143\") " pod="openshift-ingress-canary/ingress-canary-rx9dj" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.734526 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/ba947c9c-cd84-4c9b-a431-9276412058d1-plugins-dir\") pod \"csi-hostpathplugin-cmpbm\" (UID: \"ba947c9c-cd84-4c9b-a431-9276412058d1\") " pod="hostpath-provisioner/csi-hostpathplugin-cmpbm" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.735321 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/300178a7-3aa1-4fbb-a604-a37365cc968f-config\") pod \"service-ca-operator-777779d784-99n87\" (UID: \"300178a7-3aa1-4fbb-a604-a37365cc968f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-99n87" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.736008 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d5dd7cd5-c63e-460f-8d90-d91f5ec85f15-auth-proxy-config\") pod \"machine-config-operator-74547568cd-jkwk5\" (UID: \"d5dd7cd5-c63e-460f-8d90-d91f5ec85f15\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jkwk5" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.736063 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ba947c9c-cd84-4c9b-a431-9276412058d1-registration-dir\") pod \"csi-hostpathplugin-cmpbm\" (UID: \"ba947c9c-cd84-4c9b-a431-9276412058d1\") " pod="hostpath-provisioner/csi-hostpathplugin-cmpbm" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.736264 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ba947c9c-cd84-4c9b-a431-9276412058d1-socket-dir\") pod \"csi-hostpathplugin-cmpbm\" (UID: \"ba947c9c-cd84-4c9b-a431-9276412058d1\") " pod="hostpath-provisioner/csi-hostpathplugin-cmpbm" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.736715 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/300178a7-3aa1-4fbb-a604-a37365cc968f-serving-cert\") pod \"service-ca-operator-777779d784-99n87\" (UID: \"300178a7-3aa1-4fbb-a604-a37365cc968f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-99n87" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.737561 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d5dd7cd5-c63e-460f-8d90-d91f5ec85f15-proxy-tls\") pod \"machine-config-operator-74547568cd-jkwk5\" (UID: \"d5dd7cd5-c63e-460f-8d90-d91f5ec85f15\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jkwk5" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.739021 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhs55\" (UniqueName: \"kubernetes.io/projected/e09b8f75-2286-4a12-9571-5342b681690f-kube-api-access-lhs55\") pod \"route-controller-manager-6576b87f9c-mqvqx\" (UID: \"e09b8f75-2286-4a12-9571-5342b681690f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mqvqx" Nov 25 15:03:53 crc kubenswrapper[4890]: E1125 15:03:53.739845 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:03:54.239799448 +0000 UTC m=+92.682262268 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.742097 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/ba947c9c-cd84-4c9b-a431-9276412058d1-csi-data-dir\") pod \"csi-hostpathplugin-cmpbm\" (UID: \"ba947c9c-cd84-4c9b-a431-9276412058d1\") " pod="hostpath-provisioner/csi-hostpathplugin-cmpbm" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.742178 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8ef9a0ab-2131-42ff-823d-17d517929e77-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xmhq9\" (UID: \"8ef9a0ab-2131-42ff-823d-17d517929e77\") " pod="openshift-marketplace/marketplace-operator-79b997595-xmhq9" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.740385 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwrgg\" (UniqueName: \"kubernetes.io/projected/cb6875f9-0bd5-4076-848f-fd734d4f04df-kube-api-access-gwrgg\") pod \"package-server-manager-789f6589d5-z5tfd\" (UID: \"cb6875f9-0bd5-4076-848f-fd734d4f04df\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-z5tfd" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.742214 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4622b25d-fb03-4e9f-a66c-7f90def9b5f7-secret-volume\") pod \"collect-profiles-29401380-96zpb\" (UID: \"4622b25d-fb03-4e9f-a66c-7f90def9b5f7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401380-96zpb" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.742254 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjjcv\" (UniqueName: \"kubernetes.io/projected/4622b25d-fb03-4e9f-a66c-7f90def9b5f7-kube-api-access-qjjcv\") pod \"collect-profiles-29401380-96zpb\" (UID: \"4622b25d-fb03-4e9f-a66c-7f90def9b5f7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401380-96zpb" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.742286 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xn8v6\" (UniqueName: \"kubernetes.io/projected/815dd5c7-dda6-42c5-b63f-8e354c5d1143-kube-api-access-xn8v6\") pod \"ingress-canary-rx9dj\" (UID: \"815dd5c7-dda6-42c5-b63f-8e354c5d1143\") " pod="openshift-ingress-canary/ingress-canary-rx9dj" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.742303 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/429acdcb-8b70-482c-abad-82f4867a9266-node-bootstrap-token\") pod \"machine-config-server-sm46z\" (UID: \"429acdcb-8b70-482c-abad-82f4867a9266\") " pod="openshift-machine-config-operator/machine-config-server-sm46z" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.742324 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vj4n4\" (UniqueName: \"kubernetes.io/projected/d5dd7cd5-c63e-460f-8d90-d91f5ec85f15-kube-api-access-vj4n4\") pod \"machine-config-operator-74547568cd-jkwk5\" (UID: \"d5dd7cd5-c63e-460f-8d90-d91f5ec85f15\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jkwk5" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.742351 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvnfh\" (UniqueName: \"kubernetes.io/projected/8ef9a0ab-2131-42ff-823d-17d517929e77-kube-api-access-wvnfh\") pod \"marketplace-operator-79b997595-xmhq9\" (UID: \"8ef9a0ab-2131-42ff-823d-17d517929e77\") " pod="openshift-marketplace/marketplace-operator-79b997595-xmhq9" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.742438 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4622b25d-fb03-4e9f-a66c-7f90def9b5f7-config-volume\") pod \"collect-profiles-29401380-96zpb\" (UID: \"4622b25d-fb03-4e9f-a66c-7f90def9b5f7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401380-96zpb" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.741923 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/e682b6af-7f62-4efb-917e-7453e78138bc-signing-key\") pod \"service-ca-9c57cc56f-klnk6\" (UID: \"e682b6af-7f62-4efb-917e-7453e78138bc\") " pod="openshift-service-ca/service-ca-9c57cc56f-klnk6" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.741282 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/e682b6af-7f62-4efb-917e-7453e78138bc-signing-cabundle\") pod \"service-ca-9c57cc56f-klnk6\" (UID: \"e682b6af-7f62-4efb-917e-7453e78138bc\") " pod="openshift-service-ca/service-ca-9c57cc56f-klnk6" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.742622 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/ba947c9c-cd84-4c9b-a431-9276412058d1-csi-data-dir\") pod \"csi-hostpathplugin-cmpbm\" (UID: \"ba947c9c-cd84-4c9b-a431-9276412058d1\") " pod="hostpath-provisioner/csi-hostpathplugin-cmpbm" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.746732 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8ef9a0ab-2131-42ff-823d-17d517929e77-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xmhq9\" (UID: \"8ef9a0ab-2131-42ff-823d-17d517929e77\") " pod="openshift-marketplace/marketplace-operator-79b997595-xmhq9" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.748989 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4622b25d-fb03-4e9f-a66c-7f90def9b5f7-secret-volume\") pod \"collect-profiles-29401380-96zpb\" (UID: \"4622b25d-fb03-4e9f-a66c-7f90def9b5f7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401380-96zpb" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.752040 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/429acdcb-8b70-482c-abad-82f4867a9266-certs\") pod \"machine-config-server-sm46z\" (UID: \"429acdcb-8b70-482c-abad-82f4867a9266\") " pod="openshift-machine-config-operator/machine-config-server-sm46z" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.753442 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/429acdcb-8b70-482c-abad-82f4867a9266-node-bootstrap-token\") pod \"machine-config-server-sm46z\" (UID: \"429acdcb-8b70-482c-abad-82f4867a9266\") " pod="openshift-machine-config-operator/machine-config-server-sm46z" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.754591 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8ef9a0ab-2131-42ff-823d-17d517929e77-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xmhq9\" (UID: \"8ef9a0ab-2131-42ff-823d-17d517929e77\") " pod="openshift-marketplace/marketplace-operator-79b997595-xmhq9" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.755106 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e3475b27-002a-4087-ab65-9cdbad01ca92-profile-collector-cert\") pod \"olm-operator-6b444d44fb-nfp7k\" (UID: \"e3475b27-002a-4087-ab65-9cdbad01ca92\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nfp7k" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.755107 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e3475b27-002a-4087-ab65-9cdbad01ca92-srv-cert\") pod \"olm-operator-6b444d44fb-nfp7k\" (UID: \"e3475b27-002a-4087-ab65-9cdbad01ca92\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nfp7k" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.766072 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6vvk\" (UniqueName: \"kubernetes.io/projected/ac102463-bf2a-466d-b24a-e3249661d1ee-kube-api-access-q6vvk\") pod \"machine-approver-56656f9798-d7vvw\" (UID: \"ac102463-bf2a-466d-b24a-e3249661d1ee\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-d7vvw" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.766509 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-9l4z5"] Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.773030 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mmhbh"] Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.779682 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pgn88" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.787069 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wq4rw\" (UniqueName: \"kubernetes.io/projected/21edd566-f5c3-425d-b719-d68d440884d8-kube-api-access-wq4rw\") pod \"dns-operator-744455d44c-46rkq\" (UID: \"21edd566-f5c3-425d-b719-d68d440884d8\") " pod="openshift-dns-operator/dns-operator-744455d44c-46rkq" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.792209 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-t29fd"] Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.798330 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxr96\" (UniqueName: \"kubernetes.io/projected/91dd7b10-99ef-486d-a624-21cb245226c7-kube-api-access-xxr96\") pod \"apiserver-7bbb656c7d-b8zql\" (UID: \"91dd7b10-99ef-486d-a624-21cb245226c7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8zql" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.833480 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-v2tt6" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.834475 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5ssx\" (UniqueName: \"kubernetes.io/projected/96800ba3-0ee4-42e0-b8f7-316ef57a6173-kube-api-access-p5ssx\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.843342 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-6dmh8"] Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.843627 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:03:53 crc kubenswrapper[4890]: E1125 15:03:53.844181 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:03:54.344115152 +0000 UTC m=+92.786577762 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.844323 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:53 crc kubenswrapper[4890]: E1125 15:03:53.845129 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:03:54.345097089 +0000 UTC m=+92.787559699 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.848336 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-glmvm"] Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.859250 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lc8sc\" (UniqueName: \"kubernetes.io/projected/a3f62807-e422-4d74-82c5-81756d963b77-kube-api-access-lc8sc\") pod \"control-plane-machine-set-operator-78cbb6b69f-zs6w7\" (UID: \"a3f62807-e422-4d74-82c5-81756d963b77\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zs6w7" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.867923 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmx77\" (UniqueName: \"kubernetes.io/projected/b7e0b1cd-5c13-436d-8c35-c0a659f767a0-kube-api-access-hmx77\") pod \"etcd-operator-b45778765-7qxf9\" (UID: \"b7e0b1cd-5c13-436d-8c35-c0a659f767a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7qxf9" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.871341 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-z5tfd" Nov 25 15:03:53 crc kubenswrapper[4890]: W1125 15:03:53.875601 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda6795dc3_313a_4599_a07c_588365bd8965.slice/crio-96ba5c04392462f532d7f92304b9e7d366cc3501a1fd34e9e3d92b94dbb14ba6 WatchSource:0}: Error finding container 96ba5c04392462f532d7f92304b9e7d366cc3501a1fd34e9e3d92b94dbb14ba6: Status 404 returned error can't find the container with id 96ba5c04392462f532d7f92304b9e7d366cc3501a1fd34e9e3d92b94dbb14ba6 Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.878727 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nhq4p"] Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.880452 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-m95j6"] Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.882096 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xldhl"] Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.888851 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sz4xc\" (UniqueName: \"kubernetes.io/projected/79972ac1-e98f-4cd9-ba62-dd5d3b9e81e6-kube-api-access-sz4xc\") pod \"dns-default-ms5f2\" (UID: \"79972ac1-e98f-4cd9-ba62-dd5d3b9e81e6\") " pod="openshift-dns/dns-default-ms5f2" Nov 25 15:03:53 crc kubenswrapper[4890]: W1125 15:03:53.895820 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8b5c8177_1153_453a_9110_6023d01e3fb7.slice/crio-6c423f870d9a99cf21940ddce7d039c74691e5d8de00b83f8627b02fc70264df WatchSource:0}: Error finding container 6c423f870d9a99cf21940ddce7d039c74691e5d8de00b83f8627b02fc70264df: Status 404 returned error can't find the container with id 6c423f870d9a99cf21940ddce7d039c74691e5d8de00b83f8627b02fc70264df Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.900876 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67bjm\" (UniqueName: \"kubernetes.io/projected/97fc0814-92d4-42a4-89c0-acb55bf47b4d-kube-api-access-67bjm\") pod \"kube-storage-version-migrator-operator-b67b599dd-hd47j\" (UID: \"97fc0814-92d4-42a4-89c0-acb55bf47b4d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hd47j" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.919920 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5cf1f97f-ff78-4047-b839-6e8ffb6b92b9-bound-sa-token\") pod \"ingress-operator-5b745b69d9-n7q4s\" (UID: \"5cf1f97f-ff78-4047-b839-6e8ffb6b92b9\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n7q4s" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.923362 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-476wv" Nov 25 15:03:53 crc kubenswrapper[4890]: W1125 15:03:53.924916 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6e313ec3_70c2_4220_b08a_ef79b8a25847.slice/crio-1129dc184de9e2277418c632ff648e55150aaf4685472496dc6ab548f48fdf03 WatchSource:0}: Error finding container 1129dc184de9e2277418c632ff648e55150aaf4685472496dc6ab548f48fdf03: Status 404 returned error can't find the container with id 1129dc184de9e2277418c632ff648e55150aaf4685472496dc6ab548f48fdf03 Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.945597 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:03:53 crc kubenswrapper[4890]: E1125 15:03:53.946288 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:03:54.446264727 +0000 UTC m=+92.888727337 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.949317 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:53 crc kubenswrapper[4890]: E1125 15:03:53.950800 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:03:54.450586085 +0000 UTC m=+92.893048695 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.951461 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/96800ba3-0ee4-42e0-b8f7-316ef57a6173-bound-sa-token\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.955339 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-d7vvw" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.960863 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrft7\" (UniqueName: \"kubernetes.io/projected/e23712e1-8705-4d4b-bd1a-1bea23f765df-kube-api-access-xrft7\") pod \"authentication-operator-69f744f599-k4hj2\" (UID: \"e23712e1-8705-4d4b-bd1a-1bea23f765df\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-k4hj2" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.977412 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mqvqx" Nov 25 15:03:53 crc kubenswrapper[4890]: I1125 15:03:53.983088 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4mk6\" (UniqueName: \"kubernetes.io/projected/d153ec32-0e4e-41d6-b105-38f91ab40c14-kube-api-access-n4mk6\") pod \"router-default-5444994796-kxfxn\" (UID: \"d153ec32-0e4e-41d6-b105-38f91ab40c14\") " pod="openshift-ingress/router-default-5444994796-kxfxn" Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.004720 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgh2l\" (UniqueName: \"kubernetes.io/projected/62ff675c-ff12-4c14-9c6b-ce3fb25ea214-kube-api-access-jgh2l\") pod \"openshift-config-operator-7777fb866f-pshcp\" (UID: \"62ff675c-ff12-4c14-9c6b-ce3fb25ea214\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-pshcp" Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.033549 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-46rkq" Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.044033 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7gsd\" (UniqueName: \"kubernetes.io/projected/e2656d73-dfaf-4d40-a1b8-0d08297f9dc4-kube-api-access-z7gsd\") pod \"multus-admission-controller-857f4d67dd-ffl2m\" (UID: \"e2656d73-dfaf-4d40-a1b8-0d08297f9dc4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ffl2m" Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.048882 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8zql" Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.051959 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:03:54 crc kubenswrapper[4890]: E1125 15:03:54.052134 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:03:54.552103653 +0000 UTC m=+92.994566263 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.052289 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:54 crc kubenswrapper[4890]: E1125 15:03:54.052866 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:03:54.552857373 +0000 UTC m=+92.995319973 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.062724 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-7qxf9" Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.067628 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/34cf8610-cc4e-46de-a00c-ff41d7030e1f-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-h7rbb\" (UID: \"34cf8610-cc4e-46de-a00c-ff41d7030e1f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h7rbb" Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.073768 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h7rbb" Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.077795 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pgn88"] Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.081950 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvcwm\" (UniqueName: \"kubernetes.io/projected/acb1ddd4-f551-4d06-9021-28d7579bad7f-kube-api-access-fvcwm\") pod \"catalog-operator-68c6474976-c7vh9\" (UID: \"acb1ddd4-f551-4d06-9021-28d7579bad7f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-c7vh9" Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.097855 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hd47j" Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.101357 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kk9j6\" (UniqueName: \"kubernetes.io/projected/5cf1f97f-ff78-4047-b839-6e8ffb6b92b9-kube-api-access-kk9j6\") pod \"ingress-operator-5b745b69d9-n7q4s\" (UID: \"5cf1f97f-ff78-4047-b839-6e8ffb6b92b9\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n7q4s" Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.107609 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-ms5f2" Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.113213 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-kxfxn" Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.124634 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-ffl2m" Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.136173 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2w99\" (UniqueName: \"kubernetes.io/projected/10bbf031-b282-4a4e-8fb9-e757afa8c022-kube-api-access-q2w99\") pod \"packageserver-d55dfcdfc-fc2g9\" (UID: \"10bbf031-b282-4a4e-8fb9-e757afa8c022\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fc2g9" Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.138848 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0826463c-2c8a-4b52-8727-4b07de3d86c3-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-ktld7\" (UID: \"0826463c-2c8a-4b52-8727-4b07de3d86c3\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ktld7" Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.139721 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-c7vh9" Nov 25 15:03:54 crc kubenswrapper[4890]: W1125 15:03:54.144366 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod82ece3ff_72cb_48b3_92ad_251d7bf980c0.slice/crio-331d4f3b911bd3e64d1738f9f3f5aa4f9eeea8e42926e74cfeb7f2483425cdc9 WatchSource:0}: Error finding container 331d4f3b911bd3e64d1738f9f3f5aa4f9eeea8e42926e74cfeb7f2483425cdc9: Status 404 returned error can't find the container with id 331d4f3b911bd3e64d1738f9f3f5aa4f9eeea8e42926e74cfeb7f2483425cdc9 Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.149487 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zs6w7" Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.153475 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:03:54 crc kubenswrapper[4890]: E1125 15:03:54.153896 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:03:54.653856137 +0000 UTC m=+93.096318757 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.154175 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:54 crc kubenswrapper[4890]: E1125 15:03:54.154548 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:03:54.654537506 +0000 UTC m=+93.097000306 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.156429 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ktld7" Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.162654 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fc2g9" Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.167296 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvshl\" (UniqueName: \"kubernetes.io/projected/1a324db0-9f3e-417e-931f-5b74a44775ed-kube-api-access-zvshl\") pod \"migrator-59844c95c7-t5v4f\" (UID: \"1a324db0-9f3e-417e-931f-5b74a44775ed\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t5v4f" Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.183916 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t5v4f" Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.184397 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57qks\" (UniqueName: \"kubernetes.io/projected/e682b6af-7f62-4efb-917e-7453e78138bc-kube-api-access-57qks\") pod \"service-ca-9c57cc56f-klnk6\" (UID: \"e682b6af-7f62-4efb-917e-7453e78138bc\") " pod="openshift-service-ca/service-ca-9c57cc56f-klnk6" Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.191213 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-k4hj2" Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.200048 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtx7v\" (UniqueName: \"kubernetes.io/projected/e3475b27-002a-4087-ab65-9cdbad01ca92-kube-api-access-qtx7v\") pod \"olm-operator-6b444d44fb-nfp7k\" (UID: \"e3475b27-002a-4087-ab65-9cdbad01ca92\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nfp7k" Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.227442 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpgtd\" (UniqueName: \"kubernetes.io/projected/300178a7-3aa1-4fbb-a604-a37365cc968f-kube-api-access-fpgtd\") pod \"service-ca-operator-777779d784-99n87\" (UID: \"300178a7-3aa1-4fbb-a604-a37365cc968f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-99n87" Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.233566 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-pshcp" Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.236826 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-klnk6" Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.243910 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbx85\" (UniqueName: \"kubernetes.io/projected/429acdcb-8b70-482c-abad-82f4867a9266-kube-api-access-rbx85\") pod \"machine-config-server-sm46z\" (UID: \"429acdcb-8b70-482c-abad-82f4867a9266\") " pod="openshift-machine-config-operator/machine-config-server-sm46z" Nov 25 15:03:54 crc kubenswrapper[4890]: W1125 15:03:54.249745 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac102463_bf2a_466d_b24a_e3249661d1ee.slice/crio-c7aa9e6abe148063e41e46d95ff62a5dc0b295b22d3439cab211480c9a6e8980 WatchSource:0}: Error finding container c7aa9e6abe148063e41e46d95ff62a5dc0b295b22d3439cab211480c9a6e8980: Status 404 returned error can't find the container with id c7aa9e6abe148063e41e46d95ff62a5dc0b295b22d3439cab211480c9a6e8980 Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.261978 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:03:54 crc kubenswrapper[4890]: E1125 15:03:54.263102 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:03:54.763087145 +0000 UTC m=+93.205549755 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.278983 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-v2tt6"] Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.282631 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9hld\" (UniqueName: \"kubernetes.io/projected/ba947c9c-cd84-4c9b-a431-9276412058d1-kube-api-access-g9hld\") pod \"csi-hostpathplugin-cmpbm\" (UID: \"ba947c9c-cd84-4c9b-a431-9276412058d1\") " pod="hostpath-provisioner/csi-hostpathplugin-cmpbm" Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.284616 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvnfh\" (UniqueName: \"kubernetes.io/projected/8ef9a0ab-2131-42ff-823d-17d517929e77-kube-api-access-wvnfh\") pod \"marketplace-operator-79b997595-xmhq9\" (UID: \"8ef9a0ab-2131-42ff-823d-17d517929e77\") " pod="openshift-marketplace/marketplace-operator-79b997595-xmhq9" Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.299878 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xn8v6\" (UniqueName: \"kubernetes.io/projected/815dd5c7-dda6-42c5-b63f-8e354c5d1143-kube-api-access-xn8v6\") pod \"ingress-canary-rx9dj\" (UID: \"815dd5c7-dda6-42c5-b63f-8e354c5d1143\") " pod="openshift-ingress-canary/ingress-canary-rx9dj" Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.329259 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vj4n4\" (UniqueName: \"kubernetes.io/projected/d5dd7cd5-c63e-460f-8d90-d91f5ec85f15-kube-api-access-vj4n4\") pod \"machine-config-operator-74547568cd-jkwk5\" (UID: \"d5dd7cd5-c63e-460f-8d90-d91f5ec85f15\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jkwk5" Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.342537 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-sm46z" Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.354509 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjjcv\" (UniqueName: \"kubernetes.io/projected/4622b25d-fb03-4e9f-a66c-7f90def9b5f7-kube-api-access-qjjcv\") pod \"collect-profiles-29401380-96zpb\" (UID: \"4622b25d-fb03-4e9f-a66c-7f90def9b5f7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401380-96zpb" Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.354896 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n7q4s" Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.364188 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:54 crc kubenswrapper[4890]: E1125 15:03:54.364474 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:03:54.864462099 +0000 UTC m=+93.306924709 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.394391 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-z5tfd"] Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.429246 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-6dmh8" event={"ID":"8b5c8177-1153-453a-9110-6023d01e3fb7","Type":"ContainerStarted","Data":"ef38ed031c8521e377cc93711b1425a04a788db3039ebdc9dee1cae71aee1f74"} Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.429297 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-6dmh8" event={"ID":"8b5c8177-1153-453a-9110-6023d01e3fb7","Type":"ContainerStarted","Data":"6c423f870d9a99cf21940ddce7d039c74691e5d8de00b83f8627b02fc70264df"} Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.437754 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xldhl" event={"ID":"6e313ec3-70c2-4220-b08a-ef79b8a25847","Type":"ContainerStarted","Data":"1129dc184de9e2277418c632ff648e55150aaf4685472496dc6ab548f48fdf03"} Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.443983 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-m95j6" event={"ID":"49586ef2-b2a8-4deb-8438-1e23cc0609df","Type":"ContainerStarted","Data":"17f49ec98267da51c2979bcf8a8716a7fdbe299b383c05c882accb689a847410"} Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.449907 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v6ps9" event={"ID":"bc80d112-21a3-4721-9585-7b5c0a41d14d","Type":"ContainerStarted","Data":"ae3d572c5f3e18d4e6bcba2e536ab4c56a0f91a31d745a76f510054043c55f29"} Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.449955 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v6ps9" event={"ID":"bc80d112-21a3-4721-9585-7b5c0a41d14d","Type":"ContainerStarted","Data":"3aa9547ae9a1ee3a5b687aaf93e8074cf479767ee0c5a4bec494725c0aac0766"} Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.452276 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nhq4p" event={"ID":"c32eac2f-1dd7-4de9-97cf-32dc872da26b","Type":"ContainerStarted","Data":"572e467497a376e6ce687c8d828915c534ec1ac1cfcd7062857f2463e157e088"} Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.452991 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nhq4p" event={"ID":"c32eac2f-1dd7-4de9-97cf-32dc872da26b","Type":"ContainerStarted","Data":"d3032b2c201da77dd3edc2b9a7b444aebcc090dce40324af78fbf2e7c33711a0"} Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.454648 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-glmvm" event={"ID":"a6795dc3-313a-4599-a07c-588365bd8965","Type":"ContainerStarted","Data":"96ba5c04392462f532d7f92304b9e7d366cc3501a1fd34e9e3d92b94dbb14ba6"} Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.460568 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-8bln9" event={"ID":"89bebf40-d718-4168-8377-4fa334afaa17","Type":"ContainerStarted","Data":"ac4de552c16246d6d002b5da380964667314829a0a7f6bc3be68a941cbd20f46"} Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.460609 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-8bln9" event={"ID":"89bebf40-d718-4168-8377-4fa334afaa17","Type":"ContainerStarted","Data":"3ae35a6151e476060611f1db417b171d3979bb851134c4f8581037f34e569128"} Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.461428 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-8bln9" Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.464421 4890 patch_prober.go:28] interesting pod/downloads-7954f5f757-8bln9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.464502 4890 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-8bln9" podUID="89bebf40-d718-4168-8377-4fa334afaa17" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.465417 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:03:54 crc kubenswrapper[4890]: E1125 15:03:54.466077 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:03:54.966061659 +0000 UTC m=+93.408524269 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.477811 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pgn88" event={"ID":"82ece3ff-72cb-48b3-92ad-251d7bf980c0","Type":"ContainerStarted","Data":"331d4f3b911bd3e64d1738f9f3f5aa4f9eeea8e42926e74cfeb7f2483425cdc9"} Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.488585 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-9l4z5" event={"ID":"d6ec53ec-533b-49d7-966b-79b8c4f13e5f","Type":"ContainerStarted","Data":"0034580c319efb4fcf7e6278cb40d19528c58d228c11ce847856eddcbc8d7ef1"} Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.488759 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-9l4z5" event={"ID":"d6ec53ec-533b-49d7-966b-79b8c4f13e5f","Type":"ContainerStarted","Data":"4ddb1980e002a7793247760a27c1c7440542da50f6955aeb95d2d83c0628f097"} Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.489476 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-9l4z5" Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.493651 4890 patch_prober.go:28] interesting pod/console-operator-58897d9998-9l4z5 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.25:8443/readyz\": dial tcp 10.217.0.25:8443: connect: connection refused" start-of-body= Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.493718 4890 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-9l4z5" podUID="d6ec53ec-533b-49d7-966b-79b8c4f13e5f" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.25:8443/readyz\": dial tcp 10.217.0.25:8443: connect: connection refused" Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.499980 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nfp7k" Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.503929 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-t29fd" event={"ID":"f748296c-d668-46bd-bcb7-4eb21822a244","Type":"ContainerStarted","Data":"a5a26bf2f348abf2477c73bf6de69eb2f00258d478a7795da8b21e776c8eeae1"} Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.503970 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-t29fd" event={"ID":"f748296c-d668-46bd-bcb7-4eb21822a244","Type":"ContainerStarted","Data":"9d0c93c82198fc751619075a936ae9aa328dc51d85d284cdc40d56221bbf7362"} Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.507135 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xmhq9" Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.512922 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-d7vvw" event={"ID":"ac102463-bf2a-466d-b24a-e3249661d1ee","Type":"ContainerStarted","Data":"c7aa9e6abe148063e41e46d95ff62a5dc0b295b22d3439cab211480c9a6e8980"} Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.513423 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-99n87" Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.525013 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401380-96zpb" Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.526600 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" event={"ID":"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9","Type":"ContainerStarted","Data":"91dd36a079d7a52a961a4d8623a3e71a5f18069a9b6943b9349e886f2069b95f"} Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.527591 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.528959 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jkwk5" Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.529174 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-mqvqx"] Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.539677 4890 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-mmhbh container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.28:6443/healthz\": dial tcp 10.217.0.28:6443: connect: connection refused" start-of-body= Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.539736 4890 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" podUID="a27aab8b-a74f-4a62-b5a3-27d8a7d46be9" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.28:6443/healthz\": dial tcp 10.217.0.28:6443: connect: connection refused" Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.549839 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-rx9dj" Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.567399 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:54 crc kubenswrapper[4890]: E1125 15:03:54.573888 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:03:55.073867709 +0000 UTC m=+93.516330319 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.576214 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-cmpbm" Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.603745 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-476wv"] Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.675947 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:03:54 crc kubenswrapper[4890]: E1125 15:03:54.676140 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:03:55.176111036 +0000 UTC m=+93.618573656 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.677555 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:54 crc kubenswrapper[4890]: E1125 15:03:54.679524 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:03:55.179504939 +0000 UTC m=+93.621967629 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.779878 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:03:54 crc kubenswrapper[4890]: E1125 15:03:54.780235 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:03:55.280219775 +0000 UTC m=+93.722682385 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.881637 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:54 crc kubenswrapper[4890]: E1125 15:03:54.881997 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:03:55.381981389 +0000 UTC m=+93.824443999 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.893646 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hd47j"] Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.893712 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h7rbb"] Nov 25 15:03:54 crc kubenswrapper[4890]: W1125 15:03:54.907570 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod429acdcb_8b70_482c_abad_82f4867a9266.slice/crio-5fc3df88b67f39ced9a90250c8814122f07ab3377957c53ba4acef2653e9223d WatchSource:0}: Error finding container 5fc3df88b67f39ced9a90250c8814122f07ab3377957c53ba4acef2653e9223d: Status 404 returned error can't find the container with id 5fc3df88b67f39ced9a90250c8814122f07ab3377957c53ba4acef2653e9223d Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.947926 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-46rkq"] Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.953021 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zs6w7"] Nov 25 15:03:54 crc kubenswrapper[4890]: I1125 15:03:54.982275 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:03:54 crc kubenswrapper[4890]: E1125 15:03:54.982692 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:03:55.482672974 +0000 UTC m=+93.925135574 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.085840 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:55 crc kubenswrapper[4890]: E1125 15:03:55.086444 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:03:55.586428863 +0000 UTC m=+94.028891473 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.188989 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:03:55 crc kubenswrapper[4890]: E1125 15:03:55.189026 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:03:55.689006909 +0000 UTC m=+94.131469519 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.195334 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:55 crc kubenswrapper[4890]: E1125 15:03:55.195580 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:03:55.695568269 +0000 UTC m=+94.138030879 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.224740 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-c7vh9"] Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.241311 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-7qxf9"] Nov 25 15:03:55 crc kubenswrapper[4890]: W1125 15:03:55.277797 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda3f62807_e422_4d74_82c5_81756d963b77.slice/crio-b96aa6598f8a4d2939a069c798538f5dfee7af33353b3513c9aa74d9490a0219 WatchSource:0}: Error finding container b96aa6598f8a4d2939a069c798538f5dfee7af33353b3513c9aa74d9490a0219: Status 404 returned error can't find the container with id b96aa6598f8a4d2939a069c798538f5dfee7af33353b3513c9aa74d9490a0219 Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.298432 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:03:55 crc kubenswrapper[4890]: E1125 15:03:55.298912 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:03:55.798892616 +0000 UTC m=+94.241355226 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.400139 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:55 crc kubenswrapper[4890]: E1125 15:03:55.401034 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:03:55.90101873 +0000 UTC m=+94.343481340 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.406304 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-b8zql"] Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.501547 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:03:55 crc kubenswrapper[4890]: E1125 15:03:55.501938 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:03:56.001915801 +0000 UTC m=+94.444378411 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.532986 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ktld7"] Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.543207 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-7qxf9" event={"ID":"b7e0b1cd-5c13-436d-8c35-c0a659f767a0","Type":"ContainerStarted","Data":"ef59bb9e017a963aaecaa671a93e96680a68477137489bc49db1d26cfc4af67c"} Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.552029 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hd47j" event={"ID":"97fc0814-92d4-42a4-89c0-acb55bf47b4d","Type":"ContainerStarted","Data":"8375806944edeb90808a45b1d45feb7bfd738b1ba0e8268c6ec5149f9a60d5cd"} Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.579631 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" event={"ID":"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9","Type":"ContainerStarted","Data":"62c26f47dd493084e7a893c76b6c6c564775542abea1bc2a7bd94c95c34e92bf"} Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.585728 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-sm46z" event={"ID":"429acdcb-8b70-482c-abad-82f4867a9266","Type":"ContainerStarted","Data":"5fc3df88b67f39ced9a90250c8814122f07ab3377957c53ba4acef2653e9223d"} Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.599491 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-m95j6" event={"ID":"49586ef2-b2a8-4deb-8438-1e23cc0609df","Type":"ContainerStarted","Data":"afb128cb42d3eead9db569bcc05b96081a98fd61ad4b5129b66c6c81bf0cd2ac"} Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.599543 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-m95j6" event={"ID":"49586ef2-b2a8-4deb-8438-1e23cc0609df","Type":"ContainerStarted","Data":"e256feea7455da41e4c20822d6e772bb410052fa8c1b44acfd7e9306295f5b31"} Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.602431 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:55 crc kubenswrapper[4890]: E1125 15:03:55.602699 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:03:56.102688908 +0000 UTC m=+94.545151518 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.613077 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-476wv" event={"ID":"5863644d-bf1b-4dbf-9c89-99ec5bff14e3","Type":"ContainerStarted","Data":"0cf4ee0837f5abd81491fb82edc99c69541677ff09e0b8f4d6075f690ef2d0b0"} Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.613119 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-476wv" event={"ID":"5863644d-bf1b-4dbf-9c89-99ec5bff14e3","Type":"ContainerStarted","Data":"8a0d887895ada168eadd5d42e2ecbde9ad01ef0352326887ad6b402d0922cdc9"} Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.618675 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xldhl" event={"ID":"6e313ec3-70c2-4220-b08a-ef79b8a25847","Type":"ContainerStarted","Data":"9fbede642703b66da34b00feaaa8a19598b017d4abe7de06ff6b12a64a93636f"} Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.619003 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-xldhl" Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.620102 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h7rbb" event={"ID":"34cf8610-cc4e-46de-a00c-ff41d7030e1f","Type":"ContainerStarted","Data":"78789a8728d6c1a45fa4a7917028657de844e637252e8ab7f09dad15027cc547"} Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.621792 4890 generic.go:334] "Generic (PLEG): container finished" podID="a6795dc3-313a-4599-a07c-588365bd8965" containerID="a3a2a345a7d3baf940a0b16764702c4c32e1cc2a0355fa3b0a6761ae450434d7" exitCode=0 Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.622006 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-glmvm" event={"ID":"a6795dc3-313a-4599-a07c-588365bd8965","Type":"ContainerDied","Data":"a3a2a345a7d3baf940a0b16764702c4c32e1cc2a0355fa3b0a6761ae450434d7"} Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.622652 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-c7vh9" event={"ID":"acb1ddd4-f551-4d06-9021-28d7579bad7f","Type":"ContainerStarted","Data":"afd85b355b2da40abab4bd4db2a3d1de0fe5b678ee756f7ae19d17bf01767123"} Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.623493 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mqvqx" event={"ID":"e09b8f75-2286-4a12-9571-5342b681690f","Type":"ContainerStarted","Data":"54c69ecc23762a33f8d6eeca868dd59ac28361f509178948acf91cd5a6241e65"} Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.623521 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mqvqx" event={"ID":"e09b8f75-2286-4a12-9571-5342b681690f","Type":"ContainerStarted","Data":"7503f9b1edca457c0b527a0174f9a5cc859a63fdb99e56b388c0337d94c57ab8"} Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.627214 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mqvqx" Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.632301 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-v2tt6" event={"ID":"a0b0911d-7b7c-4746-bcfc-d15069434bb1","Type":"ContainerStarted","Data":"bcbc9085450a20bb6e191d08e481792b723396ff1e8daa4a981e18bbd0377342"} Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.637988 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zs6w7" event={"ID":"a3f62807-e422-4d74-82c5-81756d963b77","Type":"ContainerStarted","Data":"b96aa6598f8a4d2939a069c798538f5dfee7af33353b3513c9aa74d9490a0219"} Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.640522 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pgn88" event={"ID":"82ece3ff-72cb-48b3-92ad-251d7bf980c0","Type":"ContainerStarted","Data":"66be00b4282f77054a7bfa199987b27fc6e51cbc5232757e22a67fed6929ba0b"} Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.642294 4890 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-xldhl container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.642425 4890 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-xldhl" podUID="6e313ec3-70c2-4220-b08a-ef79b8a25847" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.647675 4890 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-mqvqx container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" start-of-body= Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.647739 4890 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mqvqx" podUID="e09b8f75-2286-4a12-9571-5342b681690f" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.650862 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-6dmh8" event={"ID":"8b5c8177-1153-453a-9110-6023d01e3fb7","Type":"ContainerStarted","Data":"d4083356f2499775c573c3c64a53092292d66261471beed139bbd0380450c42c"} Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.671392 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-z5tfd" event={"ID":"cb6875f9-0bd5-4076-848f-fd734d4f04df","Type":"ContainerStarted","Data":"cd2e5e45cd62b3835fa2cfdd127b4bb0c1dc03529bafdb4aec57db3d8c07e0cd"} Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.671450 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-z5tfd" event={"ID":"cb6875f9-0bd5-4076-848f-fd734d4f04df","Type":"ContainerStarted","Data":"685c5e8f941e354cc649b7cf7089ad67f25728908bc2d9b172ca63f285004b79"} Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.676481 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-46rkq" event={"ID":"21edd566-f5c3-425d-b719-d68d440884d8","Type":"ContainerStarted","Data":"5f4c7c413273f1ef6eedeccce1d1e834239b75debed56543e338d8d57a9e7d31"} Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.680561 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-kxfxn" event={"ID":"d153ec32-0e4e-41d6-b105-38f91ab40c14","Type":"ContainerStarted","Data":"9a67825a42e49c8dc9b71f0337033cd94d9682c4d98ae28e35833944bef5a31b"} Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.680625 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-kxfxn" event={"ID":"d153ec32-0e4e-41d6-b105-38f91ab40c14","Type":"ContainerStarted","Data":"3037982a861f3e86fb53e20adee495a0201c7a11e1c7116ea7a911ca62fdf7a1"} Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.683290 4890 patch_prober.go:28] interesting pod/downloads-7954f5f757-8bln9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.683359 4890 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-8bln9" podUID="89bebf40-d718-4168-8377-4fa334afaa17" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.684954 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jrdzr" podStartSLOduration=74.684937698 podStartE2EDuration="1m14.684937698s" podCreationTimestamp="2025-11-25 15:02:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:55.680399794 +0000 UTC m=+94.122862404" watchObservedRunningTime="2025-11-25 15:03:55.684937698 +0000 UTC m=+94.127400308" Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.704718 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:03:55 crc kubenswrapper[4890]: E1125 15:03:55.706513 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:03:56.206497828 +0000 UTC m=+94.648960438 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.813041 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:55 crc kubenswrapper[4890]: E1125 15:03:55.844980 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:03:56.344962087 +0000 UTC m=+94.787424697 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.848894 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-9l4z5" Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.866332 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-6dmh8" podStartSLOduration=73.86630504 podStartE2EDuration="1m13.86630504s" podCreationTimestamp="2025-11-25 15:02:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:55.863673638 +0000 UTC m=+94.306136248" watchObservedRunningTime="2025-11-25 15:03:55.86630504 +0000 UTC m=+94.308767650" Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.885917 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pgn88" podStartSLOduration=74.885898077 podStartE2EDuration="1m14.885898077s" podCreationTimestamp="2025-11-25 15:02:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:55.885593508 +0000 UTC m=+94.328056118" watchObservedRunningTime="2025-11-25 15:03:55.885898077 +0000 UTC m=+94.328360687" Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.927470 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:03:55 crc kubenswrapper[4890]: E1125 15:03:55.927809 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:03:56.427787153 +0000 UTC m=+94.870249763 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.928390 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:55 crc kubenswrapper[4890]: E1125 15:03:55.929710 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:03:56.429687135 +0000 UTC m=+94.872149745 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.929726 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.954987 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-8bln9" podStartSLOduration=74.954958816 podStartE2EDuration="1m14.954958816s" podCreationTimestamp="2025-11-25 15:02:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:55.954712459 +0000 UTC m=+94.397175079" watchObservedRunningTime="2025-11-25 15:03:55.954958816 +0000 UTC m=+94.397421426" Nov 25 15:03:55 crc kubenswrapper[4890]: I1125 15:03:55.960428 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-476wv" podStartSLOduration=74.960405265 podStartE2EDuration="1m14.960405265s" podCreationTimestamp="2025-11-25 15:02:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:55.928014989 +0000 UTC m=+94.370477609" watchObservedRunningTime="2025-11-25 15:03:55.960405265 +0000 UTC m=+94.402867875" Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.020339 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mqvqx" podStartSLOduration=74.020308814 podStartE2EDuration="1m14.020308814s" podCreationTimestamp="2025-11-25 15:02:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:56.011563315 +0000 UTC m=+94.454025925" watchObservedRunningTime="2025-11-25 15:03:56.020308814 +0000 UTC m=+94.462771424" Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.029527 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:03:56 crc kubenswrapper[4890]: E1125 15:03:56.030414 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:03:56.530374649 +0000 UTC m=+94.972837269 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.041066 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:56 crc kubenswrapper[4890]: E1125 15:03:56.041885 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:03:56.541849653 +0000 UTC m=+94.984312263 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.070747 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-m95j6" podStartSLOduration=75.070730284 podStartE2EDuration="1m15.070730284s" podCreationTimestamp="2025-11-25 15:02:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:56.06072561 +0000 UTC m=+94.503188210" watchObservedRunningTime="2025-11-25 15:03:56.070730284 +0000 UTC m=+94.513192894" Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.105265 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nhq4p" podStartSLOduration=75.105249988 podStartE2EDuration="1m15.105249988s" podCreationTimestamp="2025-11-25 15:02:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:56.103756857 +0000 UTC m=+94.546219457" watchObservedRunningTime="2025-11-25 15:03:56.105249988 +0000 UTC m=+94.547712598" Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.116224 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-kxfxn" Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.119520 4890 patch_prober.go:28] interesting pod/router-default-5444994796-kxfxn container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.119561 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kxfxn" podUID="d153ec32-0e4e-41d6-b105-38f91ab40c14" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.149845 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:03:56 crc kubenswrapper[4890]: E1125 15:03:56.150230 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:03:56.650215858 +0000 UTC m=+95.092678468 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.175215 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" podStartSLOduration=75.175195662 podStartE2EDuration="1m15.175195662s" podCreationTimestamp="2025-11-25 15:02:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:56.173432514 +0000 UTC m=+94.615895124" watchObservedRunningTime="2025-11-25 15:03:56.175195662 +0000 UTC m=+94.617658272" Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.176550 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-9l4z5" podStartSLOduration=75.176539108 podStartE2EDuration="1m15.176539108s" podCreationTimestamp="2025-11-25 15:02:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:56.131093515 +0000 UTC m=+94.573556125" watchObservedRunningTime="2025-11-25 15:03:56.176539108 +0000 UTC m=+94.619001718" Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.205582 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-t29fd" podStartSLOduration=75.205563153 podStartE2EDuration="1m15.205563153s" podCreationTimestamp="2025-11-25 15:02:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:56.205364497 +0000 UTC m=+94.647827187" watchObservedRunningTime="2025-11-25 15:03:56.205563153 +0000 UTC m=+94.648025763" Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.252604 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:56 crc kubenswrapper[4890]: E1125 15:03:56.253466 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:03:56.753451962 +0000 UTC m=+95.195914572 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.270477 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-k4hj2"] Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.311686 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-kxfxn" podStartSLOduration=75.311668395 podStartE2EDuration="1m15.311668395s" podCreationTimestamp="2025-11-25 15:02:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:56.272721949 +0000 UTC m=+94.715184569" watchObservedRunningTime="2025-11-25 15:03:56.311668395 +0000 UTC m=+94.754131015" Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.316026 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-pshcp"] Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.322863 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v6ps9" podStartSLOduration=75.32283821 podStartE2EDuration="1m15.32283821s" podCreationTimestamp="2025-11-25 15:02:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:56.308821527 +0000 UTC m=+94.751284147" watchObservedRunningTime="2025-11-25 15:03:56.32283821 +0000 UTC m=+94.765300820" Nov 25 15:03:56 crc kubenswrapper[4890]: W1125 15:03:56.328017 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode23712e1_8705_4d4b_bd1a_1bea23f765df.slice/crio-a66c95cd9b3e31aa5e645ebd5f809a0ee1f50c68adff74caa67be85b6ea5969c WatchSource:0}: Error finding container a66c95cd9b3e31aa5e645ebd5f809a0ee1f50c68adff74caa67be85b6ea5969c: Status 404 returned error can't find the container with id a66c95cd9b3e31aa5e645ebd5f809a0ee1f50c68adff74caa67be85b6ea5969c Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.353836 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:03:56 crc kubenswrapper[4890]: E1125 15:03:56.354141 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:03:56.854127646 +0000 UTC m=+95.296590256 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.374225 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fc2g9"] Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.374274 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-ffl2m"] Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.389486 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-klnk6"] Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.395607 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xmhq9"] Nov 25 15:03:56 crc kubenswrapper[4890]: W1125 15:03:56.402652 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod62ff675c_ff12_4c14_9c6b_ce3fb25ea214.slice/crio-e047b9e118196e1168481a0f61603cb5a19e05a9d6f3481332b500923c26002b WatchSource:0}: Error finding container e047b9e118196e1168481a0f61603cb5a19e05a9d6f3481332b500923c26002b: Status 404 returned error can't find the container with id e047b9e118196e1168481a0f61603cb5a19e05a9d6f3481332b500923c26002b Nov 25 15:03:56 crc kubenswrapper[4890]: W1125 15:03:56.408226 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod10bbf031_b282_4a4e_8fb9_e757afa8c022.slice/crio-d7d726c95e2cf6e1338d7441416b94f93a684880a0841173bfd30c32b9f0ff30 WatchSource:0}: Error finding container d7d726c95e2cf6e1338d7441416b94f93a684880a0841173bfd30c32b9f0ff30: Status 404 returned error can't find the container with id d7d726c95e2cf6e1338d7441416b94f93a684880a0841173bfd30c32b9f0ff30 Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.410212 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-t5v4f"] Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.454716 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nfp7k"] Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.455817 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:56 crc kubenswrapper[4890]: E1125 15:03:56.456114 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:03:56.956104096 +0000 UTC m=+95.398566706 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:56 crc kubenswrapper[4890]: W1125 15:03:56.525989 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode3475b27_002a_4087_ab65_9cdbad01ca92.slice/crio-064d324a6428118cca422b806653046034b997c95c37696da8599216b04dcd20 WatchSource:0}: Error finding container 064d324a6428118cca422b806653046034b997c95c37696da8599216b04dcd20: Status 404 returned error can't find the container with id 064d324a6428118cca422b806653046034b997c95c37696da8599216b04dcd20 Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.528570 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-xldhl" podStartSLOduration=75.528546229 podStartE2EDuration="1m15.528546229s" podCreationTimestamp="2025-11-25 15:02:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:56.410648693 +0000 UTC m=+94.853111313" watchObservedRunningTime="2025-11-25 15:03:56.528546229 +0000 UTC m=+94.971008839" Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.559844 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:03:56 crc kubenswrapper[4890]: E1125 15:03:56.560001 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:03:57.059980869 +0000 UTC m=+95.502443479 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.560068 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:56 crc kubenswrapper[4890]: E1125 15:03:56.560440 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:03:57.060433221 +0000 UTC m=+95.502895831 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:56 crc kubenswrapper[4890]: W1125 15:03:56.570019 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd5dd7cd5_c63e_460f_8d90_d91f5ec85f15.slice/crio-1033ef6c9d1a73f3fb64341657d7d8057d79aabc104466c1a98546299a3edbe3 WatchSource:0}: Error finding container 1033ef6c9d1a73f3fb64341657d7d8057d79aabc104466c1a98546299a3edbe3: Status 404 returned error can't find the container with id 1033ef6c9d1a73f3fb64341657d7d8057d79aabc104466c1a98546299a3edbe3 Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.587363 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-ms5f2"] Nov 25 15:03:56 crc kubenswrapper[4890]: W1125 15:03:56.587937 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4622b25d_fb03_4e9f_a66c_7f90def9b5f7.slice/crio-10b5fe4f40386180673adf770f673c3d1da6f2c20208c98dc6d43657b3f33122 WatchSource:0}: Error finding container 10b5fe4f40386180673adf770f673c3d1da6f2c20208c98dc6d43657b3f33122: Status 404 returned error can't find the container with id 10b5fe4f40386180673adf770f673c3d1da6f2c20208c98dc6d43657b3f33122 Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.592091 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-n7q4s"] Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.599099 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401380-96zpb"] Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.605404 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-99n87"] Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.613001 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-jkwk5"] Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.616630 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-rx9dj"] Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.662837 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:03:56 crc kubenswrapper[4890]: E1125 15:03:56.663255 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:03:57.163236384 +0000 UTC m=+95.605698994 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.690776 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-cmpbm"] Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.718456 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n7q4s" event={"ID":"5cf1f97f-ff78-4047-b839-6e8ffb6b92b9","Type":"ContainerStarted","Data":"c665574d4675e8dabec22f39969c3ddaf41ba60131852aed955fe9d69189b6a6"} Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.763910 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:56 crc kubenswrapper[4890]: E1125 15:03:56.764318 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:03:57.264305929 +0000 UTC m=+95.706768539 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.765976 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-c7vh9" event={"ID":"acb1ddd4-f551-4d06-9021-28d7579bad7f","Type":"ContainerStarted","Data":"9fdcd653154cff121c1bca50ac360e89e63320449438efef29aaaf7378821f28"} Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.767105 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-c7vh9" Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.768433 4890 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-c7vh9 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.18:8443/healthz\": dial tcp 10.217.0.18:8443: connect: connection refused" start-of-body= Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.768476 4890 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-c7vh9" podUID="acb1ddd4-f551-4d06-9021-28d7579bad7f" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.18:8443/healthz\": dial tcp 10.217.0.18:8443: connect: connection refused" Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.775064 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-v2tt6" event={"ID":"a0b0911d-7b7c-4746-bcfc-d15069434bb1","Type":"ContainerStarted","Data":"48cc524490cf04576eb5c218568c8bb54081b7f992c21151ba9cb5fece7ace70"} Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.775115 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-v2tt6" event={"ID":"a0b0911d-7b7c-4746-bcfc-d15069434bb1","Type":"ContainerStarted","Data":"74e78bc7747d588f8e3527110cededa2a86377bd70b1a6f12e9ac077dbfacb55"} Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.793429 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-c7vh9" podStartSLOduration=74.793414225 podStartE2EDuration="1m14.793414225s" podCreationTimestamp="2025-11-25 15:02:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:56.791182324 +0000 UTC m=+95.233644934" watchObservedRunningTime="2025-11-25 15:03:56.793414225 +0000 UTC m=+95.235876835" Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.801276 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401380-96zpb" event={"ID":"4622b25d-fb03-4e9f-a66c-7f90def9b5f7","Type":"ContainerStarted","Data":"10b5fe4f40386180673adf770f673c3d1da6f2c20208c98dc6d43657b3f33122"} Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.815463 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h7rbb" event={"ID":"34cf8610-cc4e-46de-a00c-ff41d7030e1f","Type":"ContainerStarted","Data":"0791527e972727aa7c988dd8b253d5394e6a1f9128522363ce2f91f8d19de2fd"} Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.816071 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-v2tt6" podStartSLOduration=74.816051555 podStartE2EDuration="1m14.816051555s" podCreationTimestamp="2025-11-25 15:02:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:56.815977163 +0000 UTC m=+95.258439773" watchObservedRunningTime="2025-11-25 15:03:56.816051555 +0000 UTC m=+95.258514175" Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.855427 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-rx9dj" event={"ID":"815dd5c7-dda6-42c5-b63f-8e354c5d1143","Type":"ContainerStarted","Data":"0bc8fc54a90f4d22e23bd1fe1a70ad57ddf13b2e32fbd8c524509442bab25ba7"} Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.867596 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:03:56 crc kubenswrapper[4890]: E1125 15:03:56.868708 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:03:57.368692725 +0000 UTC m=+95.811155335 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.898946 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-d7vvw" event={"ID":"ac102463-bf2a-466d-b24a-e3249661d1ee","Type":"ContainerStarted","Data":"e24105ec292a330940a7d8aedbd4387d1f7acc1befaeaa2b54248136f46549e5"} Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.902676 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-glmvm" event={"ID":"a6795dc3-313a-4599-a07c-588365bd8965","Type":"ContainerStarted","Data":"b692c88ab43c31065ac8619daf1fe1c9901afd8c4ec7ba69e4738c567887f0e1"} Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.911372 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-z5tfd" event={"ID":"cb6875f9-0bd5-4076-848f-fd734d4f04df","Type":"ContainerStarted","Data":"3d47d74b3f143801c4dd29cbaba97dc53f4e24a9da7830109ecaf6dc95266647"} Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.911636 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-z5tfd" Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.912580 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fc2g9" event={"ID":"10bbf031-b282-4a4e-8fb9-e757afa8c022","Type":"ContainerStarted","Data":"d7d726c95e2cf6e1338d7441416b94f93a684880a0841173bfd30c32b9f0ff30"} Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.913455 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-pshcp" event={"ID":"62ff675c-ff12-4c14-9c6b-ce3fb25ea214","Type":"ContainerStarted","Data":"e047b9e118196e1168481a0f61603cb5a19e05a9d6f3481332b500923c26002b"} Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.915825 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-99n87" event={"ID":"300178a7-3aa1-4fbb-a604-a37365cc968f","Type":"ContainerStarted","Data":"21ca5a474c8298778f8b8729a7e2b38765ea81d753fe168d75c880410976b387"} Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.921711 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-ms5f2" event={"ID":"79972ac1-e98f-4cd9-ba62-dd5d3b9e81e6","Type":"ContainerStarted","Data":"fd04dcfc0ff3c448c31b67fd849261f784acb7f9b46001e69633360f1b2c5877"} Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.925569 4890 generic.go:334] "Generic (PLEG): container finished" podID="91dd7b10-99ef-486d-a624-21cb245226c7" containerID="f062955bb28bfe4d059d73684fc80b0c88c7bb665b82e5a11dab9c391066a251" exitCode=0 Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.925673 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8zql" event={"ID":"91dd7b10-99ef-486d-a624-21cb245226c7","Type":"ContainerDied","Data":"f062955bb28bfe4d059d73684fc80b0c88c7bb665b82e5a11dab9c391066a251"} Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.925704 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8zql" event={"ID":"91dd7b10-99ef-486d-a624-21cb245226c7","Type":"ContainerStarted","Data":"64fd258f22d9b77711d5d1c36566c3be3e633d96776a8e22150af1a78d26b411"} Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.927581 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-k4hj2" event={"ID":"e23712e1-8705-4d4b-bd1a-1bea23f765df","Type":"ContainerStarted","Data":"a66c95cd9b3e31aa5e645ebd5f809a0ee1f50c68adff74caa67be85b6ea5969c"} Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.929072 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ktld7" event={"ID":"0826463c-2c8a-4b52-8727-4b07de3d86c3","Type":"ContainerStarted","Data":"af71e8cf085bec6d5c2a7ab60772926e2539246181916f15e30200f5b65e59b4"} Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.929114 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ktld7" event={"ID":"0826463c-2c8a-4b52-8727-4b07de3d86c3","Type":"ContainerStarted","Data":"5edcc50816642522acfd47f8b9d2ccc8849333a8c25e717b9c0809b68b9adc9c"} Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.934438 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h7rbb" podStartSLOduration=75.934417033 podStartE2EDuration="1m15.934417033s" podCreationTimestamp="2025-11-25 15:02:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:56.851173556 +0000 UTC m=+95.293636166" watchObservedRunningTime="2025-11-25 15:03:56.934417033 +0000 UTC m=+95.376879643" Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.936761 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-7qxf9" event={"ID":"b7e0b1cd-5c13-436d-8c35-c0a659f767a0","Type":"ContainerStarted","Data":"dd75971b87d2b128545f0a1fdaa77345ea9ee90cb02b3c397957e9938165376b"} Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.939208 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-z5tfd" podStartSLOduration=74.939194074 podStartE2EDuration="1m14.939194074s" podCreationTimestamp="2025-11-25 15:02:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:56.933717684 +0000 UTC m=+95.376180294" watchObservedRunningTime="2025-11-25 15:03:56.939194074 +0000 UTC m=+95.381656684" Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.941111 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-46rkq" event={"ID":"21edd566-f5c3-425d-b719-d68d440884d8","Type":"ContainerStarted","Data":"dc871ee5792811b2d29aeccc710b6db275c05fe59a97c49c20eaee26aa42235d"} Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.942175 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xmhq9" event={"ID":"8ef9a0ab-2131-42ff-823d-17d517929e77","Type":"ContainerStarted","Data":"d36a06d3be2645d92c3e0dac63b8d50beedb8861e77d4730d26ad04a620e0028"} Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.947122 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zs6w7" event={"ID":"a3f62807-e422-4d74-82c5-81756d963b77","Type":"ContainerStarted","Data":"f7b27039842b0223c44d9cae01efb06c71ffd4cf16e63f6653f08b64f0b22f3b"} Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.949233 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-ffl2m" event={"ID":"e2656d73-dfaf-4d40-a1b8-0d08297f9dc4","Type":"ContainerStarted","Data":"51f0585c614d4aa0a87748919b5d41787351f2ab9081060e7a385f4cb6ca95eb"} Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.957419 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jkwk5" event={"ID":"d5dd7cd5-c63e-460f-8d90-d91f5ec85f15","Type":"ContainerStarted","Data":"1033ef6c9d1a73f3fb64341657d7d8057d79aabc104466c1a98546299a3edbe3"} Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.961522 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-sm46z" event={"ID":"429acdcb-8b70-482c-abad-82f4867a9266","Type":"ContainerStarted","Data":"dfd4488c2b620383b1b17ac96e76e5a75045fe26a23ea2fdab2f59140f0601f3"} Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.972823 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t5v4f" event={"ID":"1a324db0-9f3e-417e-931f-5b74a44775ed","Type":"ContainerStarted","Data":"f4da298e788cac8db2d6b5a73103d0b594fedd79b00d18d15bddb9cbcb5a60fd"} Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.976025 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:56 crc kubenswrapper[4890]: E1125 15:03:56.977357 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:03:57.477334338 +0000 UTC m=+95.919797018 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:56 crc kubenswrapper[4890]: I1125 15:03:56.985358 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ktld7" podStartSLOduration=75.985337137 podStartE2EDuration="1m15.985337137s" podCreationTimestamp="2025-11-25 15:02:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:56.977986515 +0000 UTC m=+95.420449135" watchObservedRunningTime="2025-11-25 15:03:56.985337137 +0000 UTC m=+95.427799737" Nov 25 15:03:57 crc kubenswrapper[4890]: I1125 15:03:57.002549 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nfp7k" event={"ID":"e3475b27-002a-4087-ab65-9cdbad01ca92","Type":"ContainerStarted","Data":"064d324a6428118cca422b806653046034b997c95c37696da8599216b04dcd20"} Nov 25 15:03:57 crc kubenswrapper[4890]: I1125 15:03:57.008785 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-klnk6" event={"ID":"e682b6af-7f62-4efb-917e-7453e78138bc","Type":"ContainerStarted","Data":"dade232885b2ae9063311f9a9d473f6948afa649e60d1508272ef78ac537bd2e"} Nov 25 15:03:57 crc kubenswrapper[4890]: I1125 15:03:57.015711 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hd47j" event={"ID":"97fc0814-92d4-42a4-89c0-acb55bf47b4d","Type":"ContainerStarted","Data":"9a7464a0d70bb889ed3d438bfab5d5d5409c48af8154b8610f67f5f97270119a"} Nov 25 15:03:57 crc kubenswrapper[4890]: I1125 15:03:57.018707 4890 patch_prober.go:28] interesting pod/downloads-7954f5f757-8bln9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Nov 25 15:03:57 crc kubenswrapper[4890]: I1125 15:03:57.018738 4890 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-8bln9" podUID="89bebf40-d718-4168-8377-4fa334afaa17" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Nov 25 15:03:57 crc kubenswrapper[4890]: I1125 15:03:57.029364 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-xldhl" Nov 25 15:03:57 crc kubenswrapper[4890]: I1125 15:03:57.033061 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mqvqx" Nov 25 15:03:57 crc kubenswrapper[4890]: I1125 15:03:57.064198 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zs6w7" podStartSLOduration=75.064151623 podStartE2EDuration="1m15.064151623s" podCreationTimestamp="2025-11-25 15:02:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:57.007501213 +0000 UTC m=+95.449963843" watchObservedRunningTime="2025-11-25 15:03:57.064151623 +0000 UTC m=+95.506614233" Nov 25 15:03:57 crc kubenswrapper[4890]: I1125 15:03:57.107627 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:03:57 crc kubenswrapper[4890]: E1125 15:03:57.146344 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:03:57.646317131 +0000 UTC m=+96.088779741 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:57 crc kubenswrapper[4890]: I1125 15:03:57.146638 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:57 crc kubenswrapper[4890]: I1125 15:03:57.147538 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-7qxf9" podStartSLOduration=76.147519214 podStartE2EDuration="1m16.147519214s" podCreationTimestamp="2025-11-25 15:02:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:57.070833356 +0000 UTC m=+95.513295966" watchObservedRunningTime="2025-11-25 15:03:57.147519214 +0000 UTC m=+95.589981834" Nov 25 15:03:57 crc kubenswrapper[4890]: I1125 15:03:57.148116 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-sm46z" podStartSLOduration=6.14811047 podStartE2EDuration="6.14811047s" podCreationTimestamp="2025-11-25 15:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:57.145994912 +0000 UTC m=+95.588457522" watchObservedRunningTime="2025-11-25 15:03:57.14811047 +0000 UTC m=+95.590573080" Nov 25 15:03:57 crc kubenswrapper[4890]: I1125 15:03:57.162447 4890 patch_prober.go:28] interesting pod/router-default-5444994796-kxfxn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 15:03:57 crc kubenswrapper[4890]: [-]has-synced failed: reason withheld Nov 25 15:03:57 crc kubenswrapper[4890]: [+]process-running ok Nov 25 15:03:57 crc kubenswrapper[4890]: healthz check failed Nov 25 15:03:57 crc kubenswrapper[4890]: I1125 15:03:57.162536 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kxfxn" podUID="d153ec32-0e4e-41d6-b105-38f91ab40c14" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 15:03:57 crc kubenswrapper[4890]: E1125 15:03:57.165250 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:03:57.665229969 +0000 UTC m=+96.107692579 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:57 crc kubenswrapper[4890]: I1125 15:03:57.239418 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hd47j" podStartSLOduration=76.239391518 podStartE2EDuration="1m16.239391518s" podCreationTimestamp="2025-11-25 15:02:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:57.238526424 +0000 UTC m=+95.680989034" watchObservedRunningTime="2025-11-25 15:03:57.239391518 +0000 UTC m=+95.681854128" Nov 25 15:03:57 crc kubenswrapper[4890]: I1125 15:03:57.249012 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:03:57 crc kubenswrapper[4890]: E1125 15:03:57.249449 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:03:57.749432942 +0000 UTC m=+96.191895552 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:57 crc kubenswrapper[4890]: I1125 15:03:57.351874 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:57 crc kubenswrapper[4890]: E1125 15:03:57.353002 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:03:57.85278941 +0000 UTC m=+96.295252010 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:57 crc kubenswrapper[4890]: I1125 15:03:57.472949 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:03:57 crc kubenswrapper[4890]: E1125 15:03:57.473639 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:03:57.973619966 +0000 UTC m=+96.416082576 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:57 crc kubenswrapper[4890]: I1125 15:03:57.574413 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:57 crc kubenswrapper[4890]: E1125 15:03:57.574791 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:03:58.074771934 +0000 UTC m=+96.517234624 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:57 crc kubenswrapper[4890]: I1125 15:03:57.675758 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:03:57 crc kubenswrapper[4890]: E1125 15:03:57.676524 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:03:58.176505877 +0000 UTC m=+96.618968487 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:57 crc kubenswrapper[4890]: I1125 15:03:57.797371 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:57 crc kubenswrapper[4890]: E1125 15:03:57.798380 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:03:58.298368511 +0000 UTC m=+96.740831121 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:57 crc kubenswrapper[4890]: I1125 15:03:57.899775 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:03:57 crc kubenswrapper[4890]: E1125 15:03:57.900631 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:03:58.400609819 +0000 UTC m=+96.843072439 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.006873 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:58 crc kubenswrapper[4890]: E1125 15:03:58.007202 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:03:58.507190295 +0000 UTC m=+96.949652905 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.063212 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-glmvm" event={"ID":"a6795dc3-313a-4599-a07c-588365bd8965","Type":"ContainerStarted","Data":"963be7cb1b45e3dd0e34f2fe27d8a5c6cf80f14bcc2e78a402cd0c3ae3702bd5"} Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.076379 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nfp7k" event={"ID":"e3475b27-002a-4087-ab65-9cdbad01ca92","Type":"ContainerStarted","Data":"3ad457782335d2a96b4019822e86ca391f1423ea7b8f43308e0ac7ca20279057"} Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.077046 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nfp7k" Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.078412 4890 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-nfp7k container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.37:8443/healthz\": dial tcp 10.217.0.37:8443: connect: connection refused" start-of-body= Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.078468 4890 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nfp7k" podUID="e3475b27-002a-4087-ab65-9cdbad01ca92" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.37:8443/healthz\": dial tcp 10.217.0.37:8443: connect: connection refused" Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.087523 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-99n87" event={"ID":"300178a7-3aa1-4fbb-a604-a37365cc968f","Type":"ContainerStarted","Data":"a1f6789cb27c5fdbc9c52893628e631572d209abe05f12a9a9010b9506c8cc57"} Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.108313 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:03:58 crc kubenswrapper[4890]: E1125 15:03:58.110264 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:03:58.610205133 +0000 UTC m=+97.052667743 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.111855 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-cmpbm" event={"ID":"ba947c9c-cd84-4c9b-a431-9276412058d1","Type":"ContainerStarted","Data":"5bb55d77893909d7ad1bbc9cf1ec5e00e2ef58308027952d402efb14dbb14efc"} Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.127996 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-ffl2m" event={"ID":"e2656d73-dfaf-4d40-a1b8-0d08297f9dc4","Type":"ContainerStarted","Data":"4415d4cbef2d63fad8d8f2b2e7fc8da7b3d3c02dbc5b8d28aa000958752d18b7"} Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.141694 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-glmvm" podStartSLOduration=77.141675664 podStartE2EDuration="1m17.141675664s" podCreationTimestamp="2025-11-25 15:02:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:58.12359874 +0000 UTC m=+96.566061350" watchObservedRunningTime="2025-11-25 15:03:58.141675664 +0000 UTC m=+96.584138274" Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.142373 4890 patch_prober.go:28] interesting pod/router-default-5444994796-kxfxn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 15:03:58 crc kubenswrapper[4890]: [-]has-synced failed: reason withheld Nov 25 15:03:58 crc kubenswrapper[4890]: [+]process-running ok Nov 25 15:03:58 crc kubenswrapper[4890]: healthz check failed Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.142430 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kxfxn" podUID="d153ec32-0e4e-41d6-b105-38f91ab40c14" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.154436 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t5v4f" event={"ID":"1a324db0-9f3e-417e-931f-5b74a44775ed","Type":"ContainerStarted","Data":"75b18b19f291c96c960602b0f97c786cae7bb4247d1d4501293b43c870dbfe86"} Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.169654 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n7q4s" event={"ID":"5cf1f97f-ff78-4047-b839-6e8ffb6b92b9","Type":"ContainerStarted","Data":"e4813d16e3c4aa2cb83c2b5195c656da71f86e68ab9213c1e13e9e3a14257916"} Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.169706 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n7q4s" event={"ID":"5cf1f97f-ff78-4047-b839-6e8ffb6b92b9","Type":"ContainerStarted","Data":"b77b709f5951f0bb8d25f9b17ee4c48a08da8946a6aa6c3a85b8fcc59000a3f6"} Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.198515 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-99n87" podStartSLOduration=76.198497749 podStartE2EDuration="1m16.198497749s" podCreationTimestamp="2025-11-25 15:02:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:58.143077963 +0000 UTC m=+96.585540573" watchObservedRunningTime="2025-11-25 15:03:58.198497749 +0000 UTC m=+96.640960359" Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.210890 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:58 crc kubenswrapper[4890]: E1125 15:03:58.214124 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:03:58.714109906 +0000 UTC m=+97.156572516 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.233330 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nfp7k" podStartSLOduration=76.233311142 podStartE2EDuration="1m16.233311142s" podCreationTimestamp="2025-11-25 15:02:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:58.200280318 +0000 UTC m=+96.642742928" watchObservedRunningTime="2025-11-25 15:03:58.233311142 +0000 UTC m=+96.675773772" Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.234299 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n7q4s" podStartSLOduration=77.234292438 podStartE2EDuration="1m17.234292438s" podCreationTimestamp="2025-11-25 15:02:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:58.23141069 +0000 UTC m=+96.673873300" watchObservedRunningTime="2025-11-25 15:03:58.234292438 +0000 UTC m=+96.676755068" Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.250459 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-d7vvw" event={"ID":"ac102463-bf2a-466d-b24a-e3249661d1ee","Type":"ContainerStarted","Data":"4f7c37456c2833284f545b12be019e7788c61e24d8d0da5ac8141a1dc0f2ab49"} Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.270611 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-k4hj2" event={"ID":"e23712e1-8705-4d4b-bd1a-1bea23f765df","Type":"ContainerStarted","Data":"601f892bb6f81bd5633adc02f9b40dc623de4749b1bc2601b8c644534bc5030e"} Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.317047 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:03:58 crc kubenswrapper[4890]: E1125 15:03:58.317328 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:03:58.817292239 +0000 UTC m=+97.259754859 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.317666 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.320411 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jkwk5" event={"ID":"d5dd7cd5-c63e-460f-8d90-d91f5ec85f15","Type":"ContainerStarted","Data":"ab42d6cc070e0a94660e6f56f4ab1688d262590eb7066c0b600805f1c10d0271"} Nov 25 15:03:58 crc kubenswrapper[4890]: E1125 15:03:58.321127 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:03:58.821109634 +0000 UTC m=+97.263572244 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.331215 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-d7vvw" podStartSLOduration=77.331129898 podStartE2EDuration="1m17.331129898s" podCreationTimestamp="2025-11-25 15:02:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:58.328900587 +0000 UTC m=+96.771363197" watchObservedRunningTime="2025-11-25 15:03:58.331129898 +0000 UTC m=+96.773592508" Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.374872 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401380-96zpb" event={"ID":"4622b25d-fb03-4e9f-a66c-7f90def9b5f7","Type":"ContainerStarted","Data":"d69aed072dd6ab3043fb1ed07aa7e23803f193be0d9017879940b070a3669203"} Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.384702 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xmhq9" event={"ID":"8ef9a0ab-2131-42ff-823d-17d517929e77","Type":"ContainerStarted","Data":"dca75b5378b4c5d924995da03d2d3dd4b2ea0a61457719ab5b6cd93de1c5349e"} Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.385685 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-xmhq9" Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.392105 4890 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-xmhq9 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" start-of-body= Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.392190 4890 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-xmhq9" podUID="8ef9a0ab-2131-42ff-823d-17d517929e77" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.399448 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fc2g9" event={"ID":"10bbf031-b282-4a4e-8fb9-e757afa8c022","Type":"ContainerStarted","Data":"30ddcf9ce1050a668679e68e987a41b9815e329c64f5c69574cbb977214d917f"} Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.400530 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fc2g9" Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.401768 4890 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-fc2g9 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.27:5443/healthz\": dial tcp 10.217.0.27:5443: connect: connection refused" start-of-body= Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.401812 4890 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fc2g9" podUID="10bbf031-b282-4a4e-8fb9-e757afa8c022" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.27:5443/healthz\": dial tcp 10.217.0.27:5443: connect: connection refused" Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.420884 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:03:58 crc kubenswrapper[4890]: E1125 15:03:58.421452 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:03:58.921397118 +0000 UTC m=+97.363859728 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.428531 4890 generic.go:334] "Generic (PLEG): container finished" podID="62ff675c-ff12-4c14-9c6b-ce3fb25ea214" containerID="cf54ce46f7d6c05c7ec38d4f07f9feae917a832c544e99960d2b90d989e9bc76" exitCode=0 Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.428606 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-pshcp" event={"ID":"62ff675c-ff12-4c14-9c6b-ce3fb25ea214","Type":"ContainerDied","Data":"cf54ce46f7d6c05c7ec38d4f07f9feae917a832c544e99960d2b90d989e9bc76"} Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.454866 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-klnk6" event={"ID":"e682b6af-7f62-4efb-917e-7453e78138bc","Type":"ContainerStarted","Data":"fc93e3b30ddc420cf57e2144d7b2abb7a47fadfdb321f11b3edf64b46b65a9b1"} Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.461307 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-rx9dj" event={"ID":"815dd5c7-dda6-42c5-b63f-8e354c5d1143","Type":"ContainerStarted","Data":"828957beee821647c03eaa8cf7edb8aef0dcd9f631d0cafc6426b8ad1bc842f5"} Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.492664 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-glmvm" Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.492764 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-glmvm" Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.520699 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-c7vh9" Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.528411 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:58 crc kubenswrapper[4890]: E1125 15:03:58.531798 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:03:59.031782128 +0000 UTC m=+97.474244738 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.583970 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-k4hj2" podStartSLOduration=77.583949545 podStartE2EDuration="1m17.583949545s" podCreationTimestamp="2025-11-25 15:02:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:58.403977641 +0000 UTC m=+96.846440251" watchObservedRunningTime="2025-11-25 15:03:58.583949545 +0000 UTC m=+97.026412155" Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.634371 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:03:58 crc kubenswrapper[4890]: E1125 15:03:58.634740 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:03:59.134724755 +0000 UTC m=+97.577187365 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.699118 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jkwk5" podStartSLOduration=76.699098636 podStartE2EDuration="1m16.699098636s" podCreationTimestamp="2025-11-25 15:02:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:58.584516241 +0000 UTC m=+97.026978851" watchObservedRunningTime="2025-11-25 15:03:58.699098636 +0000 UTC m=+97.141561246" Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.700459 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-xmhq9" podStartSLOduration=76.700452353 podStartE2EDuration="1m16.700452353s" podCreationTimestamp="2025-11-25 15:02:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:58.698467759 +0000 UTC m=+97.140930389" watchObservedRunningTime="2025-11-25 15:03:58.700452353 +0000 UTC m=+97.142914963" Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.735566 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:58 crc kubenswrapper[4890]: E1125 15:03:58.736430 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:03:59.236416027 +0000 UTC m=+97.678878637 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.840507 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:03:58 crc kubenswrapper[4890]: E1125 15:03:58.840911 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:03:59.340895995 +0000 UTC m=+97.783358605 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.942170 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:58 crc kubenswrapper[4890]: E1125 15:03:58.942621 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:03:59.442603758 +0000 UTC m=+97.885066368 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.979384 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fc2g9" podStartSLOduration=76.979366884 podStartE2EDuration="1m16.979366884s" podCreationTimestamp="2025-11-25 15:02:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:58.921677026 +0000 UTC m=+97.364139666" watchObservedRunningTime="2025-11-25 15:03:58.979366884 +0000 UTC m=+97.421829494" Nov 25 15:03:58 crc kubenswrapper[4890]: I1125 15:03:58.979997 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-klnk6" podStartSLOduration=76.979992521 podStartE2EDuration="1m16.979992521s" podCreationTimestamp="2025-11-25 15:02:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:58.973822582 +0000 UTC m=+97.416285192" watchObservedRunningTime="2025-11-25 15:03:58.979992521 +0000 UTC m=+97.422455131" Nov 25 15:03:59 crc kubenswrapper[4890]: I1125 15:03:59.002872 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29401380-96zpb" podStartSLOduration=77.002856757 podStartE2EDuration="1m17.002856757s" podCreationTimestamp="2025-11-25 15:02:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:59.002088756 +0000 UTC m=+97.444551356" watchObservedRunningTime="2025-11-25 15:03:59.002856757 +0000 UTC m=+97.445319367" Nov 25 15:03:59 crc kubenswrapper[4890]: I1125 15:03:59.043212 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:03:59 crc kubenswrapper[4890]: E1125 15:03:59.043604 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:03:59.543587731 +0000 UTC m=+97.986050341 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:59 crc kubenswrapper[4890]: I1125 15:03:59.105810 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-rx9dj" podStartSLOduration=8.105791973 podStartE2EDuration="8.105791973s" podCreationTimestamp="2025-11-25 15:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:59.104665762 +0000 UTC m=+97.547128372" watchObservedRunningTime="2025-11-25 15:03:59.105791973 +0000 UTC m=+97.548254583" Nov 25 15:03:59 crc kubenswrapper[4890]: I1125 15:03:59.118757 4890 patch_prober.go:28] interesting pod/router-default-5444994796-kxfxn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 15:03:59 crc kubenswrapper[4890]: [-]has-synced failed: reason withheld Nov 25 15:03:59 crc kubenswrapper[4890]: [+]process-running ok Nov 25 15:03:59 crc kubenswrapper[4890]: healthz check failed Nov 25 15:03:59 crc kubenswrapper[4890]: I1125 15:03:59.119043 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kxfxn" podUID="d153ec32-0e4e-41d6-b105-38f91ab40c14" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 15:03:59 crc kubenswrapper[4890]: I1125 15:03:59.145109 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:59 crc kubenswrapper[4890]: E1125 15:03:59.145493 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:03:59.645482129 +0000 UTC m=+98.087944739 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:59 crc kubenswrapper[4890]: I1125 15:03:59.246534 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:03:59 crc kubenswrapper[4890]: E1125 15:03:59.246930 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:03:59.746911044 +0000 UTC m=+98.189373654 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:59 crc kubenswrapper[4890]: I1125 15:03:59.347930 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:59 crc kubenswrapper[4890]: E1125 15:03:59.348343 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:03:59.848329229 +0000 UTC m=+98.290791839 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:59 crc kubenswrapper[4890]: I1125 15:03:59.448876 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:03:59 crc kubenswrapper[4890]: E1125 15:03:59.449285 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:03:59.949267291 +0000 UTC m=+98.391729901 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:59 crc kubenswrapper[4890]: I1125 15:03:59.466764 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8zql" event={"ID":"91dd7b10-99ef-486d-a624-21cb245226c7","Type":"ContainerStarted","Data":"8b572ec2251cad2d47bdb954e19b0d6aa61d2b49d0b40ecaa3dcd8c407d5f46e"} Nov 25 15:03:59 crc kubenswrapper[4890]: I1125 15:03:59.468461 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-ffl2m" event={"ID":"e2656d73-dfaf-4d40-a1b8-0d08297f9dc4","Type":"ContainerStarted","Data":"1cefc7e22ecce70825ffa8b4cc2ffdd8f145444a8864a4d3257ab648b677a573"} Nov 25 15:03:59 crc kubenswrapper[4890]: I1125 15:03:59.470190 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jkwk5" event={"ID":"d5dd7cd5-c63e-460f-8d90-d91f5ec85f15","Type":"ContainerStarted","Data":"0f0933d56a752e407abfd5b8196c36e450b0be341f97046e24eb6fb353f2bd57"} Nov 25 15:03:59 crc kubenswrapper[4890]: I1125 15:03:59.472065 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-46rkq" event={"ID":"21edd566-f5c3-425d-b719-d68d440884d8","Type":"ContainerStarted","Data":"394657a750760c572cad043b19d6667654a546d472e895b7017a43e4a9fb75c7"} Nov 25 15:03:59 crc kubenswrapper[4890]: I1125 15:03:59.473567 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t5v4f" event={"ID":"1a324db0-9f3e-417e-931f-5b74a44775ed","Type":"ContainerStarted","Data":"1cfe7406dea50b1fca5b2c45b4ef19dfd70907e04db76fd5e4a2f8bcf254a13c"} Nov 25 15:03:59 crc kubenswrapper[4890]: I1125 15:03:59.478444 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-ms5f2" event={"ID":"79972ac1-e98f-4cd9-ba62-dd5d3b9e81e6","Type":"ContainerStarted","Data":"33aaaa6b5ae7bace7d641cb0aff367b2be986af98486d471686af62fae59f3ba"} Nov 25 15:03:59 crc kubenswrapper[4890]: I1125 15:03:59.478493 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-ms5f2" event={"ID":"79972ac1-e98f-4cd9-ba62-dd5d3b9e81e6","Type":"ContainerStarted","Data":"69e683bc5783ccbd4f54bb5ee2ff8052d7c18739e2a4c018dbe6f8706bbaa7f6"} Nov 25 15:03:59 crc kubenswrapper[4890]: I1125 15:03:59.478614 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-ms5f2" Nov 25 15:03:59 crc kubenswrapper[4890]: I1125 15:03:59.480574 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-pshcp" event={"ID":"62ff675c-ff12-4c14-9c6b-ce3fb25ea214","Type":"ContainerStarted","Data":"0af4244f067bf615d54f4074575777b68a17eaac1124453b7603859693014b64"} Nov 25 15:03:59 crc kubenswrapper[4890]: I1125 15:03:59.482744 4890 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-nfp7k container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.37:8443/healthz\": dial tcp 10.217.0.37:8443: connect: connection refused" start-of-body= Nov 25 15:03:59 crc kubenswrapper[4890]: I1125 15:03:59.482765 4890 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-xmhq9 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" start-of-body= Nov 25 15:03:59 crc kubenswrapper[4890]: I1125 15:03:59.482789 4890 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nfp7k" podUID="e3475b27-002a-4087-ab65-9cdbad01ca92" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.37:8443/healthz\": dial tcp 10.217.0.37:8443: connect: connection refused" Nov 25 15:03:59 crc kubenswrapper[4890]: I1125 15:03:59.482817 4890 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-xmhq9" podUID="8ef9a0ab-2131-42ff-823d-17d517929e77" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" Nov 25 15:03:59 crc kubenswrapper[4890]: I1125 15:03:59.552116 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:59 crc kubenswrapper[4890]: E1125 15:03:59.553105 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:04:00.053091081 +0000 UTC m=+98.495553691 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:59 crc kubenswrapper[4890]: I1125 15:03:59.587508 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-ffl2m" podStartSLOduration=77.587491392 podStartE2EDuration="1m17.587491392s" podCreationTimestamp="2025-11-25 15:02:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:59.585618851 +0000 UTC m=+98.028081461" watchObservedRunningTime="2025-11-25 15:03:59.587491392 +0000 UTC m=+98.029954003" Nov 25 15:03:59 crc kubenswrapper[4890]: I1125 15:03:59.602262 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8zql" podStartSLOduration=77.602218715 podStartE2EDuration="1m17.602218715s" podCreationTimestamp="2025-11-25 15:02:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:59.531073409 +0000 UTC m=+97.973536029" watchObservedRunningTime="2025-11-25 15:03:59.602218715 +0000 UTC m=+98.044681325" Nov 25 15:03:59 crc kubenswrapper[4890]: I1125 15:03:59.655419 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:03:59 crc kubenswrapper[4890]: E1125 15:03:59.655856 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:04:00.155838113 +0000 UTC m=+98.598300733 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:59 crc kubenswrapper[4890]: I1125 15:03:59.667230 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-pshcp" podStartSLOduration=78.667210294 podStartE2EDuration="1m18.667210294s" podCreationTimestamp="2025-11-25 15:02:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:59.666893935 +0000 UTC m=+98.109356545" watchObservedRunningTime="2025-11-25 15:03:59.667210294 +0000 UTC m=+98.109672904" Nov 25 15:03:59 crc kubenswrapper[4890]: I1125 15:03:59.741493 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-ms5f2" podStartSLOduration=8.741471805 podStartE2EDuration="8.741471805s" podCreationTimestamp="2025-11-25 15:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:59.738642168 +0000 UTC m=+98.181104778" watchObservedRunningTime="2025-11-25 15:03:59.741471805 +0000 UTC m=+98.183934415" Nov 25 15:03:59 crc kubenswrapper[4890]: I1125 15:03:59.756834 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:59 crc kubenswrapper[4890]: E1125 15:03:59.757445 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:04:00.257430712 +0000 UTC m=+98.699893322 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:59 crc kubenswrapper[4890]: I1125 15:03:59.845020 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-46rkq" podStartSLOduration=78.844991817 podStartE2EDuration="1m18.844991817s" podCreationTimestamp="2025-11-25 15:02:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:59.84182691 +0000 UTC m=+98.284289530" watchObservedRunningTime="2025-11-25 15:03:59.844991817 +0000 UTC m=+98.287454427" Nov 25 15:03:59 crc kubenswrapper[4890]: I1125 15:03:59.858315 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:03:59 crc kubenswrapper[4890]: E1125 15:03:59.858536 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:04:00.358503376 +0000 UTC m=+98.800966006 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:59 crc kubenswrapper[4890]: I1125 15:03:59.858985 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:03:59 crc kubenswrapper[4890]: E1125 15:03:59.859538 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:04:00.359521454 +0000 UTC m=+98.801984064 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:59 crc kubenswrapper[4890]: I1125 15:03:59.961001 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:03:59 crc kubenswrapper[4890]: E1125 15:03:59.961558 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:04:00.461523755 +0000 UTC m=+98.903986365 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:03:59 crc kubenswrapper[4890]: I1125 15:03:59.970895 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t5v4f" podStartSLOduration=77.970870901 podStartE2EDuration="1m17.970870901s" podCreationTimestamp="2025-11-25 15:02:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:03:59.97046206 +0000 UTC m=+98.412924670" watchObservedRunningTime="2025-11-25 15:03:59.970870901 +0000 UTC m=+98.413333511" Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.063108 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:04:00 crc kubenswrapper[4890]: E1125 15:04:00.063593 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:04:00.563572097 +0000 UTC m=+99.006034767 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.121183 4890 patch_prober.go:28] interesting pod/router-default-5444994796-kxfxn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 15:04:00 crc kubenswrapper[4890]: [-]has-synced failed: reason withheld Nov 25 15:04:00 crc kubenswrapper[4890]: [+]process-running ok Nov 25 15:04:00 crc kubenswrapper[4890]: healthz check failed Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.121288 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kxfxn" podUID="d153ec32-0e4e-41d6-b105-38f91ab40c14" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.154055 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-z4b8p"] Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.155191 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z4b8p" Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.164029 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:04:00 crc kubenswrapper[4890]: E1125 15:04:00.164298 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:04:00.664254732 +0000 UTC m=+99.106717342 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.164551 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:04:00 crc kubenswrapper[4890]: E1125 15:04:00.165041 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:04:00.665031793 +0000 UTC m=+99.107494403 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.170450 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.213713 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fc2g9" Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.234684 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-pshcp" Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.258886 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-z4b8p"] Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.265793 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.266219 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/16f52653-c067-4186-bc40-9c0f7c383b64-metrics-certs\") pod \"network-metrics-daemon-2nlkp\" (UID: \"16f52653-c067-4186-bc40-9c0f7c383b64\") " pod="openshift-multus/network-metrics-daemon-2nlkp" Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.266299 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2drz\" (UniqueName: \"kubernetes.io/projected/b1983120-f62d-46b0-833e-db30a81973c1-kube-api-access-x2drz\") pod \"certified-operators-z4b8p\" (UID: \"b1983120-f62d-46b0-833e-db30a81973c1\") " pod="openshift-marketplace/certified-operators-z4b8p" Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.266326 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1983120-f62d-46b0-833e-db30a81973c1-catalog-content\") pod \"certified-operators-z4b8p\" (UID: \"b1983120-f62d-46b0-833e-db30a81973c1\") " pod="openshift-marketplace/certified-operators-z4b8p" Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.266355 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1983120-f62d-46b0-833e-db30a81973c1-utilities\") pod \"certified-operators-z4b8p\" (UID: \"b1983120-f62d-46b0-833e-db30a81973c1\") " pod="openshift-marketplace/certified-operators-z4b8p" Nov 25 15:04:00 crc kubenswrapper[4890]: E1125 15:04:00.266503 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:04:00.766478249 +0000 UTC m=+99.208940859 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.273148 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/16f52653-c067-4186-bc40-9c0f7c383b64-metrics-certs\") pod \"network-metrics-daemon-2nlkp\" (UID: \"16f52653-c067-4186-bc40-9c0f7c383b64\") " pod="openshift-multus/network-metrics-daemon-2nlkp" Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.277080 4890 patch_prober.go:28] interesting pod/apiserver-76f77b778f-glmvm container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Nov 25 15:04:00 crc kubenswrapper[4890]: [+]log ok Nov 25 15:04:00 crc kubenswrapper[4890]: [+]etcd ok Nov 25 15:04:00 crc kubenswrapper[4890]: [+]poststarthook/start-apiserver-admission-initializer ok Nov 25 15:04:00 crc kubenswrapper[4890]: [+]poststarthook/generic-apiserver-start-informers ok Nov 25 15:04:00 crc kubenswrapper[4890]: [+]poststarthook/max-in-flight-filter ok Nov 25 15:04:00 crc kubenswrapper[4890]: [+]poststarthook/storage-object-count-tracker-hook ok Nov 25 15:04:00 crc kubenswrapper[4890]: [+]poststarthook/image.openshift.io-apiserver-caches ok Nov 25 15:04:00 crc kubenswrapper[4890]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Nov 25 15:04:00 crc kubenswrapper[4890]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Nov 25 15:04:00 crc kubenswrapper[4890]: [+]poststarthook/project.openshift.io-projectcache ok Nov 25 15:04:00 crc kubenswrapper[4890]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Nov 25 15:04:00 crc kubenswrapper[4890]: [+]poststarthook/openshift.io-startinformers ok Nov 25 15:04:00 crc kubenswrapper[4890]: [+]poststarthook/openshift.io-restmapperupdater ok Nov 25 15:04:00 crc kubenswrapper[4890]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Nov 25 15:04:00 crc kubenswrapper[4890]: livez check failed Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.277146 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-glmvm" podUID="a6795dc3-313a-4599-a07c-588365bd8965" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.300406 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2nlkp" Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.333983 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tfnr6"] Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.335571 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tfnr6" Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.356677 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tfnr6"] Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.368071 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1983120-f62d-46b0-833e-db30a81973c1-utilities\") pod \"certified-operators-z4b8p\" (UID: \"b1983120-f62d-46b0-833e-db30a81973c1\") " pod="openshift-marketplace/certified-operators-z4b8p" Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.368148 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:04:00 crc kubenswrapper[4890]: E1125 15:04:00.368667 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:04:00.868648234 +0000 UTC m=+99.311111024 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.368850 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2drz\" (UniqueName: \"kubernetes.io/projected/b1983120-f62d-46b0-833e-db30a81973c1-kube-api-access-x2drz\") pod \"certified-operators-z4b8p\" (UID: \"b1983120-f62d-46b0-833e-db30a81973c1\") " pod="openshift-marketplace/certified-operators-z4b8p" Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.368904 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1983120-f62d-46b0-833e-db30a81973c1-catalog-content\") pod \"certified-operators-z4b8p\" (UID: \"b1983120-f62d-46b0-833e-db30a81973c1\") " pod="openshift-marketplace/certified-operators-z4b8p" Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.369307 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1983120-f62d-46b0-833e-db30a81973c1-utilities\") pod \"certified-operators-z4b8p\" (UID: \"b1983120-f62d-46b0-833e-db30a81973c1\") " pod="openshift-marketplace/certified-operators-z4b8p" Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.369389 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1983120-f62d-46b0-833e-db30a81973c1-catalog-content\") pod \"certified-operators-z4b8p\" (UID: \"b1983120-f62d-46b0-833e-db30a81973c1\") " pod="openshift-marketplace/certified-operators-z4b8p" Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.377668 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.403150 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2drz\" (UniqueName: \"kubernetes.io/projected/b1983120-f62d-46b0-833e-db30a81973c1-kube-api-access-x2drz\") pod \"certified-operators-z4b8p\" (UID: \"b1983120-f62d-46b0-833e-db30a81973c1\") " pod="openshift-marketplace/certified-operators-z4b8p" Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.465883 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-t2pmf"] Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.466962 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t2pmf" Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.470681 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.470929 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f052eff-1a42-4c20-8644-d575fc557060-catalog-content\") pod \"community-operators-tfnr6\" (UID: \"8f052eff-1a42-4c20-8644-d575fc557060\") " pod="openshift-marketplace/community-operators-tfnr6" Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.471047 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlz6n\" (UniqueName: \"kubernetes.io/projected/8f052eff-1a42-4c20-8644-d575fc557060-kube-api-access-mlz6n\") pod \"community-operators-tfnr6\" (UID: \"8f052eff-1a42-4c20-8644-d575fc557060\") " pod="openshift-marketplace/community-operators-tfnr6" Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.471071 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f052eff-1a42-4c20-8644-d575fc557060-utilities\") pod \"community-operators-tfnr6\" (UID: \"8f052eff-1a42-4c20-8644-d575fc557060\") " pod="openshift-marketplace/community-operators-tfnr6" Nov 25 15:04:00 crc kubenswrapper[4890]: E1125 15:04:00.471182 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:04:00.971151439 +0000 UTC m=+99.413614049 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.475872 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z4b8p" Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.506125 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-t2pmf"] Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.538569 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-cmpbm" event={"ID":"ba947c9c-cd84-4c9b-a431-9276412058d1","Type":"ContainerStarted","Data":"1b0b81e7ff60286a2e6e29cccee9ed8b4430376eff5d4d4b375e5540bb0067af"} Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.539238 4890 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-xmhq9 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" start-of-body= Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.539267 4890 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-xmhq9" podUID="8ef9a0ab-2131-42ff-823d-17d517929e77" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.583942 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f56ea937-8ed8-4912-b721-2ed83a15fa2b-utilities\") pod \"certified-operators-t2pmf\" (UID: \"f56ea937-8ed8-4912-b721-2ed83a15fa2b\") " pod="openshift-marketplace/certified-operators-t2pmf" Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.584025 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kw68\" (UniqueName: \"kubernetes.io/projected/f56ea937-8ed8-4912-b721-2ed83a15fa2b-kube-api-access-4kw68\") pod \"certified-operators-t2pmf\" (UID: \"f56ea937-8ed8-4912-b721-2ed83a15fa2b\") " pod="openshift-marketplace/certified-operators-t2pmf" Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.584074 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.584131 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlz6n\" (UniqueName: \"kubernetes.io/projected/8f052eff-1a42-4c20-8644-d575fc557060-kube-api-access-mlz6n\") pod \"community-operators-tfnr6\" (UID: \"8f052eff-1a42-4c20-8644-d575fc557060\") " pod="openshift-marketplace/community-operators-tfnr6" Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.584157 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f052eff-1a42-4c20-8644-d575fc557060-utilities\") pod \"community-operators-tfnr6\" (UID: \"8f052eff-1a42-4c20-8644-d575fc557060\") " pod="openshift-marketplace/community-operators-tfnr6" Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.584201 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f56ea937-8ed8-4912-b721-2ed83a15fa2b-catalog-content\") pod \"certified-operators-t2pmf\" (UID: \"f56ea937-8ed8-4912-b721-2ed83a15fa2b\") " pod="openshift-marketplace/certified-operators-t2pmf" Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.584262 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f052eff-1a42-4c20-8644-d575fc557060-catalog-content\") pod \"community-operators-tfnr6\" (UID: \"8f052eff-1a42-4c20-8644-d575fc557060\") " pod="openshift-marketplace/community-operators-tfnr6" Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.585310 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f052eff-1a42-4c20-8644-d575fc557060-catalog-content\") pod \"community-operators-tfnr6\" (UID: \"8f052eff-1a42-4c20-8644-d575fc557060\") " pod="openshift-marketplace/community-operators-tfnr6" Nov 25 15:04:00 crc kubenswrapper[4890]: E1125 15:04:00.585616 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:04:01.08560073 +0000 UTC m=+99.528063340 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.586457 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f052eff-1a42-4c20-8644-d575fc557060-utilities\") pod \"community-operators-tfnr6\" (UID: \"8f052eff-1a42-4c20-8644-d575fc557060\") " pod="openshift-marketplace/community-operators-tfnr6" Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.656389 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlz6n\" (UniqueName: \"kubernetes.io/projected/8f052eff-1a42-4c20-8644-d575fc557060-kube-api-access-mlz6n\") pod \"community-operators-tfnr6\" (UID: \"8f052eff-1a42-4c20-8644-d575fc557060\") " pod="openshift-marketplace/community-operators-tfnr6" Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.692551 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mh5wg"] Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.694715 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mh5wg" Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.695971 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.696516 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kw68\" (UniqueName: \"kubernetes.io/projected/f56ea937-8ed8-4912-b721-2ed83a15fa2b-kube-api-access-4kw68\") pod \"certified-operators-t2pmf\" (UID: \"f56ea937-8ed8-4912-b721-2ed83a15fa2b\") " pod="openshift-marketplace/certified-operators-t2pmf" Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.728187 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f56ea937-8ed8-4912-b721-2ed83a15fa2b-catalog-content\") pod \"certified-operators-t2pmf\" (UID: \"f56ea937-8ed8-4912-b721-2ed83a15fa2b\") " pod="openshift-marketplace/certified-operators-t2pmf" Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.728671 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f56ea937-8ed8-4912-b721-2ed83a15fa2b-catalog-content\") pod \"certified-operators-t2pmf\" (UID: \"f56ea937-8ed8-4912-b721-2ed83a15fa2b\") " pod="openshift-marketplace/certified-operators-t2pmf" Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.730848 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f56ea937-8ed8-4912-b721-2ed83a15fa2b-utilities\") pod \"certified-operators-t2pmf\" (UID: \"f56ea937-8ed8-4912-b721-2ed83a15fa2b\") " pod="openshift-marketplace/certified-operators-t2pmf" Nov 25 15:04:00 crc kubenswrapper[4890]: E1125 15:04:00.731209 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:04:01.230989728 +0000 UTC m=+99.673452338 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.732027 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mh5wg"] Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.732243 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tfnr6" Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.733273 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f56ea937-8ed8-4912-b721-2ed83a15fa2b-utilities\") pod \"certified-operators-t2pmf\" (UID: \"f56ea937-8ed8-4912-b721-2ed83a15fa2b\") " pod="openshift-marketplace/certified-operators-t2pmf" Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.777271 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kw68\" (UniqueName: \"kubernetes.io/projected/f56ea937-8ed8-4912-b721-2ed83a15fa2b-kube-api-access-4kw68\") pod \"certified-operators-t2pmf\" (UID: \"f56ea937-8ed8-4912-b721-2ed83a15fa2b\") " pod="openshift-marketplace/certified-operators-t2pmf" Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.803433 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t2pmf" Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.831890 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68222e1b-b4d5-4e64-853e-a8af2ad8383a-catalog-content\") pod \"community-operators-mh5wg\" (UID: \"68222e1b-b4d5-4e64-853e-a8af2ad8383a\") " pod="openshift-marketplace/community-operators-mh5wg" Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.832240 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.832288 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68222e1b-b4d5-4e64-853e-a8af2ad8383a-utilities\") pod \"community-operators-mh5wg\" (UID: \"68222e1b-b4d5-4e64-853e-a8af2ad8383a\") " pod="openshift-marketplace/community-operators-mh5wg" Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.832309 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45v9w\" (UniqueName: \"kubernetes.io/projected/68222e1b-b4d5-4e64-853e-a8af2ad8383a-kube-api-access-45v9w\") pod \"community-operators-mh5wg\" (UID: \"68222e1b-b4d5-4e64-853e-a8af2ad8383a\") " pod="openshift-marketplace/community-operators-mh5wg" Nov 25 15:04:00 crc kubenswrapper[4890]: E1125 15:04:00.832588 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:04:01.332576877 +0000 UTC m=+99.775039487 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.938978 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.939352 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68222e1b-b4d5-4e64-853e-a8af2ad8383a-catalog-content\") pod \"community-operators-mh5wg\" (UID: \"68222e1b-b4d5-4e64-853e-a8af2ad8383a\") " pod="openshift-marketplace/community-operators-mh5wg" Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.939465 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68222e1b-b4d5-4e64-853e-a8af2ad8383a-utilities\") pod \"community-operators-mh5wg\" (UID: \"68222e1b-b4d5-4e64-853e-a8af2ad8383a\") " pod="openshift-marketplace/community-operators-mh5wg" Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.939498 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45v9w\" (UniqueName: \"kubernetes.io/projected/68222e1b-b4d5-4e64-853e-a8af2ad8383a-kube-api-access-45v9w\") pod \"community-operators-mh5wg\" (UID: \"68222e1b-b4d5-4e64-853e-a8af2ad8383a\") " pod="openshift-marketplace/community-operators-mh5wg" Nov 25 15:04:00 crc kubenswrapper[4890]: E1125 15:04:00.940201 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:04:01.440180642 +0000 UTC m=+99.882643252 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.940663 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68222e1b-b4d5-4e64-853e-a8af2ad8383a-catalog-content\") pod \"community-operators-mh5wg\" (UID: \"68222e1b-b4d5-4e64-853e-a8af2ad8383a\") " pod="openshift-marketplace/community-operators-mh5wg" Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.940878 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68222e1b-b4d5-4e64-853e-a8af2ad8383a-utilities\") pod \"community-operators-mh5wg\" (UID: \"68222e1b-b4d5-4e64-853e-a8af2ad8383a\") " pod="openshift-marketplace/community-operators-mh5wg" Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.944961 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-2nlkp"] Nov 25 15:04:00 crc kubenswrapper[4890]: W1125 15:04:00.975406 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod16f52653_c067_4186_bc40_9c0f7c383b64.slice/crio-7be08db81b8dc201460205d92e620c12c0568b62127209f0652ade362bb6c914 WatchSource:0}: Error finding container 7be08db81b8dc201460205d92e620c12c0568b62127209f0652ade362bb6c914: Status 404 returned error can't find the container with id 7be08db81b8dc201460205d92e620c12c0568b62127209f0652ade362bb6c914 Nov 25 15:04:00 crc kubenswrapper[4890]: I1125 15:04:00.990598 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45v9w\" (UniqueName: \"kubernetes.io/projected/68222e1b-b4d5-4e64-853e-a8af2ad8383a-kube-api-access-45v9w\") pod \"community-operators-mh5wg\" (UID: \"68222e1b-b4d5-4e64-853e-a8af2ad8383a\") " pod="openshift-marketplace/community-operators-mh5wg" Nov 25 15:04:01 crc kubenswrapper[4890]: I1125 15:04:01.041720 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:04:01 crc kubenswrapper[4890]: E1125 15:04:01.042065 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:04:01.542050829 +0000 UTC m=+99.984513439 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:04:01 crc kubenswrapper[4890]: I1125 15:04:01.117968 4890 patch_prober.go:28] interesting pod/router-default-5444994796-kxfxn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 15:04:01 crc kubenswrapper[4890]: [-]has-synced failed: reason withheld Nov 25 15:04:01 crc kubenswrapper[4890]: [+]process-running ok Nov 25 15:04:01 crc kubenswrapper[4890]: healthz check failed Nov 25 15:04:01 crc kubenswrapper[4890]: I1125 15:04:01.118020 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kxfxn" podUID="d153ec32-0e4e-41d6-b105-38f91ab40c14" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 15:04:01 crc kubenswrapper[4890]: I1125 15:04:01.143336 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mh5wg" Nov 25 15:04:01 crc kubenswrapper[4890]: I1125 15:04:01.143584 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:04:01 crc kubenswrapper[4890]: E1125 15:04:01.143940 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:04:01.643886515 +0000 UTC m=+100.086349125 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:04:01 crc kubenswrapper[4890]: I1125 15:04:01.144045 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:04:01 crc kubenswrapper[4890]: E1125 15:04:01.144295 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:04:01.644287606 +0000 UTC m=+100.086750206 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:04:01 crc kubenswrapper[4890]: I1125 15:04:01.225629 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-t2pmf"] Nov 25 15:04:01 crc kubenswrapper[4890]: I1125 15:04:01.249756 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-z4b8p"] Nov 25 15:04:01 crc kubenswrapper[4890]: I1125 15:04:01.253763 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:04:01 crc kubenswrapper[4890]: E1125 15:04:01.254188 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:04:01.754172172 +0000 UTC m=+100.196634782 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:04:01 crc kubenswrapper[4890]: I1125 15:04:01.319761 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tfnr6"] Nov 25 15:04:01 crc kubenswrapper[4890]: I1125 15:04:01.355378 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:04:01 crc kubenswrapper[4890]: E1125 15:04:01.355993 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:04:01.855945637 +0000 UTC m=+100.298408307 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:04:01 crc kubenswrapper[4890]: W1125 15:04:01.360545 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f052eff_1a42_4c20_8644_d575fc557060.slice/crio-a962d0961c21b943f81dba768a61952a6f35aad61ced29a057a068240353946d WatchSource:0}: Error finding container a962d0961c21b943f81dba768a61952a6f35aad61ced29a057a068240353946d: Status 404 returned error can't find the container with id a962d0961c21b943f81dba768a61952a6f35aad61ced29a057a068240353946d Nov 25 15:04:01 crc kubenswrapper[4890]: I1125 15:04:01.420650 4890 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Nov 25 15:04:01 crc kubenswrapper[4890]: I1125 15:04:01.456890 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:04:01 crc kubenswrapper[4890]: E1125 15:04:01.457433 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:04:01.957386342 +0000 UTC m=+100.399848952 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:04:01 crc kubenswrapper[4890]: I1125 15:04:01.466475 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:04:01 crc kubenswrapper[4890]: E1125 15:04:01.466972 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:04:01.966957794 +0000 UTC m=+100.409420404 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:04:01 crc kubenswrapper[4890]: I1125 15:04:01.560330 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-2nlkp" event={"ID":"16f52653-c067-4186-bc40-9c0f7c383b64","Type":"ContainerStarted","Data":"cae223ba5b4dce15ef06fb38166e7cd39420c82a5a74e738e818a2828e99c098"} Nov 25 15:04:01 crc kubenswrapper[4890]: I1125 15:04:01.560414 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-2nlkp" event={"ID":"16f52653-c067-4186-bc40-9c0f7c383b64","Type":"ContainerStarted","Data":"7be08db81b8dc201460205d92e620c12c0568b62127209f0652ade362bb6c914"} Nov 25 15:04:01 crc kubenswrapper[4890]: I1125 15:04:01.567772 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:04:01 crc kubenswrapper[4890]: E1125 15:04:01.568134 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:04:02.068117642 +0000 UTC m=+100.510580252 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:04:01 crc kubenswrapper[4890]: I1125 15:04:01.574715 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tfnr6" event={"ID":"8f052eff-1a42-4c20-8644-d575fc557060","Type":"ContainerStarted","Data":"a962d0961c21b943f81dba768a61952a6f35aad61ced29a057a068240353946d"} Nov 25 15:04:01 crc kubenswrapper[4890]: I1125 15:04:01.595945 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-cmpbm" event={"ID":"ba947c9c-cd84-4c9b-a431-9276412058d1","Type":"ContainerStarted","Data":"8704f5a202e605e5bc80946dd4f8cbbadce00db01d9369ce826c801fc3a6a86a"} Nov 25 15:04:01 crc kubenswrapper[4890]: I1125 15:04:01.596028 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-cmpbm" event={"ID":"ba947c9c-cd84-4c9b-a431-9276412058d1","Type":"ContainerStarted","Data":"7c08d8b58d88dca7104ba76fe9b0680abd17115a2eb5b81472c08100e8b7d6d3"} Nov 25 15:04:01 crc kubenswrapper[4890]: I1125 15:04:01.613690 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t2pmf" event={"ID":"f56ea937-8ed8-4912-b721-2ed83a15fa2b","Type":"ContainerStarted","Data":"4d40823fb1444992ffb52d2e52c051f274cbaa484f86cc7e3b6e9e8163c43eac"} Nov 25 15:04:01 crc kubenswrapper[4890]: I1125 15:04:01.620578 4890 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 15:04:01 crc kubenswrapper[4890]: I1125 15:04:01.629284 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mh5wg"] Nov 25 15:04:01 crc kubenswrapper[4890]: I1125 15:04:01.632375 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z4b8p" event={"ID":"b1983120-f62d-46b0-833e-db30a81973c1","Type":"ContainerStarted","Data":"8f7dc004940b2f5161509321d60d64ff691d261bd46b885a6de2a8b2d887d5a8"} Nov 25 15:04:01 crc kubenswrapper[4890]: I1125 15:04:01.646796 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-pshcp" Nov 25 15:04:01 crc kubenswrapper[4890]: I1125 15:04:01.669366 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:04:01 crc kubenswrapper[4890]: E1125 15:04:01.669944 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:04:02.169917957 +0000 UTC m=+100.612380567 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:04:01 crc kubenswrapper[4890]: W1125 15:04:01.714969 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod68222e1b_b4d5_4e64_853e_a8af2ad8383a.slice/crio-f01695da1108260f00d78b82bd23de7fe506f51da8d3ff7e5a930f0614864644 WatchSource:0}: Error finding container f01695da1108260f00d78b82bd23de7fe506f51da8d3ff7e5a930f0614864644: Status 404 returned error can't find the container with id f01695da1108260f00d78b82bd23de7fe506f51da8d3ff7e5a930f0614864644 Nov 25 15:04:01 crc kubenswrapper[4890]: I1125 15:04:01.770426 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:04:01 crc kubenswrapper[4890]: E1125 15:04:01.770571 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:04:02.270541081 +0000 UTC m=+100.713003691 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:04:01 crc kubenswrapper[4890]: I1125 15:04:01.773854 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:04:01 crc kubenswrapper[4890]: E1125 15:04:01.775243 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:04:02.275228359 +0000 UTC m=+100.717690969 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:04:01 crc kubenswrapper[4890]: I1125 15:04:01.875435 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:04:01 crc kubenswrapper[4890]: E1125 15:04:01.875677 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:04:02.375642206 +0000 UTC m=+100.818104826 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:04:01 crc kubenswrapper[4890]: I1125 15:04:01.875937 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:04:01 crc kubenswrapper[4890]: E1125 15:04:01.876276 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:04:02.376263243 +0000 UTC m=+100.818725853 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:04:01 crc kubenswrapper[4890]: I1125 15:04:01.976932 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:04:01 crc kubenswrapper[4890]: E1125 15:04:01.977182 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:04:02.477125093 +0000 UTC m=+100.919587703 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:04:01 crc kubenswrapper[4890]: I1125 15:04:01.977566 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:04:01 crc kubenswrapper[4890]: E1125 15:04:01.977926 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:04:02.477916574 +0000 UTC m=+100.920379254 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.078580 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:04:02 crc kubenswrapper[4890]: E1125 15:04:02.078758 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:04:02.578731333 +0000 UTC m=+101.021193943 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.078968 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:04:02 crc kubenswrapper[4890]: E1125 15:04:02.079462 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:04:02.579450862 +0000 UTC m=+101.021913472 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.109420 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.111430 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.119242 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.119492 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.126804 4890 patch_prober.go:28] interesting pod/router-default-5444994796-kxfxn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 15:04:02 crc kubenswrapper[4890]: [-]has-synced failed: reason withheld Nov 25 15:04:02 crc kubenswrapper[4890]: [+]process-running ok Nov 25 15:04:02 crc kubenswrapper[4890]: healthz check failed Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.126907 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kxfxn" podUID="d153ec32-0e4e-41d6-b105-38f91ab40c14" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.133316 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.179754 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:04:02 crc kubenswrapper[4890]: E1125 15:04:02.179919 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:04:02.679897671 +0000 UTC m=+101.122360281 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.180414 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:04:02 crc kubenswrapper[4890]: E1125 15:04:02.180886 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 15:04:02.680876098 +0000 UTC m=+101.123338708 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vkrx2" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.251815 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hn24q"] Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.253146 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hn24q" Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.267770 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.272982 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hn24q"] Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.281069 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.281421 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4cc12bd5-1c51-42fe-b534-5158f7fe27d7-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"4cc12bd5-1c51-42fe-b534-5158f7fe27d7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 15:04:02 crc kubenswrapper[4890]: E1125 15:04:02.281603 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 15:04:02.781587453 +0000 UTC m=+101.224050063 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.281664 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4cc12bd5-1c51-42fe-b534-5158f7fe27d7-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"4cc12bd5-1c51-42fe-b534-5158f7fe27d7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.340798 4890 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-11-25T15:04:01.420680008Z","Handler":null,"Name":""} Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.347625 4890 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.347707 4890 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.383453 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.383518 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fbb3062-51d8-4fc0-9b25-5336de75d692-catalog-content\") pod \"redhat-marketplace-hn24q\" (UID: \"7fbb3062-51d8-4fc0-9b25-5336de75d692\") " pod="openshift-marketplace/redhat-marketplace-hn24q" Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.383577 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fbb3062-51d8-4fc0-9b25-5336de75d692-utilities\") pod \"redhat-marketplace-hn24q\" (UID: \"7fbb3062-51d8-4fc0-9b25-5336de75d692\") " pod="openshift-marketplace/redhat-marketplace-hn24q" Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.383608 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4cc12bd5-1c51-42fe-b534-5158f7fe27d7-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"4cc12bd5-1c51-42fe-b534-5158f7fe27d7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.383662 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkwzk\" (UniqueName: \"kubernetes.io/projected/7fbb3062-51d8-4fc0-9b25-5336de75d692-kube-api-access-mkwzk\") pod \"redhat-marketplace-hn24q\" (UID: \"7fbb3062-51d8-4fc0-9b25-5336de75d692\") " pod="openshift-marketplace/redhat-marketplace-hn24q" Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.383700 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4cc12bd5-1c51-42fe-b534-5158f7fe27d7-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"4cc12bd5-1c51-42fe-b534-5158f7fe27d7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.383781 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4cc12bd5-1c51-42fe-b534-5158f7fe27d7-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"4cc12bd5-1c51-42fe-b534-5158f7fe27d7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.389437 4890 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.389523 4890 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.407413 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4cc12bd5-1c51-42fe-b534-5158f7fe27d7-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"4cc12bd5-1c51-42fe-b534-5158f7fe27d7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.429436 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vkrx2\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.441995 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.484475 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.484844 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fbb3062-51d8-4fc0-9b25-5336de75d692-catalog-content\") pod \"redhat-marketplace-hn24q\" (UID: \"7fbb3062-51d8-4fc0-9b25-5336de75d692\") " pod="openshift-marketplace/redhat-marketplace-hn24q" Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.484919 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fbb3062-51d8-4fc0-9b25-5336de75d692-utilities\") pod \"redhat-marketplace-hn24q\" (UID: \"7fbb3062-51d8-4fc0-9b25-5336de75d692\") " pod="openshift-marketplace/redhat-marketplace-hn24q" Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.485145 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkwzk\" (UniqueName: \"kubernetes.io/projected/7fbb3062-51d8-4fc0-9b25-5336de75d692-kube-api-access-mkwzk\") pod \"redhat-marketplace-hn24q\" (UID: \"7fbb3062-51d8-4fc0-9b25-5336de75d692\") " pod="openshift-marketplace/redhat-marketplace-hn24q" Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.504771 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fbb3062-51d8-4fc0-9b25-5336de75d692-catalog-content\") pod \"redhat-marketplace-hn24q\" (UID: \"7fbb3062-51d8-4fc0-9b25-5336de75d692\") " pod="openshift-marketplace/redhat-marketplace-hn24q" Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.505835 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fbb3062-51d8-4fc0-9b25-5336de75d692-utilities\") pod \"redhat-marketplace-hn24q\" (UID: \"7fbb3062-51d8-4fc0-9b25-5336de75d692\") " pod="openshift-marketplace/redhat-marketplace-hn24q" Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.507612 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.516571 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkwzk\" (UniqueName: \"kubernetes.io/projected/7fbb3062-51d8-4fc0-9b25-5336de75d692-kube-api-access-mkwzk\") pod \"redhat-marketplace-hn24q\" (UID: \"7fbb3062-51d8-4fc0-9b25-5336de75d692\") " pod="openshift-marketplace/redhat-marketplace-hn24q" Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.583173 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hn24q" Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.655360 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zt6cj"] Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.656561 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zt6cj" Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.665893 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.669242 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zt6cj"] Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.671126 4890 generic.go:334] "Generic (PLEG): container finished" podID="f56ea937-8ed8-4912-b721-2ed83a15fa2b" containerID="f284f9fe799e44084a0da77f9d43c13d505cd30ee11d0ea81b9615874110ad1e" exitCode=0 Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.671183 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t2pmf" event={"ID":"f56ea937-8ed8-4912-b721-2ed83a15fa2b","Type":"ContainerDied","Data":"f284f9fe799e44084a0da77f9d43c13d505cd30ee11d0ea81b9615874110ad1e"} Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.673852 4890 generic.go:334] "Generic (PLEG): container finished" podID="b1983120-f62d-46b0-833e-db30a81973c1" containerID="0ed3bd68e3b6f9f33a2a5c10e9c36a0b8bb1426015599ef33853a89b7932a57a" exitCode=0 Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.673924 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z4b8p" event={"ID":"b1983120-f62d-46b0-833e-db30a81973c1","Type":"ContainerDied","Data":"0ed3bd68e3b6f9f33a2a5c10e9c36a0b8bb1426015599ef33853a89b7932a57a"} Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.726911 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.732896 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-2nlkp" event={"ID":"16f52653-c067-4186-bc40-9c0f7c383b64","Type":"ContainerStarted","Data":"6bfe35174ac25c7446bb6d50530b0b39bba53fef7be88fc4439f007c825bd1f3"} Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.736383 4890 generic.go:334] "Generic (PLEG): container finished" podID="8f052eff-1a42-4c20-8644-d575fc557060" containerID="9c61813686fa358cdacad333b64a1c4530ac05fa7b48bcdd7491621c9e449f32" exitCode=0 Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.736450 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tfnr6" event={"ID":"8f052eff-1a42-4c20-8644-d575fc557060","Type":"ContainerDied","Data":"9c61813686fa358cdacad333b64a1c4530ac05fa7b48bcdd7491621c9e449f32"} Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.745281 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-cmpbm" event={"ID":"ba947c9c-cd84-4c9b-a431-9276412058d1","Type":"ContainerStarted","Data":"891df54a743fa66c2ca24043940571d3a4d9f47e5014f315968e4d5032db8f38"} Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.750214 4890 generic.go:334] "Generic (PLEG): container finished" podID="68222e1b-b4d5-4e64-853e-a8af2ad8383a" containerID="26be49e805a1db93198aba9739155a7a7e942b9401dbf28ae551cba63c892244" exitCode=0 Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.750411 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mh5wg" event={"ID":"68222e1b-b4d5-4e64-853e-a8af2ad8383a","Type":"ContainerDied","Data":"26be49e805a1db93198aba9739155a7a7e942b9401dbf28ae551cba63c892244"} Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.750462 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mh5wg" event={"ID":"68222e1b-b4d5-4e64-853e-a8af2ad8383a","Type":"ContainerStarted","Data":"f01695da1108260f00d78b82bd23de7fe506f51da8d3ff7e5a930f0614864644"} Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.762843 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-2nlkp" podStartSLOduration=81.76282352 podStartE2EDuration="1m21.76282352s" podCreationTimestamp="2025-11-25 15:02:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:04:02.753356791 +0000 UTC m=+101.195819401" watchObservedRunningTime="2025-11-25 15:04:02.76282352 +0000 UTC m=+101.205286130" Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.790976 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b3f25c9-5630-4623-8ee4-49e52c356530-catalog-content\") pod \"redhat-marketplace-zt6cj\" (UID: \"1b3f25c9-5630-4623-8ee4-49e52c356530\") " pod="openshift-marketplace/redhat-marketplace-zt6cj" Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.794535 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8799j\" (UniqueName: \"kubernetes.io/projected/1b3f25c9-5630-4623-8ee4-49e52c356530-kube-api-access-8799j\") pod \"redhat-marketplace-zt6cj\" (UID: \"1b3f25c9-5630-4623-8ee4-49e52c356530\") " pod="openshift-marketplace/redhat-marketplace-zt6cj" Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.794664 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b3f25c9-5630-4623-8ee4-49e52c356530-utilities\") pod \"redhat-marketplace-zt6cj\" (UID: \"1b3f25c9-5630-4623-8ee4-49e52c356530\") " pod="openshift-marketplace/redhat-marketplace-zt6cj" Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.827783 4890 generic.go:334] "Generic (PLEG): container finished" podID="4622b25d-fb03-4e9f-a66c-7f90def9b5f7" containerID="d69aed072dd6ab3043fb1ed07aa7e23803f193be0d9017879940b070a3669203" exitCode=0 Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.828134 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401380-96zpb" event={"ID":"4622b25d-fb03-4e9f-a66c-7f90def9b5f7","Type":"ContainerDied","Data":"d69aed072dd6ab3043fb1ed07aa7e23803f193be0d9017879940b070a3669203"} Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.843044 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-cmpbm" podStartSLOduration=11.843011894 podStartE2EDuration="11.843011894s" podCreationTimestamp="2025-11-25 15:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:04:02.8008574 +0000 UTC m=+101.243320030" watchObservedRunningTime="2025-11-25 15:04:02.843011894 +0000 UTC m=+101.285474504" Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.899133 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b3f25c9-5630-4623-8ee4-49e52c356530-catalog-content\") pod \"redhat-marketplace-zt6cj\" (UID: \"1b3f25c9-5630-4623-8ee4-49e52c356530\") " pod="openshift-marketplace/redhat-marketplace-zt6cj" Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.899895 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8799j\" (UniqueName: \"kubernetes.io/projected/1b3f25c9-5630-4623-8ee4-49e52c356530-kube-api-access-8799j\") pod \"redhat-marketplace-zt6cj\" (UID: \"1b3f25c9-5630-4623-8ee4-49e52c356530\") " pod="openshift-marketplace/redhat-marketplace-zt6cj" Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.900062 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b3f25c9-5630-4623-8ee4-49e52c356530-utilities\") pod \"redhat-marketplace-zt6cj\" (UID: \"1b3f25c9-5630-4623-8ee4-49e52c356530\") " pod="openshift-marketplace/redhat-marketplace-zt6cj" Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.900333 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b3f25c9-5630-4623-8ee4-49e52c356530-catalog-content\") pod \"redhat-marketplace-zt6cj\" (UID: \"1b3f25c9-5630-4623-8ee4-49e52c356530\") " pod="openshift-marketplace/redhat-marketplace-zt6cj" Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.902516 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b3f25c9-5630-4623-8ee4-49e52c356530-utilities\") pod \"redhat-marketplace-zt6cj\" (UID: \"1b3f25c9-5630-4623-8ee4-49e52c356530\") " pod="openshift-marketplace/redhat-marketplace-zt6cj" Nov 25 15:04:02 crc kubenswrapper[4890]: I1125 15:04:02.970966 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8799j\" (UniqueName: \"kubernetes.io/projected/1b3f25c9-5630-4623-8ee4-49e52c356530-kube-api-access-8799j\") pod \"redhat-marketplace-zt6cj\" (UID: \"1b3f25c9-5630-4623-8ee4-49e52c356530\") " pod="openshift-marketplace/redhat-marketplace-zt6cj" Nov 25 15:04:03 crc kubenswrapper[4890]: I1125 15:04:03.027864 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zt6cj" Nov 25 15:04:03 crc kubenswrapper[4890]: I1125 15:04:03.082085 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-vkrx2"] Nov 25 15:04:03 crc kubenswrapper[4890]: W1125 15:04:03.098554 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod96800ba3_0ee4_42e0_b8f7_316ef57a6173.slice/crio-23eb54f5a723ab1789e822b18a86de7fdb10770fc00dcbacfd094343feabd298 WatchSource:0}: Error finding container 23eb54f5a723ab1789e822b18a86de7fdb10770fc00dcbacfd094343feabd298: Status 404 returned error can't find the container with id 23eb54f5a723ab1789e822b18a86de7fdb10770fc00dcbacfd094343feabd298 Nov 25 15:04:03 crc kubenswrapper[4890]: I1125 15:04:03.118209 4890 patch_prober.go:28] interesting pod/router-default-5444994796-kxfxn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 15:04:03 crc kubenswrapper[4890]: [-]has-synced failed: reason withheld Nov 25 15:04:03 crc kubenswrapper[4890]: [+]process-running ok Nov 25 15:04:03 crc kubenswrapper[4890]: healthz check failed Nov 25 15:04:03 crc kubenswrapper[4890]: I1125 15:04:03.118283 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kxfxn" podUID="d153ec32-0e4e-41d6-b105-38f91ab40c14" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 15:04:03 crc kubenswrapper[4890]: I1125 15:04:03.171004 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hn24q"] Nov 25 15:04:03 crc kubenswrapper[4890]: W1125 15:04:03.188229 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7fbb3062_51d8_4fc0_9b25_5336de75d692.slice/crio-7f868b7bd6132c1baff4a2bc246ad9bfbb5e1054ca25426be1b0e889f197a1e6 WatchSource:0}: Error finding container 7f868b7bd6132c1baff4a2bc246ad9bfbb5e1054ca25426be1b0e889f197a1e6: Status 404 returned error can't find the container with id 7f868b7bd6132c1baff4a2bc246ad9bfbb5e1054ca25426be1b0e889f197a1e6 Nov 25 15:04:03 crc kubenswrapper[4890]: I1125 15:04:03.216744 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-t29fd" Nov 25 15:04:03 crc kubenswrapper[4890]: I1125 15:04:03.216825 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-t29fd" Nov 25 15:04:03 crc kubenswrapper[4890]: I1125 15:04:03.221656 4890 patch_prober.go:28] interesting pod/console-f9d7485db-t29fd container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.8:8443/health\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Nov 25 15:04:03 crc kubenswrapper[4890]: I1125 15:04:03.221716 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-t29fd" podUID="f748296c-d668-46bd-bcb7-4eb21822a244" containerName="console" probeResult="failure" output="Get \"https://10.217.0.8:8443/health\": dial tcp 10.217.0.8:8443: connect: connection refused" Nov 25 15:04:03 crc kubenswrapper[4890]: I1125 15:04:03.252565 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vf44g"] Nov 25 15:04:03 crc kubenswrapper[4890]: I1125 15:04:03.257010 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vf44g" Nov 25 15:04:03 crc kubenswrapper[4890]: I1125 15:04:03.264474 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 25 15:04:03 crc kubenswrapper[4890]: I1125 15:04:03.267026 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vf44g"] Nov 25 15:04:03 crc kubenswrapper[4890]: I1125 15:04:03.269072 4890 patch_prober.go:28] interesting pod/downloads-7954f5f757-8bln9 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Nov 25 15:04:03 crc kubenswrapper[4890]: I1125 15:04:03.269075 4890 patch_prober.go:28] interesting pod/downloads-7954f5f757-8bln9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Nov 25 15:04:03 crc kubenswrapper[4890]: I1125 15:04:03.269119 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-8bln9" podUID="89bebf40-d718-4168-8377-4fa334afaa17" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Nov 25 15:04:03 crc kubenswrapper[4890]: I1125 15:04:03.269130 4890 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-8bln9" podUID="89bebf40-d718-4168-8377-4fa334afaa17" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Nov 25 15:04:03 crc kubenswrapper[4890]: I1125 15:04:03.406945 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmphq\" (UniqueName: \"kubernetes.io/projected/22b61785-937e-4562-8c48-5e3d390f3389-kube-api-access-vmphq\") pod \"redhat-operators-vf44g\" (UID: \"22b61785-937e-4562-8c48-5e3d390f3389\") " pod="openshift-marketplace/redhat-operators-vf44g" Nov 25 15:04:03 crc kubenswrapper[4890]: I1125 15:04:03.407059 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22b61785-937e-4562-8c48-5e3d390f3389-catalog-content\") pod \"redhat-operators-vf44g\" (UID: \"22b61785-937e-4562-8c48-5e3d390f3389\") " pod="openshift-marketplace/redhat-operators-vf44g" Nov 25 15:04:03 crc kubenswrapper[4890]: I1125 15:04:03.407087 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22b61785-937e-4562-8c48-5e3d390f3389-utilities\") pod \"redhat-operators-vf44g\" (UID: \"22b61785-937e-4562-8c48-5e3d390f3389\") " pod="openshift-marketplace/redhat-operators-vf44g" Nov 25 15:04:03 crc kubenswrapper[4890]: I1125 15:04:03.493136 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-glmvm" Nov 25 15:04:03 crc kubenswrapper[4890]: I1125 15:04:03.499209 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-glmvm" Nov 25 15:04:03 crc kubenswrapper[4890]: I1125 15:04:03.507931 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmphq\" (UniqueName: \"kubernetes.io/projected/22b61785-937e-4562-8c48-5e3d390f3389-kube-api-access-vmphq\") pod \"redhat-operators-vf44g\" (UID: \"22b61785-937e-4562-8c48-5e3d390f3389\") " pod="openshift-marketplace/redhat-operators-vf44g" Nov 25 15:04:03 crc kubenswrapper[4890]: I1125 15:04:03.507986 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22b61785-937e-4562-8c48-5e3d390f3389-catalog-content\") pod \"redhat-operators-vf44g\" (UID: \"22b61785-937e-4562-8c48-5e3d390f3389\") " pod="openshift-marketplace/redhat-operators-vf44g" Nov 25 15:04:03 crc kubenswrapper[4890]: I1125 15:04:03.508008 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22b61785-937e-4562-8c48-5e3d390f3389-utilities\") pod \"redhat-operators-vf44g\" (UID: \"22b61785-937e-4562-8c48-5e3d390f3389\") " pod="openshift-marketplace/redhat-operators-vf44g" Nov 25 15:04:03 crc kubenswrapper[4890]: I1125 15:04:03.550597 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmphq\" (UniqueName: \"kubernetes.io/projected/22b61785-937e-4562-8c48-5e3d390f3389-kube-api-access-vmphq\") pod \"redhat-operators-vf44g\" (UID: \"22b61785-937e-4562-8c48-5e3d390f3389\") " pod="openshift-marketplace/redhat-operators-vf44g" Nov 25 15:04:03 crc kubenswrapper[4890]: I1125 15:04:03.652571 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kq8dw"] Nov 25 15:04:03 crc kubenswrapper[4890]: I1125 15:04:03.653975 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kq8dw" Nov 25 15:04:03 crc kubenswrapper[4890]: I1125 15:04:03.812294 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f0bd74d-3205-422a-8c40-28ae9d29dc3a-catalog-content\") pod \"redhat-operators-kq8dw\" (UID: \"2f0bd74d-3205-422a-8c40-28ae9d29dc3a\") " pod="openshift-marketplace/redhat-operators-kq8dw" Nov 25 15:04:03 crc kubenswrapper[4890]: I1125 15:04:03.812768 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjwk9\" (UniqueName: \"kubernetes.io/projected/2f0bd74d-3205-422a-8c40-28ae9d29dc3a-kube-api-access-jjwk9\") pod \"redhat-operators-kq8dw\" (UID: \"2f0bd74d-3205-422a-8c40-28ae9d29dc3a\") " pod="openshift-marketplace/redhat-operators-kq8dw" Nov 25 15:04:03 crc kubenswrapper[4890]: I1125 15:04:03.812812 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f0bd74d-3205-422a-8c40-28ae9d29dc3a-utilities\") pod \"redhat-operators-kq8dw\" (UID: \"2f0bd74d-3205-422a-8c40-28ae9d29dc3a\") " pod="openshift-marketplace/redhat-operators-kq8dw" Nov 25 15:04:03 crc kubenswrapper[4890]: I1125 15:04:03.831365 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22b61785-937e-4562-8c48-5e3d390f3389-utilities\") pod \"redhat-operators-vf44g\" (UID: \"22b61785-937e-4562-8c48-5e3d390f3389\") " pod="openshift-marketplace/redhat-operators-vf44g" Nov 25 15:04:03 crc kubenswrapper[4890]: I1125 15:04:03.834191 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22b61785-937e-4562-8c48-5e3d390f3389-catalog-content\") pod \"redhat-operators-vf44g\" (UID: \"22b61785-937e-4562-8c48-5e3d390f3389\") " pod="openshift-marketplace/redhat-operators-vf44g" Nov 25 15:04:03 crc kubenswrapper[4890]: I1125 15:04:03.840992 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kq8dw"] Nov 25 15:04:03 crc kubenswrapper[4890]: I1125 15:04:03.862620 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zt6cj"] Nov 25 15:04:03 crc kubenswrapper[4890]: I1125 15:04:03.870685 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" event={"ID":"96800ba3-0ee4-42e0-b8f7-316ef57a6173","Type":"ContainerStarted","Data":"23eb54f5a723ab1789e822b18a86de7fdb10770fc00dcbacfd094343feabd298"} Nov 25 15:04:03 crc kubenswrapper[4890]: I1125 15:04:03.876220 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"4cc12bd5-1c51-42fe-b534-5158f7fe27d7","Type":"ContainerStarted","Data":"15e32b819108d5c06cbe745e0a76969332da425b8794ff4171e587d25834b8bd"} Nov 25 15:04:03 crc kubenswrapper[4890]: I1125 15:04:03.877997 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hn24q" event={"ID":"7fbb3062-51d8-4fc0-9b25-5336de75d692","Type":"ContainerStarted","Data":"7f868b7bd6132c1baff4a2bc246ad9bfbb5e1054ca25426be1b0e889f197a1e6"} Nov 25 15:04:03 crc kubenswrapper[4890]: I1125 15:04:03.885482 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vf44g" Nov 25 15:04:03 crc kubenswrapper[4890]: I1125 15:04:03.928554 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f0bd74d-3205-422a-8c40-28ae9d29dc3a-catalog-content\") pod \"redhat-operators-kq8dw\" (UID: \"2f0bd74d-3205-422a-8c40-28ae9d29dc3a\") " pod="openshift-marketplace/redhat-operators-kq8dw" Nov 25 15:04:03 crc kubenswrapper[4890]: I1125 15:04:03.928694 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjwk9\" (UniqueName: \"kubernetes.io/projected/2f0bd74d-3205-422a-8c40-28ae9d29dc3a-kube-api-access-jjwk9\") pod \"redhat-operators-kq8dw\" (UID: \"2f0bd74d-3205-422a-8c40-28ae9d29dc3a\") " pod="openshift-marketplace/redhat-operators-kq8dw" Nov 25 15:04:03 crc kubenswrapper[4890]: I1125 15:04:03.928734 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f0bd74d-3205-422a-8c40-28ae9d29dc3a-utilities\") pod \"redhat-operators-kq8dw\" (UID: \"2f0bd74d-3205-422a-8c40-28ae9d29dc3a\") " pod="openshift-marketplace/redhat-operators-kq8dw" Nov 25 15:04:03 crc kubenswrapper[4890]: I1125 15:04:03.929504 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f0bd74d-3205-422a-8c40-28ae9d29dc3a-utilities\") pod \"redhat-operators-kq8dw\" (UID: \"2f0bd74d-3205-422a-8c40-28ae9d29dc3a\") " pod="openshift-marketplace/redhat-operators-kq8dw" Nov 25 15:04:03 crc kubenswrapper[4890]: I1125 15:04:03.938590 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f0bd74d-3205-422a-8c40-28ae9d29dc3a-catalog-content\") pod \"redhat-operators-kq8dw\" (UID: \"2f0bd74d-3205-422a-8c40-28ae9d29dc3a\") " pod="openshift-marketplace/redhat-operators-kq8dw" Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.038943 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjwk9\" (UniqueName: \"kubernetes.io/projected/2f0bd74d-3205-422a-8c40-28ae9d29dc3a-kube-api-access-jjwk9\") pod \"redhat-operators-kq8dw\" (UID: \"2f0bd74d-3205-422a-8c40-28ae9d29dc3a\") " pod="openshift-marketplace/redhat-operators-kq8dw" Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.054268 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8zql" Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.054621 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8zql" Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.075371 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8zql" Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.111511 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.114018 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-kxfxn" Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.135401 4890 patch_prober.go:28] interesting pod/router-default-5444994796-kxfxn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 15:04:04 crc kubenswrapper[4890]: [-]has-synced failed: reason withheld Nov 25 15:04:04 crc kubenswrapper[4890]: [+]process-running ok Nov 25 15:04:04 crc kubenswrapper[4890]: healthz check failed Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.135463 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kxfxn" podUID="d153ec32-0e4e-41d6-b105-38f91ab40c14" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.207198 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.266795 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kq8dw" Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.437733 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401380-96zpb" Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.489825 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vf44g"] Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.510467 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nfp7k" Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.511852 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 25 15:04:04 crc kubenswrapper[4890]: E1125 15:04:04.512056 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4622b25d-fb03-4e9f-a66c-7f90def9b5f7" containerName="collect-profiles" Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.512072 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="4622b25d-fb03-4e9f-a66c-7f90def9b5f7" containerName="collect-profiles" Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.512187 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="4622b25d-fb03-4e9f-a66c-7f90def9b5f7" containerName="collect-profiles" Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.513729 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.518840 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.519335 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.519810 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-xmhq9" Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.528355 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.551621 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4622b25d-fb03-4e9f-a66c-7f90def9b5f7-secret-volume\") pod \"4622b25d-fb03-4e9f-a66c-7f90def9b5f7\" (UID: \"4622b25d-fb03-4e9f-a66c-7f90def9b5f7\") " Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.551679 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4622b25d-fb03-4e9f-a66c-7f90def9b5f7-config-volume\") pod \"4622b25d-fb03-4e9f-a66c-7f90def9b5f7\" (UID: \"4622b25d-fb03-4e9f-a66c-7f90def9b5f7\") " Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.552793 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qjjcv\" (UniqueName: \"kubernetes.io/projected/4622b25d-fb03-4e9f-a66c-7f90def9b5f7-kube-api-access-qjjcv\") pod \"4622b25d-fb03-4e9f-a66c-7f90def9b5f7\" (UID: \"4622b25d-fb03-4e9f-a66c-7f90def9b5f7\") " Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.552736 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4622b25d-fb03-4e9f-a66c-7f90def9b5f7-config-volume" (OuterVolumeSpecName: "config-volume") pod "4622b25d-fb03-4e9f-a66c-7f90def9b5f7" (UID: "4622b25d-fb03-4e9f-a66c-7f90def9b5f7"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.559871 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4622b25d-fb03-4e9f-a66c-7f90def9b5f7-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4622b25d-fb03-4e9f-a66c-7f90def9b5f7" (UID: "4622b25d-fb03-4e9f-a66c-7f90def9b5f7"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.560266 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4622b25d-fb03-4e9f-a66c-7f90def9b5f7-kube-api-access-qjjcv" (OuterVolumeSpecName: "kube-api-access-qjjcv") pod "4622b25d-fb03-4e9f-a66c-7f90def9b5f7" (UID: "4622b25d-fb03-4e9f-a66c-7f90def9b5f7"). InnerVolumeSpecName "kube-api-access-qjjcv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.665039 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ec3fd801-c4f4-4b19-ab98-1e8e5ac644b2-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"ec3fd801-c4f4-4b19-ab98-1e8e5ac644b2\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.665884 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ec3fd801-c4f4-4b19-ab98-1e8e5ac644b2-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"ec3fd801-c4f4-4b19-ab98-1e8e5ac644b2\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.666189 4890 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4622b25d-fb03-4e9f-a66c-7f90def9b5f7-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.666210 4890 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4622b25d-fb03-4e9f-a66c-7f90def9b5f7-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.666225 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qjjcv\" (UniqueName: \"kubernetes.io/projected/4622b25d-fb03-4e9f-a66c-7f90def9b5f7-kube-api-access-qjjcv\") on node \"crc\" DevicePath \"\"" Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.731105 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kq8dw"] Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.768253 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ec3fd801-c4f4-4b19-ab98-1e8e5ac644b2-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"ec3fd801-c4f4-4b19-ab98-1e8e5ac644b2\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.768264 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ec3fd801-c4f4-4b19-ab98-1e8e5ac644b2-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"ec3fd801-c4f4-4b19-ab98-1e8e5ac644b2\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.768415 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ec3fd801-c4f4-4b19-ab98-1e8e5ac644b2-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"ec3fd801-c4f4-4b19-ab98-1e8e5ac644b2\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.788319 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ec3fd801-c4f4-4b19-ab98-1e8e5ac644b2-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"ec3fd801-c4f4-4b19-ab98-1e8e5ac644b2\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.841638 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.893857 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" event={"ID":"96800ba3-0ee4-42e0-b8f7-316ef57a6173","Type":"ContainerStarted","Data":"7eae6ff20f6d7f1191e0f629ea1c2463746285c08eb996c6c7ba18f5a4856b72"} Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.901511 4890 generic.go:334] "Generic (PLEG): container finished" podID="4cc12bd5-1c51-42fe-b534-5158f7fe27d7" containerID="03d7aace4d4f9f33aa0de6899b16cf88719b92ea6f2c7fc47b255ebd2ecfb39c" exitCode=0 Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.905717 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"4cc12bd5-1c51-42fe-b534-5158f7fe27d7","Type":"ContainerDied","Data":"03d7aace4d4f9f33aa0de6899b16cf88719b92ea6f2c7fc47b255ebd2ecfb39c"} Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.907849 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401380-96zpb" event={"ID":"4622b25d-fb03-4e9f-a66c-7f90def9b5f7","Type":"ContainerDied","Data":"10b5fe4f40386180673adf770f673c3d1da6f2c20208c98dc6d43657b3f33122"} Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.907900 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="10b5fe4f40386180673adf770f673c3d1da6f2c20208c98dc6d43657b3f33122" Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.907976 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401380-96zpb" Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.908947 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vf44g" event={"ID":"22b61785-937e-4562-8c48-5e3d390f3389","Type":"ContainerStarted","Data":"189fdd9f5e09b7cd03c235fd64aa8fec639f58a6e2e643d6066b2d5882d8d914"} Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.936952 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" podStartSLOduration=83.936927593 podStartE2EDuration="1m23.936927593s" podCreationTimestamp="2025-11-25 15:02:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:04:04.922755305 +0000 UTC m=+103.365217915" watchObservedRunningTime="2025-11-25 15:04:04.936927593 +0000 UTC m=+103.379390203" Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.949302 4890 generic.go:334] "Generic (PLEG): container finished" podID="7fbb3062-51d8-4fc0-9b25-5336de75d692" containerID="f7e0aec046f1f257d4a70d708669775c3cca407ce449e708d87b0062817414b0" exitCode=0 Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.950343 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hn24q" event={"ID":"7fbb3062-51d8-4fc0-9b25-5336de75d692","Type":"ContainerDied","Data":"f7e0aec046f1f257d4a70d708669775c3cca407ce449e708d87b0062817414b0"} Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.958348 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kq8dw" event={"ID":"2f0bd74d-3205-422a-8c40-28ae9d29dc3a","Type":"ContainerStarted","Data":"5e454994990abe652d55378c352c61be4ffcf12798afbcf96d9791e869659b40"} Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.978726 4890 generic.go:334] "Generic (PLEG): container finished" podID="1b3f25c9-5630-4623-8ee4-49e52c356530" containerID="197fbac9c97f052e944100bf7f9f40745e24fa43e336172a29730bb5b89958f4" exitCode=0 Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.979414 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zt6cj" event={"ID":"1b3f25c9-5630-4623-8ee4-49e52c356530","Type":"ContainerDied","Data":"197fbac9c97f052e944100bf7f9f40745e24fa43e336172a29730bb5b89958f4"} Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.979457 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zt6cj" event={"ID":"1b3f25c9-5630-4623-8ee4-49e52c356530","Type":"ContainerStarted","Data":"c78ee8e6cf9ecb89952d16ee84c7a20b4db8355b72ff4d8ef6ee49cd4649ccca"} Nov 25 15:04:04 crc kubenswrapper[4890]: I1125 15:04:04.986834 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8zql" Nov 25 15:04:05 crc kubenswrapper[4890]: I1125 15:04:05.173540 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-kxfxn" Nov 25 15:04:05 crc kubenswrapper[4890]: I1125 15:04:05.177845 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-kxfxn" Nov 25 15:04:05 crc kubenswrapper[4890]: I1125 15:04:05.441340 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 25 15:04:05 crc kubenswrapper[4890]: I1125 15:04:05.986212 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kq8dw" event={"ID":"2f0bd74d-3205-422a-8c40-28ae9d29dc3a","Type":"ContainerStarted","Data":"e789da9878ec6123f3ca4c7cc9727d52d2b8e333f3db12faaa4bc941d405b7c6"} Nov 25 15:04:05 crc kubenswrapper[4890]: I1125 15:04:05.990392 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ec3fd801-c4f4-4b19-ab98-1e8e5ac644b2","Type":"ContainerStarted","Data":"ab03d4f99b0b2d933f30af09c5ef0e91ef2144d418b9fbed57395d6e5cf134e0"} Nov 25 15:04:05 crc kubenswrapper[4890]: I1125 15:04:05.999168 4890 generic.go:334] "Generic (PLEG): container finished" podID="22b61785-937e-4562-8c48-5e3d390f3389" containerID="ba7ced8954142349680e569794df37ccd3da2c9b7be0397aee6e5ac06a31693c" exitCode=0 Nov 25 15:04:06 crc kubenswrapper[4890]: I1125 15:04:06.000003 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vf44g" event={"ID":"22b61785-937e-4562-8c48-5e3d390f3389","Type":"ContainerDied","Data":"ba7ced8954142349680e569794df37ccd3da2c9b7be0397aee6e5ac06a31693c"} Nov 25 15:04:06 crc kubenswrapper[4890]: I1125 15:04:06.001000 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:04:06 crc kubenswrapper[4890]: I1125 15:04:06.430783 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 15:04:06 crc kubenswrapper[4890]: I1125 15:04:06.521866 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4cc12bd5-1c51-42fe-b534-5158f7fe27d7-kubelet-dir\") pod \"4cc12bd5-1c51-42fe-b534-5158f7fe27d7\" (UID: \"4cc12bd5-1c51-42fe-b534-5158f7fe27d7\") " Nov 25 15:04:06 crc kubenswrapper[4890]: I1125 15:04:06.522116 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4cc12bd5-1c51-42fe-b534-5158f7fe27d7-kube-api-access\") pod \"4cc12bd5-1c51-42fe-b534-5158f7fe27d7\" (UID: \"4cc12bd5-1c51-42fe-b534-5158f7fe27d7\") " Nov 25 15:04:06 crc kubenswrapper[4890]: I1125 15:04:06.524581 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4cc12bd5-1c51-42fe-b534-5158f7fe27d7-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "4cc12bd5-1c51-42fe-b534-5158f7fe27d7" (UID: "4cc12bd5-1c51-42fe-b534-5158f7fe27d7"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 15:04:06 crc kubenswrapper[4890]: I1125 15:04:06.531436 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4cc12bd5-1c51-42fe-b534-5158f7fe27d7-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "4cc12bd5-1c51-42fe-b534-5158f7fe27d7" (UID: "4cc12bd5-1c51-42fe-b534-5158f7fe27d7"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:04:06 crc kubenswrapper[4890]: I1125 15:04:06.624789 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4cc12bd5-1c51-42fe-b534-5158f7fe27d7-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 15:04:06 crc kubenswrapper[4890]: I1125 15:04:06.625309 4890 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4cc12bd5-1c51-42fe-b534-5158f7fe27d7-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 25 15:04:07 crc kubenswrapper[4890]: I1125 15:04:07.009934 4890 generic.go:334] "Generic (PLEG): container finished" podID="2f0bd74d-3205-422a-8c40-28ae9d29dc3a" containerID="e789da9878ec6123f3ca4c7cc9727d52d2b8e333f3db12faaa4bc941d405b7c6" exitCode=0 Nov 25 15:04:07 crc kubenswrapper[4890]: I1125 15:04:07.010010 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kq8dw" event={"ID":"2f0bd74d-3205-422a-8c40-28ae9d29dc3a","Type":"ContainerDied","Data":"e789da9878ec6123f3ca4c7cc9727d52d2b8e333f3db12faaa4bc941d405b7c6"} Nov 25 15:04:07 crc kubenswrapper[4890]: I1125 15:04:07.014408 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ec3fd801-c4f4-4b19-ab98-1e8e5ac644b2","Type":"ContainerStarted","Data":"b9cf0e7d2455f63ae8b765fb922864d9a301663ed2a5fcc5d0b17e4bd6dfe960"} Nov 25 15:04:07 crc kubenswrapper[4890]: I1125 15:04:07.024417 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 15:04:07 crc kubenswrapper[4890]: I1125 15:04:07.024404 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"4cc12bd5-1c51-42fe-b534-5158f7fe27d7","Type":"ContainerDied","Data":"15e32b819108d5c06cbe745e0a76969332da425b8794ff4171e587d25834b8bd"} Nov 25 15:04:07 crc kubenswrapper[4890]: I1125 15:04:07.024573 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="15e32b819108d5c06cbe745e0a76969332da425b8794ff4171e587d25834b8bd" Nov 25 15:04:07 crc kubenswrapper[4890]: I1125 15:04:07.064381 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=3.064341359 podStartE2EDuration="3.064341359s" podCreationTimestamp="2025-11-25 15:04:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:04:07.051291702 +0000 UTC m=+105.493754322" watchObservedRunningTime="2025-11-25 15:04:07.064341359 +0000 UTC m=+105.506803969" Nov 25 15:04:08 crc kubenswrapper[4890]: I1125 15:04:08.058470 4890 generic.go:334] "Generic (PLEG): container finished" podID="ec3fd801-c4f4-4b19-ab98-1e8e5ac644b2" containerID="b9cf0e7d2455f63ae8b765fb922864d9a301663ed2a5fcc5d0b17e4bd6dfe960" exitCode=0 Nov 25 15:04:08 crc kubenswrapper[4890]: I1125 15:04:08.058525 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ec3fd801-c4f4-4b19-ab98-1e8e5ac644b2","Type":"ContainerDied","Data":"b9cf0e7d2455f63ae8b765fb922864d9a301663ed2a5fcc5d0b17e4bd6dfe960"} Nov 25 15:04:09 crc kubenswrapper[4890]: I1125 15:04:09.112812 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-ms5f2" Nov 25 15:04:12 crc kubenswrapper[4890]: I1125 15:04:12.749050 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 15:04:12 crc kubenswrapper[4890]: I1125 15:04:12.953768 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ec3fd801-c4f4-4b19-ab98-1e8e5ac644b2-kubelet-dir\") pod \"ec3fd801-c4f4-4b19-ab98-1e8e5ac644b2\" (UID: \"ec3fd801-c4f4-4b19-ab98-1e8e5ac644b2\") " Nov 25 15:04:12 crc kubenswrapper[4890]: I1125 15:04:12.953835 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ec3fd801-c4f4-4b19-ab98-1e8e5ac644b2-kube-api-access\") pod \"ec3fd801-c4f4-4b19-ab98-1e8e5ac644b2\" (UID: \"ec3fd801-c4f4-4b19-ab98-1e8e5ac644b2\") " Nov 25 15:04:12 crc kubenswrapper[4890]: I1125 15:04:12.958273 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ec3fd801-c4f4-4b19-ab98-1e8e5ac644b2-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "ec3fd801-c4f4-4b19-ab98-1e8e5ac644b2" (UID: "ec3fd801-c4f4-4b19-ab98-1e8e5ac644b2"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 15:04:12 crc kubenswrapper[4890]: I1125 15:04:12.971809 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec3fd801-c4f4-4b19-ab98-1e8e5ac644b2-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "ec3fd801-c4f4-4b19-ab98-1e8e5ac644b2" (UID: "ec3fd801-c4f4-4b19-ab98-1e8e5ac644b2"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:04:13 crc kubenswrapper[4890]: I1125 15:04:13.055055 4890 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ec3fd801-c4f4-4b19-ab98-1e8e5ac644b2-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 25 15:04:13 crc kubenswrapper[4890]: I1125 15:04:13.055090 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ec3fd801-c4f4-4b19-ab98-1e8e5ac644b2-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 15:04:13 crc kubenswrapper[4890]: I1125 15:04:13.107340 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ec3fd801-c4f4-4b19-ab98-1e8e5ac644b2","Type":"ContainerDied","Data":"ab03d4f99b0b2d933f30af09c5ef0e91ef2144d418b9fbed57395d6e5cf134e0"} Nov 25 15:04:13 crc kubenswrapper[4890]: I1125 15:04:13.107386 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab03d4f99b0b2d933f30af09c5ef0e91ef2144d418b9fbed57395d6e5cf134e0" Nov 25 15:04:13 crc kubenswrapper[4890]: I1125 15:04:13.107400 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 15:04:13 crc kubenswrapper[4890]: I1125 15:04:13.224044 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-t29fd" Nov 25 15:04:13 crc kubenswrapper[4890]: I1125 15:04:13.229583 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-t29fd" Nov 25 15:04:13 crc kubenswrapper[4890]: I1125 15:04:13.268876 4890 patch_prober.go:28] interesting pod/downloads-7954f5f757-8bln9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Nov 25 15:04:13 crc kubenswrapper[4890]: I1125 15:04:13.268921 4890 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-8bln9" podUID="89bebf40-d718-4168-8377-4fa334afaa17" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Nov 25 15:04:13 crc kubenswrapper[4890]: I1125 15:04:13.269231 4890 patch_prober.go:28] interesting pod/downloads-7954f5f757-8bln9 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Nov 25 15:04:13 crc kubenswrapper[4890]: I1125 15:04:13.269251 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-8bln9" podUID="89bebf40-d718-4168-8377-4fa334afaa17" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Nov 25 15:04:22 crc kubenswrapper[4890]: I1125 15:04:22.673595 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:04:23 crc kubenswrapper[4890]: I1125 15:04:23.275509 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-8bln9" Nov 25 15:04:32 crc kubenswrapper[4890]: E1125 15:04:32.595384 4890 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 25 15:04:32 crc kubenswrapper[4890]: E1125 15:04:32.595938 4890 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mlz6n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-tfnr6_openshift-marketplace(8f052eff-1a42-4c20-8644-d575fc557060): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 15:04:32 crc kubenswrapper[4890]: E1125 15:04:32.597136 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-tfnr6" podUID="8f052eff-1a42-4c20-8644-d575fc557060" Nov 25 15:04:32 crc kubenswrapper[4890]: E1125 15:04:32.717741 4890 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 25 15:04:32 crc kubenswrapper[4890]: E1125 15:04:32.717908 4890 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-45v9w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-mh5wg_openshift-marketplace(68222e1b-b4d5-4e64-853e-a8af2ad8383a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 15:04:32 crc kubenswrapper[4890]: E1125 15:04:32.719101 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-mh5wg" podUID="68222e1b-b4d5-4e64-853e-a8af2ad8383a" Nov 25 15:04:33 crc kubenswrapper[4890]: I1125 15:04:33.876339 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-z5tfd" Nov 25 15:04:34 crc kubenswrapper[4890]: E1125 15:04:34.224401 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-tfnr6" podUID="8f052eff-1a42-4c20-8644-d575fc557060" Nov 25 15:04:34 crc kubenswrapper[4890]: E1125 15:04:34.224531 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-mh5wg" podUID="68222e1b-b4d5-4e64-853e-a8af2ad8383a" Nov 25 15:04:37 crc kubenswrapper[4890]: E1125 15:04:37.934437 4890 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 25 15:04:37 crc kubenswrapper[4890]: E1125 15:04:37.934742 4890 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4kw68,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-t2pmf_openshift-marketplace(f56ea937-8ed8-4912-b721-2ed83a15fa2b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 15:04:37 crc kubenswrapper[4890]: E1125 15:04:37.935956 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-t2pmf" podUID="f56ea937-8ed8-4912-b721-2ed83a15fa2b" Nov 25 15:04:39 crc kubenswrapper[4890]: E1125 15:04:39.190348 4890 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 25 15:04:39 crc kubenswrapper[4890]: E1125 15:04:39.191574 4890 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-x2drz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-z4b8p_openshift-marketplace(b1983120-f62d-46b0-833e-db30a81973c1): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 15:04:39 crc kubenswrapper[4890]: E1125 15:04:39.193239 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-z4b8p" podUID="b1983120-f62d-46b0-833e-db30a81973c1" Nov 25 15:04:39 crc kubenswrapper[4890]: E1125 15:04:39.321470 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-z4b8p" podUID="b1983120-f62d-46b0-833e-db30a81973c1" Nov 25 15:04:39 crc kubenswrapper[4890]: E1125 15:04:39.321553 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-t2pmf" podUID="f56ea937-8ed8-4912-b721-2ed83a15fa2b" Nov 25 15:04:39 crc kubenswrapper[4890]: E1125 15:04:39.380977 4890 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 25 15:04:39 crc kubenswrapper[4890]: E1125 15:04:39.381199 4890 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mkwzk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-hn24q_openshift-marketplace(7fbb3062-51d8-4fc0-9b25-5336de75d692): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 15:04:39 crc kubenswrapper[4890]: E1125 15:04:39.383759 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-hn24q" podUID="7fbb3062-51d8-4fc0-9b25-5336de75d692" Nov 25 15:04:39 crc kubenswrapper[4890]: E1125 15:04:39.424488 4890 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 25 15:04:39 crc kubenswrapper[4890]: E1125 15:04:39.424637 4890 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8799j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-zt6cj_openshift-marketplace(1b3f25c9-5630-4623-8ee4-49e52c356530): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 15:04:39 crc kubenswrapper[4890]: E1125 15:04:39.425800 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-zt6cj" podUID="1b3f25c9-5630-4623-8ee4-49e52c356530" Nov 25 15:04:42 crc kubenswrapper[4890]: E1125 15:04:42.580744 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-zt6cj" podUID="1b3f25c9-5630-4623-8ee4-49e52c356530" Nov 25 15:04:42 crc kubenswrapper[4890]: E1125 15:04:42.581374 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-hn24q" podUID="7fbb3062-51d8-4fc0-9b25-5336de75d692" Nov 25 15:04:44 crc kubenswrapper[4890]: I1125 15:04:44.275964 4890 generic.go:334] "Generic (PLEG): container finished" podID="22b61785-937e-4562-8c48-5e3d390f3389" containerID="3dbd44e5beadfc1768f2a52baf3ed9e523cadb8edb361a7cc36b6e0f71266084" exitCode=0 Nov 25 15:04:44 crc kubenswrapper[4890]: I1125 15:04:44.276051 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vf44g" event={"ID":"22b61785-937e-4562-8c48-5e3d390f3389","Type":"ContainerDied","Data":"3dbd44e5beadfc1768f2a52baf3ed9e523cadb8edb361a7cc36b6e0f71266084"} Nov 25 15:04:44 crc kubenswrapper[4890]: I1125 15:04:44.280275 4890 generic.go:334] "Generic (PLEG): container finished" podID="2f0bd74d-3205-422a-8c40-28ae9d29dc3a" containerID="1b9a8599a20ca58994061afeba58b05c1c04643773589841cad3cb123ceeafe6" exitCode=0 Nov 25 15:04:44 crc kubenswrapper[4890]: I1125 15:04:44.280310 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kq8dw" event={"ID":"2f0bd74d-3205-422a-8c40-28ae9d29dc3a","Type":"ContainerDied","Data":"1b9a8599a20ca58994061afeba58b05c1c04643773589841cad3cb123ceeafe6"} Nov 25 15:04:46 crc kubenswrapper[4890]: I1125 15:04:46.292465 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kq8dw" event={"ID":"2f0bd74d-3205-422a-8c40-28ae9d29dc3a","Type":"ContainerStarted","Data":"3f1f92623e2a9195b28ccc29ffebdf3c1e7ca76fcd4f15de3fb621ee18d84fb3"} Nov 25 15:04:46 crc kubenswrapper[4890]: I1125 15:04:46.313740 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kq8dw" podStartSLOduration=10.42719019 podStartE2EDuration="43.313723579s" podCreationTimestamp="2025-11-25 15:04:03 +0000 UTC" firstStartedPulling="2025-11-25 15:04:12.712519884 +0000 UTC m=+111.154982494" lastFinishedPulling="2025-11-25 15:04:45.599053283 +0000 UTC m=+144.041515883" observedRunningTime="2025-11-25 15:04:46.309008624 +0000 UTC m=+144.751471224" watchObservedRunningTime="2025-11-25 15:04:46.313723579 +0000 UTC m=+144.756186199" Nov 25 15:04:47 crc kubenswrapper[4890]: I1125 15:04:47.300275 4890 generic.go:334] "Generic (PLEG): container finished" podID="8f052eff-1a42-4c20-8644-d575fc557060" containerID="97823a3f2d4dcc99f7b1980839e5e94fc7b18d857a99e776c43aa11d1c9e7c36" exitCode=0 Nov 25 15:04:47 crc kubenswrapper[4890]: I1125 15:04:47.300358 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tfnr6" event={"ID":"8f052eff-1a42-4c20-8644-d575fc557060","Type":"ContainerDied","Data":"97823a3f2d4dcc99f7b1980839e5e94fc7b18d857a99e776c43aa11d1c9e7c36"} Nov 25 15:04:47 crc kubenswrapper[4890]: I1125 15:04:47.306014 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vf44g" event={"ID":"22b61785-937e-4562-8c48-5e3d390f3389","Type":"ContainerStarted","Data":"1ad675f01e648f268e022e0e561a44eeac06d708344116d1e20e8f72fb0d5a82"} Nov 25 15:04:47 crc kubenswrapper[4890]: I1125 15:04:47.339343 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vf44g" podStartSLOduration=11.286797498 podStartE2EDuration="44.339323157s" podCreationTimestamp="2025-11-25 15:04:03 +0000 UTC" firstStartedPulling="2025-11-25 15:04:12.713670815 +0000 UTC m=+111.156133435" lastFinishedPulling="2025-11-25 15:04:45.766196484 +0000 UTC m=+144.208659094" observedRunningTime="2025-11-25 15:04:47.33806127 +0000 UTC m=+145.780523880" watchObservedRunningTime="2025-11-25 15:04:47.339323157 +0000 UTC m=+145.781785767" Nov 25 15:04:48 crc kubenswrapper[4890]: I1125 15:04:48.314187 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tfnr6" event={"ID":"8f052eff-1a42-4c20-8644-d575fc557060","Type":"ContainerStarted","Data":"e1591e78261809d129766d029d60ca310164e994d53c83ec1c55c3ca7e263a85"} Nov 25 15:04:48 crc kubenswrapper[4890]: I1125 15:04:48.348308 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tfnr6" podStartSLOduration=3.40403285 podStartE2EDuration="48.348275608s" podCreationTimestamp="2025-11-25 15:04:00 +0000 UTC" firstStartedPulling="2025-11-25 15:04:02.739709987 +0000 UTC m=+101.182172597" lastFinishedPulling="2025-11-25 15:04:47.683952745 +0000 UTC m=+146.126415355" observedRunningTime="2025-11-25 15:04:48.337339245 +0000 UTC m=+146.779801855" watchObservedRunningTime="2025-11-25 15:04:48.348275608 +0000 UTC m=+146.790738218" Nov 25 15:04:49 crc kubenswrapper[4890]: I1125 15:04:49.348387 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 15:04:49 crc kubenswrapper[4890]: I1125 15:04:49.348763 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:04:49 crc kubenswrapper[4890]: I1125 15:04:49.348826 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 15:04:49 crc kubenswrapper[4890]: I1125 15:04:49.348855 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:04:49 crc kubenswrapper[4890]: I1125 15:04:49.353515 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 25 15:04:49 crc kubenswrapper[4890]: I1125 15:04:49.353994 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 25 15:04:49 crc kubenswrapper[4890]: I1125 15:04:49.354068 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 25 15:04:49 crc kubenswrapper[4890]: I1125 15:04:49.360263 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 25 15:04:49 crc kubenswrapper[4890]: I1125 15:04:49.361260 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:04:49 crc kubenswrapper[4890]: I1125 15:04:49.365309 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:04:49 crc kubenswrapper[4890]: I1125 15:04:49.372772 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 15:04:49 crc kubenswrapper[4890]: I1125 15:04:49.373217 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 15:04:49 crc kubenswrapper[4890]: I1125 15:04:49.493387 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 15:04:49 crc kubenswrapper[4890]: I1125 15:04:49.511004 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 15:04:49 crc kubenswrapper[4890]: I1125 15:04:49.519691 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 15:04:49 crc kubenswrapper[4890]: W1125 15:04:49.998657 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-19290bb2f9c82ba44069f36d7f5a81cf82e68ed9cc03e78c2eca040ccea048e8 WatchSource:0}: Error finding container 19290bb2f9c82ba44069f36d7f5a81cf82e68ed9cc03e78c2eca040ccea048e8: Status 404 returned error can't find the container with id 19290bb2f9c82ba44069f36d7f5a81cf82e68ed9cc03e78c2eca040ccea048e8 Nov 25 15:04:50 crc kubenswrapper[4890]: W1125 15:04:50.057108 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-876da85eeb3d4152d45ea2e9c97f5d660c4ee151d75d21934b94a89c7f760920 WatchSource:0}: Error finding container 876da85eeb3d4152d45ea2e9c97f5d660c4ee151d75d21934b94a89c7f760920: Status 404 returned error can't find the container with id 876da85eeb3d4152d45ea2e9c97f5d660c4ee151d75d21934b94a89c7f760920 Nov 25 15:04:50 crc kubenswrapper[4890]: W1125 15:04:50.058837 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-ea170392fe459754398a15ff97b5397ad89dd9bf2356b8cd3c2745a9e4934ac4 WatchSource:0}: Error finding container ea170392fe459754398a15ff97b5397ad89dd9bf2356b8cd3c2745a9e4934ac4: Status 404 returned error can't find the container with id ea170392fe459754398a15ff97b5397ad89dd9bf2356b8cd3c2745a9e4934ac4 Nov 25 15:04:50 crc kubenswrapper[4890]: I1125 15:04:50.334896 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"19290bb2f9c82ba44069f36d7f5a81cf82e68ed9cc03e78c2eca040ccea048e8"} Nov 25 15:04:50 crc kubenswrapper[4890]: I1125 15:04:50.340026 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"ea170392fe459754398a15ff97b5397ad89dd9bf2356b8cd3c2745a9e4934ac4"} Nov 25 15:04:50 crc kubenswrapper[4890]: I1125 15:04:50.341466 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"876da85eeb3d4152d45ea2e9c97f5d660c4ee151d75d21934b94a89c7f760920"} Nov 25 15:04:50 crc kubenswrapper[4890]: I1125 15:04:50.734575 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tfnr6" Nov 25 15:04:50 crc kubenswrapper[4890]: I1125 15:04:50.735088 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tfnr6" Nov 25 15:04:50 crc kubenswrapper[4890]: I1125 15:04:50.972127 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tfnr6" Nov 25 15:04:51 crc kubenswrapper[4890]: I1125 15:04:51.359985 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"3f91abe5a3c6a5df3e5cb1431b9279ed662e612ad3a5d08384ec3f11cfbdcac7"} Nov 25 15:04:51 crc kubenswrapper[4890]: I1125 15:04:51.363703 4890 generic.go:334] "Generic (PLEG): container finished" podID="68222e1b-b4d5-4e64-853e-a8af2ad8383a" containerID="60511c7ad7128b68077b90ca64cdbcea4f5d823ca886be330f26bff4d4fd6c4a" exitCode=0 Nov 25 15:04:51 crc kubenswrapper[4890]: I1125 15:04:51.363779 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mh5wg" event={"ID":"68222e1b-b4d5-4e64-853e-a8af2ad8383a","Type":"ContainerDied","Data":"60511c7ad7128b68077b90ca64cdbcea4f5d823ca886be330f26bff4d4fd6c4a"} Nov 25 15:04:51 crc kubenswrapper[4890]: I1125 15:04:51.367274 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"e80e598333cf37e00dbd1ccfca39a959fe385ec8cb07621e175e5f6f2592825d"} Nov 25 15:04:51 crc kubenswrapper[4890]: I1125 15:04:51.367468 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 15:04:51 crc kubenswrapper[4890]: I1125 15:04:51.370336 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"5d62cc292da9c7e7959e9ac2609f4baaea52fd6621aa397a6d4c8169a75a1d7e"} Nov 25 15:04:52 crc kubenswrapper[4890]: I1125 15:04:52.378786 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mh5wg" event={"ID":"68222e1b-b4d5-4e64-853e-a8af2ad8383a","Type":"ContainerStarted","Data":"2155b21428b6abbe0ec9ddbb0c673297f55251bf2d310e46838eeb37bd0224d4"} Nov 25 15:04:52 crc kubenswrapper[4890]: I1125 15:04:52.424936 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tfnr6" Nov 25 15:04:52 crc kubenswrapper[4890]: I1125 15:04:52.447761 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mh5wg" podStartSLOduration=3.218904282 podStartE2EDuration="52.447726957s" podCreationTimestamp="2025-11-25 15:04:00 +0000 UTC" firstStartedPulling="2025-11-25 15:04:02.764032053 +0000 UTC m=+101.206494663" lastFinishedPulling="2025-11-25 15:04:51.992854718 +0000 UTC m=+150.435317338" observedRunningTime="2025-11-25 15:04:52.398187097 +0000 UTC m=+150.840649707" watchObservedRunningTime="2025-11-25 15:04:52.447726957 +0000 UTC m=+150.890189567" Nov 25 15:04:53 crc kubenswrapper[4890]: I1125 15:04:53.386495 4890 generic.go:334] "Generic (PLEG): container finished" podID="f56ea937-8ed8-4912-b721-2ed83a15fa2b" containerID="c1cf5d9f4bc7429e2ecd9b0b4917e8014df37c6e4afaca13e1fc262d445442cb" exitCode=0 Nov 25 15:04:53 crc kubenswrapper[4890]: I1125 15:04:53.386575 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t2pmf" event={"ID":"f56ea937-8ed8-4912-b721-2ed83a15fa2b","Type":"ContainerDied","Data":"c1cf5d9f4bc7429e2ecd9b0b4917e8014df37c6e4afaca13e1fc262d445442cb"} Nov 25 15:04:53 crc kubenswrapper[4890]: I1125 15:04:53.886484 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vf44g" Nov 25 15:04:53 crc kubenswrapper[4890]: I1125 15:04:53.886893 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vf44g" Nov 25 15:04:53 crc kubenswrapper[4890]: I1125 15:04:53.933657 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vf44g" Nov 25 15:04:54 crc kubenswrapper[4890]: I1125 15:04:54.267917 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kq8dw" Nov 25 15:04:54 crc kubenswrapper[4890]: I1125 15:04:54.268281 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kq8dw" Nov 25 15:04:54 crc kubenswrapper[4890]: I1125 15:04:54.382983 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kq8dw" Nov 25 15:04:54 crc kubenswrapper[4890]: I1125 15:04:54.440517 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kq8dw" Nov 25 15:04:54 crc kubenswrapper[4890]: I1125 15:04:54.441027 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vf44g" Nov 25 15:04:56 crc kubenswrapper[4890]: I1125 15:04:56.406973 4890 generic.go:334] "Generic (PLEG): container finished" podID="b1983120-f62d-46b0-833e-db30a81973c1" containerID="317401cc98b8211f11251559a4e5bf87764469a03304557c3931d7c9d3f1c13f" exitCode=0 Nov 25 15:04:56 crc kubenswrapper[4890]: I1125 15:04:56.407220 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z4b8p" event={"ID":"b1983120-f62d-46b0-833e-db30a81973c1","Type":"ContainerDied","Data":"317401cc98b8211f11251559a4e5bf87764469a03304557c3931d7c9d3f1c13f"} Nov 25 15:04:56 crc kubenswrapper[4890]: I1125 15:04:56.447583 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 15:04:56 crc kubenswrapper[4890]: I1125 15:04:56.447695 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 15:04:56 crc kubenswrapper[4890]: I1125 15:04:56.706073 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kq8dw"] Nov 25 15:04:56 crc kubenswrapper[4890]: I1125 15:04:56.706337 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kq8dw" podUID="2f0bd74d-3205-422a-8c40-28ae9d29dc3a" containerName="registry-server" containerID="cri-o://3f1f92623e2a9195b28ccc29ffebdf3c1e7ca76fcd4f15de3fb621ee18d84fb3" gracePeriod=2 Nov 25 15:04:57 crc kubenswrapper[4890]: I1125 15:04:57.448241 4890 generic.go:334] "Generic (PLEG): container finished" podID="2f0bd74d-3205-422a-8c40-28ae9d29dc3a" containerID="3f1f92623e2a9195b28ccc29ffebdf3c1e7ca76fcd4f15de3fb621ee18d84fb3" exitCode=0 Nov 25 15:04:57 crc kubenswrapper[4890]: I1125 15:04:57.448311 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kq8dw" event={"ID":"2f0bd74d-3205-422a-8c40-28ae9d29dc3a","Type":"ContainerDied","Data":"3f1f92623e2a9195b28ccc29ffebdf3c1e7ca76fcd4f15de3fb621ee18d84fb3"} Nov 25 15:04:58 crc kubenswrapper[4890]: I1125 15:04:58.411842 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kq8dw" Nov 25 15:04:58 crc kubenswrapper[4890]: I1125 15:04:58.455696 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kq8dw" event={"ID":"2f0bd74d-3205-422a-8c40-28ae9d29dc3a","Type":"ContainerDied","Data":"5e454994990abe652d55378c352c61be4ffcf12798afbcf96d9791e869659b40"} Nov 25 15:04:58 crc kubenswrapper[4890]: I1125 15:04:58.455759 4890 scope.go:117] "RemoveContainer" containerID="3f1f92623e2a9195b28ccc29ffebdf3c1e7ca76fcd4f15de3fb621ee18d84fb3" Nov 25 15:04:58 crc kubenswrapper[4890]: I1125 15:04:58.455720 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kq8dw" Nov 25 15:04:58 crc kubenswrapper[4890]: I1125 15:04:58.572218 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f0bd74d-3205-422a-8c40-28ae9d29dc3a-utilities\") pod \"2f0bd74d-3205-422a-8c40-28ae9d29dc3a\" (UID: \"2f0bd74d-3205-422a-8c40-28ae9d29dc3a\") " Nov 25 15:04:58 crc kubenswrapper[4890]: I1125 15:04:58.572379 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jjwk9\" (UniqueName: \"kubernetes.io/projected/2f0bd74d-3205-422a-8c40-28ae9d29dc3a-kube-api-access-jjwk9\") pod \"2f0bd74d-3205-422a-8c40-28ae9d29dc3a\" (UID: \"2f0bd74d-3205-422a-8c40-28ae9d29dc3a\") " Nov 25 15:04:58 crc kubenswrapper[4890]: I1125 15:04:58.572420 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f0bd74d-3205-422a-8c40-28ae9d29dc3a-catalog-content\") pod \"2f0bd74d-3205-422a-8c40-28ae9d29dc3a\" (UID: \"2f0bd74d-3205-422a-8c40-28ae9d29dc3a\") " Nov 25 15:04:58 crc kubenswrapper[4890]: I1125 15:04:58.573390 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f0bd74d-3205-422a-8c40-28ae9d29dc3a-utilities" (OuterVolumeSpecName: "utilities") pod "2f0bd74d-3205-422a-8c40-28ae9d29dc3a" (UID: "2f0bd74d-3205-422a-8c40-28ae9d29dc3a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:04:58 crc kubenswrapper[4890]: I1125 15:04:58.578070 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f0bd74d-3205-422a-8c40-28ae9d29dc3a-kube-api-access-jjwk9" (OuterVolumeSpecName: "kube-api-access-jjwk9") pod "2f0bd74d-3205-422a-8c40-28ae9d29dc3a" (UID: "2f0bd74d-3205-422a-8c40-28ae9d29dc3a"). InnerVolumeSpecName "kube-api-access-jjwk9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:04:58 crc kubenswrapper[4890]: I1125 15:04:58.674267 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jjwk9\" (UniqueName: \"kubernetes.io/projected/2f0bd74d-3205-422a-8c40-28ae9d29dc3a-kube-api-access-jjwk9\") on node \"crc\" DevicePath \"\"" Nov 25 15:04:58 crc kubenswrapper[4890]: I1125 15:04:58.674310 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f0bd74d-3205-422a-8c40-28ae9d29dc3a-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 15:04:58 crc kubenswrapper[4890]: I1125 15:04:58.725802 4890 scope.go:117] "RemoveContainer" containerID="1b9a8599a20ca58994061afeba58b05c1c04643773589841cad3cb123ceeafe6" Nov 25 15:04:58 crc kubenswrapper[4890]: I1125 15:04:58.747918 4890 scope.go:117] "RemoveContainer" containerID="e789da9878ec6123f3ca4c7cc9727d52d2b8e333f3db12faaa4bc941d405b7c6" Nov 25 15:04:59 crc kubenswrapper[4890]: I1125 15:04:59.986194 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f0bd74d-3205-422a-8c40-28ae9d29dc3a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2f0bd74d-3205-422a-8c40-28ae9d29dc3a" (UID: "2f0bd74d-3205-422a-8c40-28ae9d29dc3a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:04:59 crc kubenswrapper[4890]: I1125 15:04:59.990942 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f0bd74d-3205-422a-8c40-28ae9d29dc3a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 15:05:00 crc kubenswrapper[4890]: I1125 15:05:00.280773 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kq8dw"] Nov 25 15:05:00 crc kubenswrapper[4890]: I1125 15:05:00.290782 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kq8dw"] Nov 25 15:05:00 crc kubenswrapper[4890]: I1125 15:05:00.472485 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t2pmf" event={"ID":"f56ea937-8ed8-4912-b721-2ed83a15fa2b","Type":"ContainerStarted","Data":"bef03c35383a2a3ab221cd21cb092f50a8ddfc2924d06f5f4a0ad00934ebe1fb"} Nov 25 15:05:01 crc kubenswrapper[4890]: I1125 15:05:01.144538 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mh5wg" Nov 25 15:05:01 crc kubenswrapper[4890]: I1125 15:05:01.144598 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mh5wg" Nov 25 15:05:01 crc kubenswrapper[4890]: I1125 15:05:01.188358 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mh5wg" Nov 25 15:05:01 crc kubenswrapper[4890]: I1125 15:05:01.529400 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mh5wg" Nov 25 15:05:01 crc kubenswrapper[4890]: I1125 15:05:01.545745 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-t2pmf" podStartSLOduration=4.43984691 podStartE2EDuration="1m1.545719305s" podCreationTimestamp="2025-11-25 15:04:00 +0000 UTC" firstStartedPulling="2025-11-25 15:04:01.620150916 +0000 UTC m=+100.062613526" lastFinishedPulling="2025-11-25 15:04:58.726023311 +0000 UTC m=+157.168485921" observedRunningTime="2025-11-25 15:05:01.499213382 +0000 UTC m=+159.941676002" watchObservedRunningTime="2025-11-25 15:05:01.545719305 +0000 UTC m=+159.988181915" Nov 25 15:05:02 crc kubenswrapper[4890]: I1125 15:05:02.180821 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f0bd74d-3205-422a-8c40-28ae9d29dc3a" path="/var/lib/kubelet/pods/2f0bd74d-3205-422a-8c40-28ae9d29dc3a/volumes" Nov 25 15:05:03 crc kubenswrapper[4890]: I1125 15:05:03.107430 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mh5wg"] Nov 25 15:05:03 crc kubenswrapper[4890]: I1125 15:05:03.488061 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mh5wg" podUID="68222e1b-b4d5-4e64-853e-a8af2ad8383a" containerName="registry-server" containerID="cri-o://2155b21428b6abbe0ec9ddbb0c673297f55251bf2d310e46838eeb37bd0224d4" gracePeriod=2 Nov 25 15:05:07 crc kubenswrapper[4890]: I1125 15:05:07.107297 4890 generic.go:334] "Generic (PLEG): container finished" podID="68222e1b-b4d5-4e64-853e-a8af2ad8383a" containerID="2155b21428b6abbe0ec9ddbb0c673297f55251bf2d310e46838eeb37bd0224d4" exitCode=0 Nov 25 15:05:07 crc kubenswrapper[4890]: I1125 15:05:07.107375 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mh5wg" event={"ID":"68222e1b-b4d5-4e64-853e-a8af2ad8383a","Type":"ContainerDied","Data":"2155b21428b6abbe0ec9ddbb0c673297f55251bf2d310e46838eeb37bd0224d4"} Nov 25 15:05:08 crc kubenswrapper[4890]: I1125 15:05:08.431388 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mh5wg" Nov 25 15:05:08 crc kubenswrapper[4890]: I1125 15:05:08.625251 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68222e1b-b4d5-4e64-853e-a8af2ad8383a-utilities\") pod \"68222e1b-b4d5-4e64-853e-a8af2ad8383a\" (UID: \"68222e1b-b4d5-4e64-853e-a8af2ad8383a\") " Nov 25 15:05:08 crc kubenswrapper[4890]: I1125 15:05:08.625404 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68222e1b-b4d5-4e64-853e-a8af2ad8383a-catalog-content\") pod \"68222e1b-b4d5-4e64-853e-a8af2ad8383a\" (UID: \"68222e1b-b4d5-4e64-853e-a8af2ad8383a\") " Nov 25 15:05:08 crc kubenswrapper[4890]: I1125 15:05:08.625458 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-45v9w\" (UniqueName: \"kubernetes.io/projected/68222e1b-b4d5-4e64-853e-a8af2ad8383a-kube-api-access-45v9w\") pod \"68222e1b-b4d5-4e64-853e-a8af2ad8383a\" (UID: \"68222e1b-b4d5-4e64-853e-a8af2ad8383a\") " Nov 25 15:05:08 crc kubenswrapper[4890]: I1125 15:05:08.626848 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68222e1b-b4d5-4e64-853e-a8af2ad8383a-utilities" (OuterVolumeSpecName: "utilities") pod "68222e1b-b4d5-4e64-853e-a8af2ad8383a" (UID: "68222e1b-b4d5-4e64-853e-a8af2ad8383a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:05:08 crc kubenswrapper[4890]: I1125 15:05:08.641525 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68222e1b-b4d5-4e64-853e-a8af2ad8383a-kube-api-access-45v9w" (OuterVolumeSpecName: "kube-api-access-45v9w") pod "68222e1b-b4d5-4e64-853e-a8af2ad8383a" (UID: "68222e1b-b4d5-4e64-853e-a8af2ad8383a"). InnerVolumeSpecName "kube-api-access-45v9w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:05:08 crc kubenswrapper[4890]: I1125 15:05:08.671792 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68222e1b-b4d5-4e64-853e-a8af2ad8383a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "68222e1b-b4d5-4e64-853e-a8af2ad8383a" (UID: "68222e1b-b4d5-4e64-853e-a8af2ad8383a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:05:08 crc kubenswrapper[4890]: I1125 15:05:08.727753 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68222e1b-b4d5-4e64-853e-a8af2ad8383a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 15:05:08 crc kubenswrapper[4890]: I1125 15:05:08.727782 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-45v9w\" (UniqueName: \"kubernetes.io/projected/68222e1b-b4d5-4e64-853e-a8af2ad8383a-kube-api-access-45v9w\") on node \"crc\" DevicePath \"\"" Nov 25 15:05:08 crc kubenswrapper[4890]: I1125 15:05:08.727794 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68222e1b-b4d5-4e64-853e-a8af2ad8383a-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 15:05:09 crc kubenswrapper[4890]: I1125 15:05:09.121668 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mh5wg" event={"ID":"68222e1b-b4d5-4e64-853e-a8af2ad8383a","Type":"ContainerDied","Data":"f01695da1108260f00d78b82bd23de7fe506f51da8d3ff7e5a930f0614864644"} Nov 25 15:05:09 crc kubenswrapper[4890]: I1125 15:05:09.121721 4890 scope.go:117] "RemoveContainer" containerID="2155b21428b6abbe0ec9ddbb0c673297f55251bf2d310e46838eeb37bd0224d4" Nov 25 15:05:09 crc kubenswrapper[4890]: I1125 15:05:09.121727 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mh5wg" Nov 25 15:05:09 crc kubenswrapper[4890]: I1125 15:05:09.147764 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mh5wg"] Nov 25 15:05:09 crc kubenswrapper[4890]: I1125 15:05:09.150709 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mh5wg"] Nov 25 15:05:09 crc kubenswrapper[4890]: I1125 15:05:09.713909 4890 scope.go:117] "RemoveContainer" containerID="60511c7ad7128b68077b90ca64cdbcea4f5d823ca886be330f26bff4d4fd6c4a" Nov 25 15:05:10 crc kubenswrapper[4890]: I1125 15:05:10.181309 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68222e1b-b4d5-4e64-853e-a8af2ad8383a" path="/var/lib/kubelet/pods/68222e1b-b4d5-4e64-853e-a8af2ad8383a/volumes" Nov 25 15:05:10 crc kubenswrapper[4890]: I1125 15:05:10.806577 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-t2pmf" Nov 25 15:05:10 crc kubenswrapper[4890]: I1125 15:05:10.806656 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-t2pmf" Nov 25 15:05:10 crc kubenswrapper[4890]: I1125 15:05:10.847079 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-t2pmf" Nov 25 15:05:11 crc kubenswrapper[4890]: I1125 15:05:11.176982 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-t2pmf" Nov 25 15:05:11 crc kubenswrapper[4890]: I1125 15:05:11.372260 4890 scope.go:117] "RemoveContainer" containerID="26be49e805a1db93198aba9739155a7a7e942b9401dbf28ae551cba63c892244" Nov 25 15:05:13 crc kubenswrapper[4890]: I1125 15:05:13.108068 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-t2pmf"] Nov 25 15:05:13 crc kubenswrapper[4890]: I1125 15:05:13.147429 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z4b8p" event={"ID":"b1983120-f62d-46b0-833e-db30a81973c1","Type":"ContainerStarted","Data":"59347c6caa2d3aff1243c63d5b531762bea390ab80bcad117cda132ae2a5f867"} Nov 25 15:05:13 crc kubenswrapper[4890]: I1125 15:05:13.149283 4890 generic.go:334] "Generic (PLEG): container finished" podID="1b3f25c9-5630-4623-8ee4-49e52c356530" containerID="2881d08dcb14cf46e92395b413d4f520c050fc12114718381f3eb48d52af8208" exitCode=0 Nov 25 15:05:13 crc kubenswrapper[4890]: I1125 15:05:13.149343 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zt6cj" event={"ID":"1b3f25c9-5630-4623-8ee4-49e52c356530","Type":"ContainerDied","Data":"2881d08dcb14cf46e92395b413d4f520c050fc12114718381f3eb48d52af8208"} Nov 25 15:05:13 crc kubenswrapper[4890]: I1125 15:05:13.151423 4890 generic.go:334] "Generic (PLEG): container finished" podID="7fbb3062-51d8-4fc0-9b25-5336de75d692" containerID="62a36226c66383b36f4eda2fe0aa72ab055026ff30490aef7a4b72c4e7610471" exitCode=0 Nov 25 15:05:13 crc kubenswrapper[4890]: I1125 15:05:13.151620 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-t2pmf" podUID="f56ea937-8ed8-4912-b721-2ed83a15fa2b" containerName="registry-server" containerID="cri-o://bef03c35383a2a3ab221cd21cb092f50a8ddfc2924d06f5f4a0ad00934ebe1fb" gracePeriod=2 Nov 25 15:05:13 crc kubenswrapper[4890]: I1125 15:05:13.151690 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hn24q" event={"ID":"7fbb3062-51d8-4fc0-9b25-5336de75d692","Type":"ContainerDied","Data":"62a36226c66383b36f4eda2fe0aa72ab055026ff30490aef7a4b72c4e7610471"} Nov 25 15:05:13 crc kubenswrapper[4890]: I1125 15:05:13.191578 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-z4b8p" podStartSLOduration=4.548045202 podStartE2EDuration="1m13.191559428s" podCreationTimestamp="2025-11-25 15:04:00 +0000 UTC" firstStartedPulling="2025-11-25 15:04:02.723914165 +0000 UTC m=+101.166376765" lastFinishedPulling="2025-11-25 15:05:11.367428391 +0000 UTC m=+169.809890991" observedRunningTime="2025-11-25 15:05:13.168896908 +0000 UTC m=+171.611359528" watchObservedRunningTime="2025-11-25 15:05:13.191559428 +0000 UTC m=+171.634022038" Nov 25 15:05:13 crc kubenswrapper[4890]: I1125 15:05:13.472582 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t2pmf" Nov 25 15:05:13 crc kubenswrapper[4890]: I1125 15:05:13.590732 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f56ea937-8ed8-4912-b721-2ed83a15fa2b-catalog-content\") pod \"f56ea937-8ed8-4912-b721-2ed83a15fa2b\" (UID: \"f56ea937-8ed8-4912-b721-2ed83a15fa2b\") " Nov 25 15:05:13 crc kubenswrapper[4890]: I1125 15:05:13.590790 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4kw68\" (UniqueName: \"kubernetes.io/projected/f56ea937-8ed8-4912-b721-2ed83a15fa2b-kube-api-access-4kw68\") pod \"f56ea937-8ed8-4912-b721-2ed83a15fa2b\" (UID: \"f56ea937-8ed8-4912-b721-2ed83a15fa2b\") " Nov 25 15:05:13 crc kubenswrapper[4890]: I1125 15:05:13.590839 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f56ea937-8ed8-4912-b721-2ed83a15fa2b-utilities\") pod \"f56ea937-8ed8-4912-b721-2ed83a15fa2b\" (UID: \"f56ea937-8ed8-4912-b721-2ed83a15fa2b\") " Nov 25 15:05:13 crc kubenswrapper[4890]: I1125 15:05:13.591636 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f56ea937-8ed8-4912-b721-2ed83a15fa2b-utilities" (OuterVolumeSpecName: "utilities") pod "f56ea937-8ed8-4912-b721-2ed83a15fa2b" (UID: "f56ea937-8ed8-4912-b721-2ed83a15fa2b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:05:13 crc kubenswrapper[4890]: I1125 15:05:13.591902 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f56ea937-8ed8-4912-b721-2ed83a15fa2b-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 15:05:13 crc kubenswrapper[4890]: I1125 15:05:13.597391 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f56ea937-8ed8-4912-b721-2ed83a15fa2b-kube-api-access-4kw68" (OuterVolumeSpecName: "kube-api-access-4kw68") pod "f56ea937-8ed8-4912-b721-2ed83a15fa2b" (UID: "f56ea937-8ed8-4912-b721-2ed83a15fa2b"). InnerVolumeSpecName "kube-api-access-4kw68". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:05:13 crc kubenswrapper[4890]: I1125 15:05:13.639241 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f56ea937-8ed8-4912-b721-2ed83a15fa2b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f56ea937-8ed8-4912-b721-2ed83a15fa2b" (UID: "f56ea937-8ed8-4912-b721-2ed83a15fa2b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:05:13 crc kubenswrapper[4890]: I1125 15:05:13.693803 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f56ea937-8ed8-4912-b721-2ed83a15fa2b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 15:05:13 crc kubenswrapper[4890]: I1125 15:05:13.693841 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4kw68\" (UniqueName: \"kubernetes.io/projected/f56ea937-8ed8-4912-b721-2ed83a15fa2b-kube-api-access-4kw68\") on node \"crc\" DevicePath \"\"" Nov 25 15:05:14 crc kubenswrapper[4890]: I1125 15:05:14.160080 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hn24q" event={"ID":"7fbb3062-51d8-4fc0-9b25-5336de75d692","Type":"ContainerStarted","Data":"b14a44d27ed55bfd08a7795cc6c29e7e0ab52f3b0a5f017adde77b9bf4d06d8d"} Nov 25 15:05:14 crc kubenswrapper[4890]: I1125 15:05:14.163092 4890 generic.go:334] "Generic (PLEG): container finished" podID="f56ea937-8ed8-4912-b721-2ed83a15fa2b" containerID="bef03c35383a2a3ab221cd21cb092f50a8ddfc2924d06f5f4a0ad00934ebe1fb" exitCode=0 Nov 25 15:05:14 crc kubenswrapper[4890]: I1125 15:05:14.163193 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t2pmf" event={"ID":"f56ea937-8ed8-4912-b721-2ed83a15fa2b","Type":"ContainerDied","Data":"bef03c35383a2a3ab221cd21cb092f50a8ddfc2924d06f5f4a0ad00934ebe1fb"} Nov 25 15:05:14 crc kubenswrapper[4890]: I1125 15:05:14.163248 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t2pmf" event={"ID":"f56ea937-8ed8-4912-b721-2ed83a15fa2b","Type":"ContainerDied","Data":"4d40823fb1444992ffb52d2e52c051f274cbaa484f86cc7e3b6e9e8163c43eac"} Nov 25 15:05:14 crc kubenswrapper[4890]: I1125 15:05:14.163280 4890 scope.go:117] "RemoveContainer" containerID="bef03c35383a2a3ab221cd21cb092f50a8ddfc2924d06f5f4a0ad00934ebe1fb" Nov 25 15:05:14 crc kubenswrapper[4890]: I1125 15:05:14.163297 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t2pmf" Nov 25 15:05:14 crc kubenswrapper[4890]: I1125 15:05:14.166382 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zt6cj" event={"ID":"1b3f25c9-5630-4623-8ee4-49e52c356530","Type":"ContainerStarted","Data":"2b8f6c83e40c8e47b96aa4f3b88f05a93a0dc17cd08b3fa98dc47c54c2cfd0b5"} Nov 25 15:05:14 crc kubenswrapper[4890]: I1125 15:05:14.189436 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hn24q" podStartSLOduration=3.627844608 podStartE2EDuration="1m12.189417371s" podCreationTimestamp="2025-11-25 15:04:02 +0000 UTC" firstStartedPulling="2025-11-25 15:04:05.006467946 +0000 UTC m=+103.448930556" lastFinishedPulling="2025-11-25 15:05:13.568040709 +0000 UTC m=+172.010503319" observedRunningTime="2025-11-25 15:05:14.187366392 +0000 UTC m=+172.629829002" watchObservedRunningTime="2025-11-25 15:05:14.189417371 +0000 UTC m=+172.631879981" Nov 25 15:05:14 crc kubenswrapper[4890]: I1125 15:05:14.194419 4890 scope.go:117] "RemoveContainer" containerID="c1cf5d9f4bc7429e2ecd9b0b4917e8014df37c6e4afaca13e1fc262d445442cb" Nov 25 15:05:14 crc kubenswrapper[4890]: I1125 15:05:14.219403 4890 scope.go:117] "RemoveContainer" containerID="f284f9fe799e44084a0da77f9d43c13d505cd30ee11d0ea81b9615874110ad1e" Nov 25 15:05:14 crc kubenswrapper[4890]: I1125 15:05:14.221064 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zt6cj" podStartSLOduration=3.551817058 podStartE2EDuration="1m12.221046688s" podCreationTimestamp="2025-11-25 15:04:02 +0000 UTC" firstStartedPulling="2025-11-25 15:04:04.98360352 +0000 UTC m=+103.426066130" lastFinishedPulling="2025-11-25 15:05:13.65283315 +0000 UTC m=+172.095295760" observedRunningTime="2025-11-25 15:05:14.221041538 +0000 UTC m=+172.663504178" watchObservedRunningTime="2025-11-25 15:05:14.221046688 +0000 UTC m=+172.663509298" Nov 25 15:05:14 crc kubenswrapper[4890]: I1125 15:05:14.240352 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-t2pmf"] Nov 25 15:05:14 crc kubenswrapper[4890]: I1125 15:05:14.241627 4890 scope.go:117] "RemoveContainer" containerID="bef03c35383a2a3ab221cd21cb092f50a8ddfc2924d06f5f4a0ad00934ebe1fb" Nov 25 15:05:14 crc kubenswrapper[4890]: E1125 15:05:14.242413 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bef03c35383a2a3ab221cd21cb092f50a8ddfc2924d06f5f4a0ad00934ebe1fb\": container with ID starting with bef03c35383a2a3ab221cd21cb092f50a8ddfc2924d06f5f4a0ad00934ebe1fb not found: ID does not exist" containerID="bef03c35383a2a3ab221cd21cb092f50a8ddfc2924d06f5f4a0ad00934ebe1fb" Nov 25 15:05:14 crc kubenswrapper[4890]: I1125 15:05:14.242455 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bef03c35383a2a3ab221cd21cb092f50a8ddfc2924d06f5f4a0ad00934ebe1fb"} err="failed to get container status \"bef03c35383a2a3ab221cd21cb092f50a8ddfc2924d06f5f4a0ad00934ebe1fb\": rpc error: code = NotFound desc = could not find container \"bef03c35383a2a3ab221cd21cb092f50a8ddfc2924d06f5f4a0ad00934ebe1fb\": container with ID starting with bef03c35383a2a3ab221cd21cb092f50a8ddfc2924d06f5f4a0ad00934ebe1fb not found: ID does not exist" Nov 25 15:05:14 crc kubenswrapper[4890]: I1125 15:05:14.242551 4890 scope.go:117] "RemoveContainer" containerID="c1cf5d9f4bc7429e2ecd9b0b4917e8014df37c6e4afaca13e1fc262d445442cb" Nov 25 15:05:14 crc kubenswrapper[4890]: E1125 15:05:14.249370 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1cf5d9f4bc7429e2ecd9b0b4917e8014df37c6e4afaca13e1fc262d445442cb\": container with ID starting with c1cf5d9f4bc7429e2ecd9b0b4917e8014df37c6e4afaca13e1fc262d445442cb not found: ID does not exist" containerID="c1cf5d9f4bc7429e2ecd9b0b4917e8014df37c6e4afaca13e1fc262d445442cb" Nov 25 15:05:14 crc kubenswrapper[4890]: I1125 15:05:14.249435 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1cf5d9f4bc7429e2ecd9b0b4917e8014df37c6e4afaca13e1fc262d445442cb"} err="failed to get container status \"c1cf5d9f4bc7429e2ecd9b0b4917e8014df37c6e4afaca13e1fc262d445442cb\": rpc error: code = NotFound desc = could not find container \"c1cf5d9f4bc7429e2ecd9b0b4917e8014df37c6e4afaca13e1fc262d445442cb\": container with ID starting with c1cf5d9f4bc7429e2ecd9b0b4917e8014df37c6e4afaca13e1fc262d445442cb not found: ID does not exist" Nov 25 15:05:14 crc kubenswrapper[4890]: I1125 15:05:14.249479 4890 scope.go:117] "RemoveContainer" containerID="f284f9fe799e44084a0da77f9d43c13d505cd30ee11d0ea81b9615874110ad1e" Nov 25 15:05:14 crc kubenswrapper[4890]: I1125 15:05:14.252017 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-t2pmf"] Nov 25 15:05:14 crc kubenswrapper[4890]: E1125 15:05:14.253346 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f284f9fe799e44084a0da77f9d43c13d505cd30ee11d0ea81b9615874110ad1e\": container with ID starting with f284f9fe799e44084a0da77f9d43c13d505cd30ee11d0ea81b9615874110ad1e not found: ID does not exist" containerID="f284f9fe799e44084a0da77f9d43c13d505cd30ee11d0ea81b9615874110ad1e" Nov 25 15:05:14 crc kubenswrapper[4890]: I1125 15:05:14.253405 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f284f9fe799e44084a0da77f9d43c13d505cd30ee11d0ea81b9615874110ad1e"} err="failed to get container status \"f284f9fe799e44084a0da77f9d43c13d505cd30ee11d0ea81b9615874110ad1e\": rpc error: code = NotFound desc = could not find container \"f284f9fe799e44084a0da77f9d43c13d505cd30ee11d0ea81b9615874110ad1e\": container with ID starting with f284f9fe799e44084a0da77f9d43c13d505cd30ee11d0ea81b9615874110ad1e not found: ID does not exist" Nov 25 15:05:16 crc kubenswrapper[4890]: I1125 15:05:16.178455 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f56ea937-8ed8-4912-b721-2ed83a15fa2b" path="/var/lib/kubelet/pods/f56ea937-8ed8-4912-b721-2ed83a15fa2b/volumes" Nov 25 15:05:20 crc kubenswrapper[4890]: I1125 15:05:20.477511 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-z4b8p" Nov 25 15:05:20 crc kubenswrapper[4890]: I1125 15:05:20.477886 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-z4b8p" Nov 25 15:05:20 crc kubenswrapper[4890]: I1125 15:05:20.522439 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-z4b8p" Nov 25 15:05:21 crc kubenswrapper[4890]: I1125 15:05:21.243802 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-z4b8p" Nov 25 15:05:22 crc kubenswrapper[4890]: I1125 15:05:22.088522 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mmhbh"] Nov 25 15:05:22 crc kubenswrapper[4890]: I1125 15:05:22.584223 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hn24q" Nov 25 15:05:22 crc kubenswrapper[4890]: I1125 15:05:22.584427 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hn24q" Nov 25 15:05:22 crc kubenswrapper[4890]: I1125 15:05:22.622529 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hn24q" Nov 25 15:05:23 crc kubenswrapper[4890]: I1125 15:05:23.028569 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zt6cj" Nov 25 15:05:23 crc kubenswrapper[4890]: I1125 15:05:23.028941 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zt6cj" Nov 25 15:05:23 crc kubenswrapper[4890]: I1125 15:05:23.066068 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zt6cj" Nov 25 15:05:23 crc kubenswrapper[4890]: I1125 15:05:23.249903 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hn24q" Nov 25 15:05:23 crc kubenswrapper[4890]: I1125 15:05:23.253321 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zt6cj" Nov 25 15:05:25 crc kubenswrapper[4890]: I1125 15:05:25.309690 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zt6cj"] Nov 25 15:05:25 crc kubenswrapper[4890]: I1125 15:05:25.310090 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zt6cj" podUID="1b3f25c9-5630-4623-8ee4-49e52c356530" containerName="registry-server" containerID="cri-o://2b8f6c83e40c8e47b96aa4f3b88f05a93a0dc17cd08b3fa98dc47c54c2cfd0b5" gracePeriod=2 Nov 25 15:05:26 crc kubenswrapper[4890]: I1125 15:05:26.199596 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zt6cj" Nov 25 15:05:26 crc kubenswrapper[4890]: I1125 15:05:26.233737 4890 generic.go:334] "Generic (PLEG): container finished" podID="1b3f25c9-5630-4623-8ee4-49e52c356530" containerID="2b8f6c83e40c8e47b96aa4f3b88f05a93a0dc17cd08b3fa98dc47c54c2cfd0b5" exitCode=0 Nov 25 15:05:26 crc kubenswrapper[4890]: I1125 15:05:26.233798 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zt6cj" Nov 25 15:05:26 crc kubenswrapper[4890]: I1125 15:05:26.233780 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zt6cj" event={"ID":"1b3f25c9-5630-4623-8ee4-49e52c356530","Type":"ContainerDied","Data":"2b8f6c83e40c8e47b96aa4f3b88f05a93a0dc17cd08b3fa98dc47c54c2cfd0b5"} Nov 25 15:05:26 crc kubenswrapper[4890]: I1125 15:05:26.233927 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zt6cj" event={"ID":"1b3f25c9-5630-4623-8ee4-49e52c356530","Type":"ContainerDied","Data":"c78ee8e6cf9ecb89952d16ee84c7a20b4db8355b72ff4d8ef6ee49cd4649ccca"} Nov 25 15:05:26 crc kubenswrapper[4890]: I1125 15:05:26.233959 4890 scope.go:117] "RemoveContainer" containerID="2b8f6c83e40c8e47b96aa4f3b88f05a93a0dc17cd08b3fa98dc47c54c2cfd0b5" Nov 25 15:05:26 crc kubenswrapper[4890]: I1125 15:05:26.248493 4890 scope.go:117] "RemoveContainer" containerID="2881d08dcb14cf46e92395b413d4f520c050fc12114718381f3eb48d52af8208" Nov 25 15:05:26 crc kubenswrapper[4890]: I1125 15:05:26.261140 4890 scope.go:117] "RemoveContainer" containerID="197fbac9c97f052e944100bf7f9f40745e24fa43e336172a29730bb5b89958f4" Nov 25 15:05:26 crc kubenswrapper[4890]: I1125 15:05:26.274476 4890 scope.go:117] "RemoveContainer" containerID="2b8f6c83e40c8e47b96aa4f3b88f05a93a0dc17cd08b3fa98dc47c54c2cfd0b5" Nov 25 15:05:26 crc kubenswrapper[4890]: E1125 15:05:26.274820 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b8f6c83e40c8e47b96aa4f3b88f05a93a0dc17cd08b3fa98dc47c54c2cfd0b5\": container with ID starting with 2b8f6c83e40c8e47b96aa4f3b88f05a93a0dc17cd08b3fa98dc47c54c2cfd0b5 not found: ID does not exist" containerID="2b8f6c83e40c8e47b96aa4f3b88f05a93a0dc17cd08b3fa98dc47c54c2cfd0b5" Nov 25 15:05:26 crc kubenswrapper[4890]: I1125 15:05:26.274862 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b8f6c83e40c8e47b96aa4f3b88f05a93a0dc17cd08b3fa98dc47c54c2cfd0b5"} err="failed to get container status \"2b8f6c83e40c8e47b96aa4f3b88f05a93a0dc17cd08b3fa98dc47c54c2cfd0b5\": rpc error: code = NotFound desc = could not find container \"2b8f6c83e40c8e47b96aa4f3b88f05a93a0dc17cd08b3fa98dc47c54c2cfd0b5\": container with ID starting with 2b8f6c83e40c8e47b96aa4f3b88f05a93a0dc17cd08b3fa98dc47c54c2cfd0b5 not found: ID does not exist" Nov 25 15:05:26 crc kubenswrapper[4890]: I1125 15:05:26.274891 4890 scope.go:117] "RemoveContainer" containerID="2881d08dcb14cf46e92395b413d4f520c050fc12114718381f3eb48d52af8208" Nov 25 15:05:26 crc kubenswrapper[4890]: E1125 15:05:26.275191 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2881d08dcb14cf46e92395b413d4f520c050fc12114718381f3eb48d52af8208\": container with ID starting with 2881d08dcb14cf46e92395b413d4f520c050fc12114718381f3eb48d52af8208 not found: ID does not exist" containerID="2881d08dcb14cf46e92395b413d4f520c050fc12114718381f3eb48d52af8208" Nov 25 15:05:26 crc kubenswrapper[4890]: I1125 15:05:26.275212 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2881d08dcb14cf46e92395b413d4f520c050fc12114718381f3eb48d52af8208"} err="failed to get container status \"2881d08dcb14cf46e92395b413d4f520c050fc12114718381f3eb48d52af8208\": rpc error: code = NotFound desc = could not find container \"2881d08dcb14cf46e92395b413d4f520c050fc12114718381f3eb48d52af8208\": container with ID starting with 2881d08dcb14cf46e92395b413d4f520c050fc12114718381f3eb48d52af8208 not found: ID does not exist" Nov 25 15:05:26 crc kubenswrapper[4890]: I1125 15:05:26.275225 4890 scope.go:117] "RemoveContainer" containerID="197fbac9c97f052e944100bf7f9f40745e24fa43e336172a29730bb5b89958f4" Nov 25 15:05:26 crc kubenswrapper[4890]: E1125 15:05:26.275488 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"197fbac9c97f052e944100bf7f9f40745e24fa43e336172a29730bb5b89958f4\": container with ID starting with 197fbac9c97f052e944100bf7f9f40745e24fa43e336172a29730bb5b89958f4 not found: ID does not exist" containerID="197fbac9c97f052e944100bf7f9f40745e24fa43e336172a29730bb5b89958f4" Nov 25 15:05:26 crc kubenswrapper[4890]: I1125 15:05:26.275523 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"197fbac9c97f052e944100bf7f9f40745e24fa43e336172a29730bb5b89958f4"} err="failed to get container status \"197fbac9c97f052e944100bf7f9f40745e24fa43e336172a29730bb5b89958f4\": rpc error: code = NotFound desc = could not find container \"197fbac9c97f052e944100bf7f9f40745e24fa43e336172a29730bb5b89958f4\": container with ID starting with 197fbac9c97f052e944100bf7f9f40745e24fa43e336172a29730bb5b89958f4 not found: ID does not exist" Nov 25 15:05:26 crc kubenswrapper[4890]: I1125 15:05:26.382016 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b3f25c9-5630-4623-8ee4-49e52c356530-utilities\") pod \"1b3f25c9-5630-4623-8ee4-49e52c356530\" (UID: \"1b3f25c9-5630-4623-8ee4-49e52c356530\") " Nov 25 15:05:26 crc kubenswrapper[4890]: I1125 15:05:26.382100 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b3f25c9-5630-4623-8ee4-49e52c356530-catalog-content\") pod \"1b3f25c9-5630-4623-8ee4-49e52c356530\" (UID: \"1b3f25c9-5630-4623-8ee4-49e52c356530\") " Nov 25 15:05:26 crc kubenswrapper[4890]: I1125 15:05:26.382234 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8799j\" (UniqueName: \"kubernetes.io/projected/1b3f25c9-5630-4623-8ee4-49e52c356530-kube-api-access-8799j\") pod \"1b3f25c9-5630-4623-8ee4-49e52c356530\" (UID: \"1b3f25c9-5630-4623-8ee4-49e52c356530\") " Nov 25 15:05:26 crc kubenswrapper[4890]: I1125 15:05:26.383368 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b3f25c9-5630-4623-8ee4-49e52c356530-utilities" (OuterVolumeSpecName: "utilities") pod "1b3f25c9-5630-4623-8ee4-49e52c356530" (UID: "1b3f25c9-5630-4623-8ee4-49e52c356530"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:05:26 crc kubenswrapper[4890]: I1125 15:05:26.386955 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b3f25c9-5630-4623-8ee4-49e52c356530-kube-api-access-8799j" (OuterVolumeSpecName: "kube-api-access-8799j") pod "1b3f25c9-5630-4623-8ee4-49e52c356530" (UID: "1b3f25c9-5630-4623-8ee4-49e52c356530"). InnerVolumeSpecName "kube-api-access-8799j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:05:26 crc kubenswrapper[4890]: I1125 15:05:26.399466 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b3f25c9-5630-4623-8ee4-49e52c356530-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1b3f25c9-5630-4623-8ee4-49e52c356530" (UID: "1b3f25c9-5630-4623-8ee4-49e52c356530"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:05:26 crc kubenswrapper[4890]: I1125 15:05:26.447440 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 15:05:26 crc kubenswrapper[4890]: I1125 15:05:26.447504 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 15:05:26 crc kubenswrapper[4890]: I1125 15:05:26.483989 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8799j\" (UniqueName: \"kubernetes.io/projected/1b3f25c9-5630-4623-8ee4-49e52c356530-kube-api-access-8799j\") on node \"crc\" DevicePath \"\"" Nov 25 15:05:26 crc kubenswrapper[4890]: I1125 15:05:26.484032 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b3f25c9-5630-4623-8ee4-49e52c356530-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 15:05:26 crc kubenswrapper[4890]: I1125 15:05:26.484050 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b3f25c9-5630-4623-8ee4-49e52c356530-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 15:05:26 crc kubenswrapper[4890]: I1125 15:05:26.562078 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zt6cj"] Nov 25 15:05:26 crc kubenswrapper[4890]: I1125 15:05:26.565432 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zt6cj"] Nov 25 15:05:28 crc kubenswrapper[4890]: I1125 15:05:28.180975 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b3f25c9-5630-4623-8ee4-49e52c356530" path="/var/lib/kubelet/pods/1b3f25c9-5630-4623-8ee4-49e52c356530/volumes" Nov 25 15:05:29 crc kubenswrapper[4890]: I1125 15:05:29.522699 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.112627 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" podUID="a27aab8b-a74f-4a62-b5a3-27d8a7d46be9" containerName="oauth-openshift" containerID="cri-o://62c26f47dd493084e7a893c76b6c6c564775542abea1bc2a7bd94c95c34e92bf" gracePeriod=15 Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.354850 4890 generic.go:334] "Generic (PLEG): container finished" podID="a27aab8b-a74f-4a62-b5a3-27d8a7d46be9" containerID="62c26f47dd493084e7a893c76b6c6c564775542abea1bc2a7bd94c95c34e92bf" exitCode=0 Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.355305 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" event={"ID":"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9","Type":"ContainerDied","Data":"62c26f47dd493084e7a893c76b6c6c564775542abea1bc2a7bd94c95c34e92bf"} Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.538743 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.580547 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6"] Nov 25 15:05:47 crc kubenswrapper[4890]: E1125 15:05:47.581019 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f0bd74d-3205-422a-8c40-28ae9d29dc3a" containerName="extract-utilities" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.581051 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f0bd74d-3205-422a-8c40-28ae9d29dc3a" containerName="extract-utilities" Nov 25 15:05:47 crc kubenswrapper[4890]: E1125 15:05:47.581075 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f56ea937-8ed8-4912-b721-2ed83a15fa2b" containerName="extract-content" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.581088 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="f56ea937-8ed8-4912-b721-2ed83a15fa2b" containerName="extract-content" Nov 25 15:05:47 crc kubenswrapper[4890]: E1125 15:05:47.581107 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f0bd74d-3205-422a-8c40-28ae9d29dc3a" containerName="registry-server" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.581119 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f0bd74d-3205-422a-8c40-28ae9d29dc3a" containerName="registry-server" Nov 25 15:05:47 crc kubenswrapper[4890]: E1125 15:05:47.581135 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f0bd74d-3205-422a-8c40-28ae9d29dc3a" containerName="extract-content" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.581146 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f0bd74d-3205-422a-8c40-28ae9d29dc3a" containerName="extract-content" Nov 25 15:05:47 crc kubenswrapper[4890]: E1125 15:05:47.581156 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b3f25c9-5630-4623-8ee4-49e52c356530" containerName="extract-utilities" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.581188 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b3f25c9-5630-4623-8ee4-49e52c356530" containerName="extract-utilities" Nov 25 15:05:47 crc kubenswrapper[4890]: E1125 15:05:47.581203 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f56ea937-8ed8-4912-b721-2ed83a15fa2b" containerName="registry-server" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.581213 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="f56ea937-8ed8-4912-b721-2ed83a15fa2b" containerName="registry-server" Nov 25 15:05:47 crc kubenswrapper[4890]: E1125 15:05:47.581225 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b3f25c9-5630-4623-8ee4-49e52c356530" containerName="extract-content" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.581235 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b3f25c9-5630-4623-8ee4-49e52c356530" containerName="extract-content" Nov 25 15:05:47 crc kubenswrapper[4890]: E1125 15:05:47.581249 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68222e1b-b4d5-4e64-853e-a8af2ad8383a" containerName="extract-utilities" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.581261 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="68222e1b-b4d5-4e64-853e-a8af2ad8383a" containerName="extract-utilities" Nov 25 15:05:47 crc kubenswrapper[4890]: E1125 15:05:47.581278 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b3f25c9-5630-4623-8ee4-49e52c356530" containerName="registry-server" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.581291 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b3f25c9-5630-4623-8ee4-49e52c356530" containerName="registry-server" Nov 25 15:05:47 crc kubenswrapper[4890]: E1125 15:05:47.581310 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68222e1b-b4d5-4e64-853e-a8af2ad8383a" containerName="extract-content" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.581321 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="68222e1b-b4d5-4e64-853e-a8af2ad8383a" containerName="extract-content" Nov 25 15:05:47 crc kubenswrapper[4890]: E1125 15:05:47.581339 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f56ea937-8ed8-4912-b721-2ed83a15fa2b" containerName="extract-utilities" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.581350 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="f56ea937-8ed8-4912-b721-2ed83a15fa2b" containerName="extract-utilities" Nov 25 15:05:47 crc kubenswrapper[4890]: E1125 15:05:47.581363 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a27aab8b-a74f-4a62-b5a3-27d8a7d46be9" containerName="oauth-openshift" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.581373 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="a27aab8b-a74f-4a62-b5a3-27d8a7d46be9" containerName="oauth-openshift" Nov 25 15:05:47 crc kubenswrapper[4890]: E1125 15:05:47.581389 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec3fd801-c4f4-4b19-ab98-1e8e5ac644b2" containerName="pruner" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.581400 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec3fd801-c4f4-4b19-ab98-1e8e5ac644b2" containerName="pruner" Nov 25 15:05:47 crc kubenswrapper[4890]: E1125 15:05:47.581417 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cc12bd5-1c51-42fe-b534-5158f7fe27d7" containerName="pruner" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.581428 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cc12bd5-1c51-42fe-b534-5158f7fe27d7" containerName="pruner" Nov 25 15:05:47 crc kubenswrapper[4890]: E1125 15:05:47.581444 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68222e1b-b4d5-4e64-853e-a8af2ad8383a" containerName="registry-server" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.581453 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="68222e1b-b4d5-4e64-853e-a8af2ad8383a" containerName="registry-server" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.581617 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="4cc12bd5-1c51-42fe-b534-5158f7fe27d7" containerName="pruner" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.581636 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f0bd74d-3205-422a-8c40-28ae9d29dc3a" containerName="registry-server" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.581646 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="f56ea937-8ed8-4912-b721-2ed83a15fa2b" containerName="registry-server" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.581662 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="a27aab8b-a74f-4a62-b5a3-27d8a7d46be9" containerName="oauth-openshift" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.581678 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b3f25c9-5630-4623-8ee4-49e52c356530" containerName="registry-server" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.581694 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec3fd801-c4f4-4b19-ab98-1e8e5ac644b2" containerName="pruner" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.581710 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="68222e1b-b4d5-4e64-853e-a8af2ad8383a" containerName="registry-server" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.582381 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.587959 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6"] Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.653599 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-system-service-ca\") pod \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.653672 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-user-template-provider-selection\") pod \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.653699 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-user-idp-0-file-data\") pod \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.653753 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-system-router-certs\") pod \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.653790 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-system-session\") pod \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.653814 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-system-trusted-ca-bundle\") pod \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.653841 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-system-serving-cert\") pod \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.653896 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-user-template-login\") pod \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.653917 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2p5mv\" (UniqueName: \"kubernetes.io/projected/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-kube-api-access-2p5mv\") pod \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.653937 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-audit-dir\") pod \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.653972 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-user-template-error\") pod \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.654098 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-system-cliconfig\") pod \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.654117 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-audit-policies\") pod \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.654180 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-system-ocp-branding-template\") pod \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\" (UID: \"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9\") " Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.654369 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3053ea79-6147-4ff5-9e3c-470b947938fb-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7d5bfb7cdd-nzph6\" (UID: \"3053ea79-6147-4ff5-9e3c-470b947938fb\") " pod="openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.654399 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3053ea79-6147-4ff5-9e3c-470b947938fb-audit-policies\") pod \"oauth-openshift-7d5bfb7cdd-nzph6\" (UID: \"3053ea79-6147-4ff5-9e3c-470b947938fb\") " pod="openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.654422 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3053ea79-6147-4ff5-9e3c-470b947938fb-v4-0-config-user-template-login\") pod \"oauth-openshift-7d5bfb7cdd-nzph6\" (UID: \"3053ea79-6147-4ff5-9e3c-470b947938fb\") " pod="openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.654447 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3053ea79-6147-4ff5-9e3c-470b947938fb-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7d5bfb7cdd-nzph6\" (UID: \"3053ea79-6147-4ff5-9e3c-470b947938fb\") " pod="openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.654488 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3053ea79-6147-4ff5-9e3c-470b947938fb-v4-0-config-system-service-ca\") pod \"oauth-openshift-7d5bfb7cdd-nzph6\" (UID: \"3053ea79-6147-4ff5-9e3c-470b947938fb\") " pod="openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.654508 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3053ea79-6147-4ff5-9e3c-470b947938fb-v4-0-config-user-template-error\") pod \"oauth-openshift-7d5bfb7cdd-nzph6\" (UID: \"3053ea79-6147-4ff5-9e3c-470b947938fb\") " pod="openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.654535 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3053ea79-6147-4ff5-9e3c-470b947938fb-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7d5bfb7cdd-nzph6\" (UID: \"3053ea79-6147-4ff5-9e3c-470b947938fb\") " pod="openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.654560 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3053ea79-6147-4ff5-9e3c-470b947938fb-v4-0-config-system-session\") pod \"oauth-openshift-7d5bfb7cdd-nzph6\" (UID: \"3053ea79-6147-4ff5-9e3c-470b947938fb\") " pod="openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.654579 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3053ea79-6147-4ff5-9e3c-470b947938fb-v4-0-config-system-router-certs\") pod \"oauth-openshift-7d5bfb7cdd-nzph6\" (UID: \"3053ea79-6147-4ff5-9e3c-470b947938fb\") " pod="openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.654605 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3053ea79-6147-4ff5-9e3c-470b947938fb-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7d5bfb7cdd-nzph6\" (UID: \"3053ea79-6147-4ff5-9e3c-470b947938fb\") " pod="openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.654625 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3053ea79-6147-4ff5-9e3c-470b947938fb-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7d5bfb7cdd-nzph6\" (UID: \"3053ea79-6147-4ff5-9e3c-470b947938fb\") " pod="openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.654646 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3053ea79-6147-4ff5-9e3c-470b947938fb-audit-dir\") pod \"oauth-openshift-7d5bfb7cdd-nzph6\" (UID: \"3053ea79-6147-4ff5-9e3c-470b947938fb\") " pod="openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.654663 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96hh2\" (UniqueName: \"kubernetes.io/projected/3053ea79-6147-4ff5-9e3c-470b947938fb-kube-api-access-96hh2\") pod \"oauth-openshift-7d5bfb7cdd-nzph6\" (UID: \"3053ea79-6147-4ff5-9e3c-470b947938fb\") " pod="openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.654684 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3053ea79-6147-4ff5-9e3c-470b947938fb-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7d5bfb7cdd-nzph6\" (UID: \"3053ea79-6147-4ff5-9e3c-470b947938fb\") " pod="openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.654750 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "a27aab8b-a74f-4a62-b5a3-27d8a7d46be9" (UID: "a27aab8b-a74f-4a62-b5a3-27d8a7d46be9"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.654774 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "a27aab8b-a74f-4a62-b5a3-27d8a7d46be9" (UID: "a27aab8b-a74f-4a62-b5a3-27d8a7d46be9"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.655589 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "a27aab8b-a74f-4a62-b5a3-27d8a7d46be9" (UID: "a27aab8b-a74f-4a62-b5a3-27d8a7d46be9"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.655713 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "a27aab8b-a74f-4a62-b5a3-27d8a7d46be9" (UID: "a27aab8b-a74f-4a62-b5a3-27d8a7d46be9"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.656143 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "a27aab8b-a74f-4a62-b5a3-27d8a7d46be9" (UID: "a27aab8b-a74f-4a62-b5a3-27d8a7d46be9"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.661405 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "a27aab8b-a74f-4a62-b5a3-27d8a7d46be9" (UID: "a27aab8b-a74f-4a62-b5a3-27d8a7d46be9"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.662396 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "a27aab8b-a74f-4a62-b5a3-27d8a7d46be9" (UID: "a27aab8b-a74f-4a62-b5a3-27d8a7d46be9"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.662611 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-kube-api-access-2p5mv" (OuterVolumeSpecName: "kube-api-access-2p5mv") pod "a27aab8b-a74f-4a62-b5a3-27d8a7d46be9" (UID: "a27aab8b-a74f-4a62-b5a3-27d8a7d46be9"). InnerVolumeSpecName "kube-api-access-2p5mv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.662707 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "a27aab8b-a74f-4a62-b5a3-27d8a7d46be9" (UID: "a27aab8b-a74f-4a62-b5a3-27d8a7d46be9"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.662835 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "a27aab8b-a74f-4a62-b5a3-27d8a7d46be9" (UID: "a27aab8b-a74f-4a62-b5a3-27d8a7d46be9"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.662945 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "a27aab8b-a74f-4a62-b5a3-27d8a7d46be9" (UID: "a27aab8b-a74f-4a62-b5a3-27d8a7d46be9"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.663309 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "a27aab8b-a74f-4a62-b5a3-27d8a7d46be9" (UID: "a27aab8b-a74f-4a62-b5a3-27d8a7d46be9"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.663603 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "a27aab8b-a74f-4a62-b5a3-27d8a7d46be9" (UID: "a27aab8b-a74f-4a62-b5a3-27d8a7d46be9"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.663870 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "a27aab8b-a74f-4a62-b5a3-27d8a7d46be9" (UID: "a27aab8b-a74f-4a62-b5a3-27d8a7d46be9"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.755939 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3053ea79-6147-4ff5-9e3c-470b947938fb-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7d5bfb7cdd-nzph6\" (UID: \"3053ea79-6147-4ff5-9e3c-470b947938fb\") " pod="openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.755995 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3053ea79-6147-4ff5-9e3c-470b947938fb-audit-policies\") pod \"oauth-openshift-7d5bfb7cdd-nzph6\" (UID: \"3053ea79-6147-4ff5-9e3c-470b947938fb\") " pod="openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.756020 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3053ea79-6147-4ff5-9e3c-470b947938fb-v4-0-config-user-template-login\") pod \"oauth-openshift-7d5bfb7cdd-nzph6\" (UID: \"3053ea79-6147-4ff5-9e3c-470b947938fb\") " pod="openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.756058 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3053ea79-6147-4ff5-9e3c-470b947938fb-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7d5bfb7cdd-nzph6\" (UID: \"3053ea79-6147-4ff5-9e3c-470b947938fb\") " pod="openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.756103 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3053ea79-6147-4ff5-9e3c-470b947938fb-v4-0-config-system-service-ca\") pod \"oauth-openshift-7d5bfb7cdd-nzph6\" (UID: \"3053ea79-6147-4ff5-9e3c-470b947938fb\") " pod="openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.756132 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3053ea79-6147-4ff5-9e3c-470b947938fb-v4-0-config-user-template-error\") pod \"oauth-openshift-7d5bfb7cdd-nzph6\" (UID: \"3053ea79-6147-4ff5-9e3c-470b947938fb\") " pod="openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.756187 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3053ea79-6147-4ff5-9e3c-470b947938fb-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7d5bfb7cdd-nzph6\" (UID: \"3053ea79-6147-4ff5-9e3c-470b947938fb\") " pod="openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.756222 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3053ea79-6147-4ff5-9e3c-470b947938fb-v4-0-config-system-session\") pod \"oauth-openshift-7d5bfb7cdd-nzph6\" (UID: \"3053ea79-6147-4ff5-9e3c-470b947938fb\") " pod="openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.756249 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3053ea79-6147-4ff5-9e3c-470b947938fb-v4-0-config-system-router-certs\") pod \"oauth-openshift-7d5bfb7cdd-nzph6\" (UID: \"3053ea79-6147-4ff5-9e3c-470b947938fb\") " pod="openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.756287 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3053ea79-6147-4ff5-9e3c-470b947938fb-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7d5bfb7cdd-nzph6\" (UID: \"3053ea79-6147-4ff5-9e3c-470b947938fb\") " pod="openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.756313 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3053ea79-6147-4ff5-9e3c-470b947938fb-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7d5bfb7cdd-nzph6\" (UID: \"3053ea79-6147-4ff5-9e3c-470b947938fb\") " pod="openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.756343 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3053ea79-6147-4ff5-9e3c-470b947938fb-audit-dir\") pod \"oauth-openshift-7d5bfb7cdd-nzph6\" (UID: \"3053ea79-6147-4ff5-9e3c-470b947938fb\") " pod="openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.756366 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96hh2\" (UniqueName: \"kubernetes.io/projected/3053ea79-6147-4ff5-9e3c-470b947938fb-kube-api-access-96hh2\") pod \"oauth-openshift-7d5bfb7cdd-nzph6\" (UID: \"3053ea79-6147-4ff5-9e3c-470b947938fb\") " pod="openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.756398 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3053ea79-6147-4ff5-9e3c-470b947938fb-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7d5bfb7cdd-nzph6\" (UID: \"3053ea79-6147-4ff5-9e3c-470b947938fb\") " pod="openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.756468 4890 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.756484 4890 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.756500 4890 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.756515 4890 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.756530 4890 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.756547 4890 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.756562 4890 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.756576 4890 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.756591 4890 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.756604 4890 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.756618 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2p5mv\" (UniqueName: \"kubernetes.io/projected/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-kube-api-access-2p5mv\") on node \"crc\" DevicePath \"\"" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.756634 4890 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.756648 4890 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.756660 4890 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.756830 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3053ea79-6147-4ff5-9e3c-470b947938fb-audit-dir\") pod \"oauth-openshift-7d5bfb7cdd-nzph6\" (UID: \"3053ea79-6147-4ff5-9e3c-470b947938fb\") " pod="openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.757256 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3053ea79-6147-4ff5-9e3c-470b947938fb-audit-policies\") pod \"oauth-openshift-7d5bfb7cdd-nzph6\" (UID: \"3053ea79-6147-4ff5-9e3c-470b947938fb\") " pod="openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.757316 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3053ea79-6147-4ff5-9e3c-470b947938fb-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7d5bfb7cdd-nzph6\" (UID: \"3053ea79-6147-4ff5-9e3c-470b947938fb\") " pod="openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.757405 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3053ea79-6147-4ff5-9e3c-470b947938fb-v4-0-config-system-service-ca\") pod \"oauth-openshift-7d5bfb7cdd-nzph6\" (UID: \"3053ea79-6147-4ff5-9e3c-470b947938fb\") " pod="openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.757422 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3053ea79-6147-4ff5-9e3c-470b947938fb-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7d5bfb7cdd-nzph6\" (UID: \"3053ea79-6147-4ff5-9e3c-470b947938fb\") " pod="openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.761672 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3053ea79-6147-4ff5-9e3c-470b947938fb-v4-0-config-user-template-error\") pod \"oauth-openshift-7d5bfb7cdd-nzph6\" (UID: \"3053ea79-6147-4ff5-9e3c-470b947938fb\") " pod="openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.762405 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3053ea79-6147-4ff5-9e3c-470b947938fb-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7d5bfb7cdd-nzph6\" (UID: \"3053ea79-6147-4ff5-9e3c-470b947938fb\") " pod="openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.762957 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3053ea79-6147-4ff5-9e3c-470b947938fb-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7d5bfb7cdd-nzph6\" (UID: \"3053ea79-6147-4ff5-9e3c-470b947938fb\") " pod="openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.763404 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3053ea79-6147-4ff5-9e3c-470b947938fb-v4-0-config-system-router-certs\") pod \"oauth-openshift-7d5bfb7cdd-nzph6\" (UID: \"3053ea79-6147-4ff5-9e3c-470b947938fb\") " pod="openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.763603 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3053ea79-6147-4ff5-9e3c-470b947938fb-v4-0-config-user-template-login\") pod \"oauth-openshift-7d5bfb7cdd-nzph6\" (UID: \"3053ea79-6147-4ff5-9e3c-470b947938fb\") " pod="openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.769764 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3053ea79-6147-4ff5-9e3c-470b947938fb-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7d5bfb7cdd-nzph6\" (UID: \"3053ea79-6147-4ff5-9e3c-470b947938fb\") " pod="openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.773628 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3053ea79-6147-4ff5-9e3c-470b947938fb-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7d5bfb7cdd-nzph6\" (UID: \"3053ea79-6147-4ff5-9e3c-470b947938fb\") " pod="openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.778737 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3053ea79-6147-4ff5-9e3c-470b947938fb-v4-0-config-system-session\") pod \"oauth-openshift-7d5bfb7cdd-nzph6\" (UID: \"3053ea79-6147-4ff5-9e3c-470b947938fb\") " pod="openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.782461 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96hh2\" (UniqueName: \"kubernetes.io/projected/3053ea79-6147-4ff5-9e3c-470b947938fb-kube-api-access-96hh2\") pod \"oauth-openshift-7d5bfb7cdd-nzph6\" (UID: \"3053ea79-6147-4ff5-9e3c-470b947938fb\") " pod="openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6" Nov 25 15:05:47 crc kubenswrapper[4890]: I1125 15:05:47.899452 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6" Nov 25 15:05:48 crc kubenswrapper[4890]: I1125 15:05:48.133379 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6"] Nov 25 15:05:48 crc kubenswrapper[4890]: I1125 15:05:48.372715 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6" event={"ID":"3053ea79-6147-4ff5-9e3c-470b947938fb","Type":"ContainerStarted","Data":"8bdbdb2c2975f3383c5dfef38a1d0d182e1062bd39ce4f20a10b79fe0ef22411"} Nov 25 15:05:48 crc kubenswrapper[4890]: I1125 15:05:48.374998 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" event={"ID":"a27aab8b-a74f-4a62-b5a3-27d8a7d46be9","Type":"ContainerDied","Data":"91dd36a079d7a52a961a4d8623a3e71a5f18069a9b6943b9349e886f2069b95f"} Nov 25 15:05:48 crc kubenswrapper[4890]: I1125 15:05:48.375046 4890 scope.go:117] "RemoveContainer" containerID="62c26f47dd493084e7a893c76b6c6c564775542abea1bc2a7bd94c95c34e92bf" Nov 25 15:05:48 crc kubenswrapper[4890]: I1125 15:05:48.375130 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-mmhbh" Nov 25 15:05:48 crc kubenswrapper[4890]: I1125 15:05:48.399500 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mmhbh"] Nov 25 15:05:48 crc kubenswrapper[4890]: I1125 15:05:48.404322 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mmhbh"] Nov 25 15:05:49 crc kubenswrapper[4890]: I1125 15:05:49.382010 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6" event={"ID":"3053ea79-6147-4ff5-9e3c-470b947938fb","Type":"ContainerStarted","Data":"8d5502008bacad5611e3761ae20e1ac7dede134fee006f83044fde6d297b742f"} Nov 25 15:05:49 crc kubenswrapper[4890]: I1125 15:05:49.382437 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6" Nov 25 15:05:49 crc kubenswrapper[4890]: I1125 15:05:49.390147 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6" Nov 25 15:05:49 crc kubenswrapper[4890]: I1125 15:05:49.415644 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-7d5bfb7cdd-nzph6" podStartSLOduration=27.415621352 podStartE2EDuration="27.415621352s" podCreationTimestamp="2025-11-25 15:05:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:05:49.41230719 +0000 UTC m=+207.854769810" watchObservedRunningTime="2025-11-25 15:05:49.415621352 +0000 UTC m=+207.858083982" Nov 25 15:05:50 crc kubenswrapper[4890]: I1125 15:05:50.183315 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a27aab8b-a74f-4a62-b5a3-27d8a7d46be9" path="/var/lib/kubelet/pods/a27aab8b-a74f-4a62-b5a3-27d8a7d46be9/volumes" Nov 25 15:05:56 crc kubenswrapper[4890]: I1125 15:05:56.447729 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 15:05:56 crc kubenswrapper[4890]: I1125 15:05:56.448427 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 15:05:56 crc kubenswrapper[4890]: I1125 15:05:56.448488 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" Nov 25 15:05:56 crc kubenswrapper[4890]: I1125 15:05:56.449100 4890 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0902d6f8e56aecc5a2a13a7148f94a8a00ad139027b3c036aba0215ba0a67f59"} pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 15:05:56 crc kubenswrapper[4890]: I1125 15:05:56.449153 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" containerID="cri-o://0902d6f8e56aecc5a2a13a7148f94a8a00ad139027b3c036aba0215ba0a67f59" gracePeriod=600 Nov 25 15:05:57 crc kubenswrapper[4890]: I1125 15:05:57.429057 4890 generic.go:334] "Generic (PLEG): container finished" podID="4e4f849d-f239-4727-a73e-18327856929a" containerID="0902d6f8e56aecc5a2a13a7148f94a8a00ad139027b3c036aba0215ba0a67f59" exitCode=0 Nov 25 15:05:57 crc kubenswrapper[4890]: I1125 15:05:57.429278 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" event={"ID":"4e4f849d-f239-4727-a73e-18327856929a","Type":"ContainerDied","Data":"0902d6f8e56aecc5a2a13a7148f94a8a00ad139027b3c036aba0215ba0a67f59"} Nov 25 15:05:57 crc kubenswrapper[4890]: I1125 15:05:57.429491 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" event={"ID":"4e4f849d-f239-4727-a73e-18327856929a","Type":"ContainerStarted","Data":"2f432adbd81e026db39d6c93118322aceca51fcd04a65e4110d7e407d22f0ff0"} Nov 25 15:06:00 crc kubenswrapper[4890]: I1125 15:06:00.925988 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-z4b8p"] Nov 25 15:06:00 crc kubenswrapper[4890]: I1125 15:06:00.926735 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-z4b8p" podUID="b1983120-f62d-46b0-833e-db30a81973c1" containerName="registry-server" containerID="cri-o://59347c6caa2d3aff1243c63d5b531762bea390ab80bcad117cda132ae2a5f867" gracePeriod=30 Nov 25 15:06:00 crc kubenswrapper[4890]: I1125 15:06:00.936681 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tfnr6"] Nov 25 15:06:00 crc kubenswrapper[4890]: I1125 15:06:00.936966 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tfnr6" podUID="8f052eff-1a42-4c20-8644-d575fc557060" containerName="registry-server" containerID="cri-o://e1591e78261809d129766d029d60ca310164e994d53c83ec1c55c3ca7e263a85" gracePeriod=30 Nov 25 15:06:00 crc kubenswrapper[4890]: I1125 15:06:00.947084 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xmhq9"] Nov 25 15:06:00 crc kubenswrapper[4890]: I1125 15:06:00.947648 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-xmhq9" podUID="8ef9a0ab-2131-42ff-823d-17d517929e77" containerName="marketplace-operator" containerID="cri-o://dca75b5378b4c5d924995da03d2d3dd4b2ea0a61457719ab5b6cd93de1c5349e" gracePeriod=30 Nov 25 15:06:00 crc kubenswrapper[4890]: I1125 15:06:00.978133 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hn24q"] Nov 25 15:06:00 crc kubenswrapper[4890]: I1125 15:06:00.978668 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hn24q" podUID="7fbb3062-51d8-4fc0-9b25-5336de75d692" containerName="registry-server" containerID="cri-o://b14a44d27ed55bfd08a7795cc6c29e7e0ab52f3b0a5f017adde77b9bf4d06d8d" gracePeriod=30 Nov 25 15:06:00 crc kubenswrapper[4890]: I1125 15:06:00.985956 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vf44g"] Nov 25 15:06:00 crc kubenswrapper[4890]: I1125 15:06:00.986354 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vf44g" podUID="22b61785-937e-4562-8c48-5e3d390f3389" containerName="registry-server" containerID="cri-o://1ad675f01e648f268e022e0e561a44eeac06d708344116d1e20e8f72fb0d5a82" gracePeriod=30 Nov 25 15:06:00 crc kubenswrapper[4890]: I1125 15:06:00.988916 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2xzsf"] Nov 25 15:06:00 crc kubenswrapper[4890]: I1125 15:06:00.989743 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-2xzsf" Nov 25 15:06:00 crc kubenswrapper[4890]: I1125 15:06:00.993653 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2xzsf"] Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.138995 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qclbj\" (UniqueName: \"kubernetes.io/projected/4bbe88e6-b859-452a-baf1-8a963b475014-kube-api-access-qclbj\") pod \"marketplace-operator-79b997595-2xzsf\" (UID: \"4bbe88e6-b859-452a-baf1-8a963b475014\") " pod="openshift-marketplace/marketplace-operator-79b997595-2xzsf" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.139076 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4bbe88e6-b859-452a-baf1-8a963b475014-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-2xzsf\" (UID: \"4bbe88e6-b859-452a-baf1-8a963b475014\") " pod="openshift-marketplace/marketplace-operator-79b997595-2xzsf" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.139140 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4bbe88e6-b859-452a-baf1-8a963b475014-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-2xzsf\" (UID: \"4bbe88e6-b859-452a-baf1-8a963b475014\") " pod="openshift-marketplace/marketplace-operator-79b997595-2xzsf" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.240940 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qclbj\" (UniqueName: \"kubernetes.io/projected/4bbe88e6-b859-452a-baf1-8a963b475014-kube-api-access-qclbj\") pod \"marketplace-operator-79b997595-2xzsf\" (UID: \"4bbe88e6-b859-452a-baf1-8a963b475014\") " pod="openshift-marketplace/marketplace-operator-79b997595-2xzsf" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.241325 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4bbe88e6-b859-452a-baf1-8a963b475014-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-2xzsf\" (UID: \"4bbe88e6-b859-452a-baf1-8a963b475014\") " pod="openshift-marketplace/marketplace-operator-79b997595-2xzsf" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.241395 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4bbe88e6-b859-452a-baf1-8a963b475014-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-2xzsf\" (UID: \"4bbe88e6-b859-452a-baf1-8a963b475014\") " pod="openshift-marketplace/marketplace-operator-79b997595-2xzsf" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.242656 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4bbe88e6-b859-452a-baf1-8a963b475014-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-2xzsf\" (UID: \"4bbe88e6-b859-452a-baf1-8a963b475014\") " pod="openshift-marketplace/marketplace-operator-79b997595-2xzsf" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.247864 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4bbe88e6-b859-452a-baf1-8a963b475014-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-2xzsf\" (UID: \"4bbe88e6-b859-452a-baf1-8a963b475014\") " pod="openshift-marketplace/marketplace-operator-79b997595-2xzsf" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.261888 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qclbj\" (UniqueName: \"kubernetes.io/projected/4bbe88e6-b859-452a-baf1-8a963b475014-kube-api-access-qclbj\") pod \"marketplace-operator-79b997595-2xzsf\" (UID: \"4bbe88e6-b859-452a-baf1-8a963b475014\") " pod="openshift-marketplace/marketplace-operator-79b997595-2xzsf" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.363931 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-2xzsf" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.376915 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xmhq9" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.377961 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z4b8p" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.380239 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tfnr6" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.405331 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vf44g" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.421173 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hn24q" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.442914 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8ef9a0ab-2131-42ff-823d-17d517929e77-marketplace-operator-metrics\") pod \"8ef9a0ab-2131-42ff-823d-17d517929e77\" (UID: \"8ef9a0ab-2131-42ff-823d-17d517929e77\") " Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.443021 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wvnfh\" (UniqueName: \"kubernetes.io/projected/8ef9a0ab-2131-42ff-823d-17d517929e77-kube-api-access-wvnfh\") pod \"8ef9a0ab-2131-42ff-823d-17d517929e77\" (UID: \"8ef9a0ab-2131-42ff-823d-17d517929e77\") " Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.443058 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8ef9a0ab-2131-42ff-823d-17d517929e77-marketplace-trusted-ca\") pod \"8ef9a0ab-2131-42ff-823d-17d517929e77\" (UID: \"8ef9a0ab-2131-42ff-823d-17d517929e77\") " Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.443153 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1983120-f62d-46b0-833e-db30a81973c1-utilities\") pod \"b1983120-f62d-46b0-833e-db30a81973c1\" (UID: \"b1983120-f62d-46b0-833e-db30a81973c1\") " Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.443336 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2drz\" (UniqueName: \"kubernetes.io/projected/b1983120-f62d-46b0-833e-db30a81973c1-kube-api-access-x2drz\") pod \"b1983120-f62d-46b0-833e-db30a81973c1\" (UID: \"b1983120-f62d-46b0-833e-db30a81973c1\") " Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.443380 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1983120-f62d-46b0-833e-db30a81973c1-catalog-content\") pod \"b1983120-f62d-46b0-833e-db30a81973c1\" (UID: \"b1983120-f62d-46b0-833e-db30a81973c1\") " Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.444253 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ef9a0ab-2131-42ff-823d-17d517929e77-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "8ef9a0ab-2131-42ff-823d-17d517929e77" (UID: "8ef9a0ab-2131-42ff-823d-17d517929e77"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.444431 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1983120-f62d-46b0-833e-db30a81973c1-utilities" (OuterVolumeSpecName: "utilities") pod "b1983120-f62d-46b0-833e-db30a81973c1" (UID: "b1983120-f62d-46b0-833e-db30a81973c1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.448844 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ef9a0ab-2131-42ff-823d-17d517929e77-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "8ef9a0ab-2131-42ff-823d-17d517929e77" (UID: "8ef9a0ab-2131-42ff-823d-17d517929e77"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.450394 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1983120-f62d-46b0-833e-db30a81973c1-kube-api-access-x2drz" (OuterVolumeSpecName: "kube-api-access-x2drz") pod "b1983120-f62d-46b0-833e-db30a81973c1" (UID: "b1983120-f62d-46b0-833e-db30a81973c1"). InnerVolumeSpecName "kube-api-access-x2drz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.452549 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ef9a0ab-2131-42ff-823d-17d517929e77-kube-api-access-wvnfh" (OuterVolumeSpecName: "kube-api-access-wvnfh") pod "8ef9a0ab-2131-42ff-823d-17d517929e77" (UID: "8ef9a0ab-2131-42ff-823d-17d517929e77"). InnerVolumeSpecName "kube-api-access-wvnfh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.455847 4890 generic.go:334] "Generic (PLEG): container finished" podID="8ef9a0ab-2131-42ff-823d-17d517929e77" containerID="dca75b5378b4c5d924995da03d2d3dd4b2ea0a61457719ab5b6cd93de1c5349e" exitCode=0 Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.455929 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xmhq9" event={"ID":"8ef9a0ab-2131-42ff-823d-17d517929e77","Type":"ContainerDied","Data":"dca75b5378b4c5d924995da03d2d3dd4b2ea0a61457719ab5b6cd93de1c5349e"} Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.455964 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xmhq9" event={"ID":"8ef9a0ab-2131-42ff-823d-17d517929e77","Type":"ContainerDied","Data":"d36a06d3be2645d92c3e0dac63b8d50beedb8861e77d4730d26ad04a620e0028"} Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.455983 4890 scope.go:117] "RemoveContainer" containerID="dca75b5378b4c5d924995da03d2d3dd4b2ea0a61457719ab5b6cd93de1c5349e" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.456114 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xmhq9" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.487573 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hn24q" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.487547 4890 generic.go:334] "Generic (PLEG): container finished" podID="7fbb3062-51d8-4fc0-9b25-5336de75d692" containerID="b14a44d27ed55bfd08a7795cc6c29e7e0ab52f3b0a5f017adde77b9bf4d06d8d" exitCode=0 Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.487860 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hn24q" event={"ID":"7fbb3062-51d8-4fc0-9b25-5336de75d692","Type":"ContainerDied","Data":"b14a44d27ed55bfd08a7795cc6c29e7e0ab52f3b0a5f017adde77b9bf4d06d8d"} Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.488734 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hn24q" event={"ID":"7fbb3062-51d8-4fc0-9b25-5336de75d692","Type":"ContainerDied","Data":"7f868b7bd6132c1baff4a2bc246ad9bfbb5e1054ca25426be1b0e889f197a1e6"} Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.497998 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xmhq9"] Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.500411 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xmhq9"] Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.500569 4890 scope.go:117] "RemoveContainer" containerID="dca75b5378b4c5d924995da03d2d3dd4b2ea0a61457719ab5b6cd93de1c5349e" Nov 25 15:06:01 crc kubenswrapper[4890]: E1125 15:06:01.501297 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dca75b5378b4c5d924995da03d2d3dd4b2ea0a61457719ab5b6cd93de1c5349e\": container with ID starting with dca75b5378b4c5d924995da03d2d3dd4b2ea0a61457719ab5b6cd93de1c5349e not found: ID does not exist" containerID="dca75b5378b4c5d924995da03d2d3dd4b2ea0a61457719ab5b6cd93de1c5349e" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.501331 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dca75b5378b4c5d924995da03d2d3dd4b2ea0a61457719ab5b6cd93de1c5349e"} err="failed to get container status \"dca75b5378b4c5d924995da03d2d3dd4b2ea0a61457719ab5b6cd93de1c5349e\": rpc error: code = NotFound desc = could not find container \"dca75b5378b4c5d924995da03d2d3dd4b2ea0a61457719ab5b6cd93de1c5349e\": container with ID starting with dca75b5378b4c5d924995da03d2d3dd4b2ea0a61457719ab5b6cd93de1c5349e not found: ID does not exist" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.501367 4890 scope.go:117] "RemoveContainer" containerID="b14a44d27ed55bfd08a7795cc6c29e7e0ab52f3b0a5f017adde77b9bf4d06d8d" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.501715 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z4b8p" event={"ID":"b1983120-f62d-46b0-833e-db30a81973c1","Type":"ContainerDied","Data":"59347c6caa2d3aff1243c63d5b531762bea390ab80bcad117cda132ae2a5f867"} Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.501731 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z4b8p" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.501800 4890 generic.go:334] "Generic (PLEG): container finished" podID="b1983120-f62d-46b0-833e-db30a81973c1" containerID="59347c6caa2d3aff1243c63d5b531762bea390ab80bcad117cda132ae2a5f867" exitCode=0 Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.501975 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z4b8p" event={"ID":"b1983120-f62d-46b0-833e-db30a81973c1","Type":"ContainerDied","Data":"8f7dc004940b2f5161509321d60d64ff691d261bd46b885a6de2a8b2d887d5a8"} Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.516132 4890 generic.go:334] "Generic (PLEG): container finished" podID="8f052eff-1a42-4c20-8644-d575fc557060" containerID="e1591e78261809d129766d029d60ca310164e994d53c83ec1c55c3ca7e263a85" exitCode=0 Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.516204 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tfnr6" event={"ID":"8f052eff-1a42-4c20-8644-d575fc557060","Type":"ContainerDied","Data":"e1591e78261809d129766d029d60ca310164e994d53c83ec1c55c3ca7e263a85"} Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.516233 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tfnr6" event={"ID":"8f052eff-1a42-4c20-8644-d575fc557060","Type":"ContainerDied","Data":"a962d0961c21b943f81dba768a61952a6f35aad61ced29a057a068240353946d"} Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.516294 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tfnr6" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.519851 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1983120-f62d-46b0-833e-db30a81973c1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b1983120-f62d-46b0-833e-db30a81973c1" (UID: "b1983120-f62d-46b0-833e-db30a81973c1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.523021 4890 generic.go:334] "Generic (PLEG): container finished" podID="22b61785-937e-4562-8c48-5e3d390f3389" containerID="1ad675f01e648f268e022e0e561a44eeac06d708344116d1e20e8f72fb0d5a82" exitCode=0 Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.523062 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vf44g" event={"ID":"22b61785-937e-4562-8c48-5e3d390f3389","Type":"ContainerDied","Data":"1ad675f01e648f268e022e0e561a44eeac06d708344116d1e20e8f72fb0d5a82"} Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.523091 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vf44g" event={"ID":"22b61785-937e-4562-8c48-5e3d390f3389","Type":"ContainerDied","Data":"189fdd9f5e09b7cd03c235fd64aa8fec639f58a6e2e643d6066b2d5882d8d914"} Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.523112 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vf44g" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.524941 4890 scope.go:117] "RemoveContainer" containerID="62a36226c66383b36f4eda2fe0aa72ab055026ff30490aef7a4b72c4e7610471" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.545672 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mkwzk\" (UniqueName: \"kubernetes.io/projected/7fbb3062-51d8-4fc0-9b25-5336de75d692-kube-api-access-mkwzk\") pod \"7fbb3062-51d8-4fc0-9b25-5336de75d692\" (UID: \"7fbb3062-51d8-4fc0-9b25-5336de75d692\") " Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.545710 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fbb3062-51d8-4fc0-9b25-5336de75d692-catalog-content\") pod \"7fbb3062-51d8-4fc0-9b25-5336de75d692\" (UID: \"7fbb3062-51d8-4fc0-9b25-5336de75d692\") " Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.545745 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22b61785-937e-4562-8c48-5e3d390f3389-catalog-content\") pod \"22b61785-937e-4562-8c48-5e3d390f3389\" (UID: \"22b61785-937e-4562-8c48-5e3d390f3389\") " Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.545791 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmphq\" (UniqueName: \"kubernetes.io/projected/22b61785-937e-4562-8c48-5e3d390f3389-kube-api-access-vmphq\") pod \"22b61785-937e-4562-8c48-5e3d390f3389\" (UID: \"22b61785-937e-4562-8c48-5e3d390f3389\") " Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.545817 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f052eff-1a42-4c20-8644-d575fc557060-catalog-content\") pod \"8f052eff-1a42-4c20-8644-d575fc557060\" (UID: \"8f052eff-1a42-4c20-8644-d575fc557060\") " Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.545840 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mlz6n\" (UniqueName: \"kubernetes.io/projected/8f052eff-1a42-4c20-8644-d575fc557060-kube-api-access-mlz6n\") pod \"8f052eff-1a42-4c20-8644-d575fc557060\" (UID: \"8f052eff-1a42-4c20-8644-d575fc557060\") " Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.545889 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f052eff-1a42-4c20-8644-d575fc557060-utilities\") pod \"8f052eff-1a42-4c20-8644-d575fc557060\" (UID: \"8f052eff-1a42-4c20-8644-d575fc557060\") " Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.545915 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fbb3062-51d8-4fc0-9b25-5336de75d692-utilities\") pod \"7fbb3062-51d8-4fc0-9b25-5336de75d692\" (UID: \"7fbb3062-51d8-4fc0-9b25-5336de75d692\") " Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.545966 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22b61785-937e-4562-8c48-5e3d390f3389-utilities\") pod \"22b61785-937e-4562-8c48-5e3d390f3389\" (UID: \"22b61785-937e-4562-8c48-5e3d390f3389\") " Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.546321 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1983120-f62d-46b0-833e-db30a81973c1-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.546339 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2drz\" (UniqueName: \"kubernetes.io/projected/b1983120-f62d-46b0-833e-db30a81973c1-kube-api-access-x2drz\") on node \"crc\" DevicePath \"\"" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.546350 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1983120-f62d-46b0-833e-db30a81973c1-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.546359 4890 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8ef9a0ab-2131-42ff-823d-17d517929e77-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.546370 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wvnfh\" (UniqueName: \"kubernetes.io/projected/8ef9a0ab-2131-42ff-823d-17d517929e77-kube-api-access-wvnfh\") on node \"crc\" DevicePath \"\"" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.546379 4890 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8ef9a0ab-2131-42ff-823d-17d517929e77-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.547013 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22b61785-937e-4562-8c48-5e3d390f3389-utilities" (OuterVolumeSpecName: "utilities") pod "22b61785-937e-4562-8c48-5e3d390f3389" (UID: "22b61785-937e-4562-8c48-5e3d390f3389"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.549494 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f052eff-1a42-4c20-8644-d575fc557060-utilities" (OuterVolumeSpecName: "utilities") pod "8f052eff-1a42-4c20-8644-d575fc557060" (UID: "8f052eff-1a42-4c20-8644-d575fc557060"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.550559 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22b61785-937e-4562-8c48-5e3d390f3389-kube-api-access-vmphq" (OuterVolumeSpecName: "kube-api-access-vmphq") pod "22b61785-937e-4562-8c48-5e3d390f3389" (UID: "22b61785-937e-4562-8c48-5e3d390f3389"). InnerVolumeSpecName "kube-api-access-vmphq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.550829 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fbb3062-51d8-4fc0-9b25-5336de75d692-kube-api-access-mkwzk" (OuterVolumeSpecName: "kube-api-access-mkwzk") pod "7fbb3062-51d8-4fc0-9b25-5336de75d692" (UID: "7fbb3062-51d8-4fc0-9b25-5336de75d692"). InnerVolumeSpecName "kube-api-access-mkwzk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.551659 4890 scope.go:117] "RemoveContainer" containerID="f7e0aec046f1f257d4a70d708669775c3cca407ce449e708d87b0062817414b0" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.576211 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f052eff-1a42-4c20-8644-d575fc557060-kube-api-access-mlz6n" (OuterVolumeSpecName: "kube-api-access-mlz6n") pod "8f052eff-1a42-4c20-8644-d575fc557060" (UID: "8f052eff-1a42-4c20-8644-d575fc557060"). InnerVolumeSpecName "kube-api-access-mlz6n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.586028 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fbb3062-51d8-4fc0-9b25-5336de75d692-utilities" (OuterVolumeSpecName: "utilities") pod "7fbb3062-51d8-4fc0-9b25-5336de75d692" (UID: "7fbb3062-51d8-4fc0-9b25-5336de75d692"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.589060 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fbb3062-51d8-4fc0-9b25-5336de75d692-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7fbb3062-51d8-4fc0-9b25-5336de75d692" (UID: "7fbb3062-51d8-4fc0-9b25-5336de75d692"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.593264 4890 scope.go:117] "RemoveContainer" containerID="b14a44d27ed55bfd08a7795cc6c29e7e0ab52f3b0a5f017adde77b9bf4d06d8d" Nov 25 15:06:01 crc kubenswrapper[4890]: E1125 15:06:01.593832 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b14a44d27ed55bfd08a7795cc6c29e7e0ab52f3b0a5f017adde77b9bf4d06d8d\": container with ID starting with b14a44d27ed55bfd08a7795cc6c29e7e0ab52f3b0a5f017adde77b9bf4d06d8d not found: ID does not exist" containerID="b14a44d27ed55bfd08a7795cc6c29e7e0ab52f3b0a5f017adde77b9bf4d06d8d" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.593894 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b14a44d27ed55bfd08a7795cc6c29e7e0ab52f3b0a5f017adde77b9bf4d06d8d"} err="failed to get container status \"b14a44d27ed55bfd08a7795cc6c29e7e0ab52f3b0a5f017adde77b9bf4d06d8d\": rpc error: code = NotFound desc = could not find container \"b14a44d27ed55bfd08a7795cc6c29e7e0ab52f3b0a5f017adde77b9bf4d06d8d\": container with ID starting with b14a44d27ed55bfd08a7795cc6c29e7e0ab52f3b0a5f017adde77b9bf4d06d8d not found: ID does not exist" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.593958 4890 scope.go:117] "RemoveContainer" containerID="62a36226c66383b36f4eda2fe0aa72ab055026ff30490aef7a4b72c4e7610471" Nov 25 15:06:01 crc kubenswrapper[4890]: E1125 15:06:01.594505 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62a36226c66383b36f4eda2fe0aa72ab055026ff30490aef7a4b72c4e7610471\": container with ID starting with 62a36226c66383b36f4eda2fe0aa72ab055026ff30490aef7a4b72c4e7610471 not found: ID does not exist" containerID="62a36226c66383b36f4eda2fe0aa72ab055026ff30490aef7a4b72c4e7610471" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.594541 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62a36226c66383b36f4eda2fe0aa72ab055026ff30490aef7a4b72c4e7610471"} err="failed to get container status \"62a36226c66383b36f4eda2fe0aa72ab055026ff30490aef7a4b72c4e7610471\": rpc error: code = NotFound desc = could not find container \"62a36226c66383b36f4eda2fe0aa72ab055026ff30490aef7a4b72c4e7610471\": container with ID starting with 62a36226c66383b36f4eda2fe0aa72ab055026ff30490aef7a4b72c4e7610471 not found: ID does not exist" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.594564 4890 scope.go:117] "RemoveContainer" containerID="f7e0aec046f1f257d4a70d708669775c3cca407ce449e708d87b0062817414b0" Nov 25 15:06:01 crc kubenswrapper[4890]: E1125 15:06:01.595456 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7e0aec046f1f257d4a70d708669775c3cca407ce449e708d87b0062817414b0\": container with ID starting with f7e0aec046f1f257d4a70d708669775c3cca407ce449e708d87b0062817414b0 not found: ID does not exist" containerID="f7e0aec046f1f257d4a70d708669775c3cca407ce449e708d87b0062817414b0" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.595492 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7e0aec046f1f257d4a70d708669775c3cca407ce449e708d87b0062817414b0"} err="failed to get container status \"f7e0aec046f1f257d4a70d708669775c3cca407ce449e708d87b0062817414b0\": rpc error: code = NotFound desc = could not find container \"f7e0aec046f1f257d4a70d708669775c3cca407ce449e708d87b0062817414b0\": container with ID starting with f7e0aec046f1f257d4a70d708669775c3cca407ce449e708d87b0062817414b0 not found: ID does not exist" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.595516 4890 scope.go:117] "RemoveContainer" containerID="59347c6caa2d3aff1243c63d5b531762bea390ab80bcad117cda132ae2a5f867" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.612835 4890 scope.go:117] "RemoveContainer" containerID="317401cc98b8211f11251559a4e5bf87764469a03304557c3931d7c9d3f1c13f" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.629757 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f052eff-1a42-4c20-8644-d575fc557060-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8f052eff-1a42-4c20-8644-d575fc557060" (UID: "8f052eff-1a42-4c20-8644-d575fc557060"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.636500 4890 scope.go:117] "RemoveContainer" containerID="0ed3bd68e3b6f9f33a2a5c10e9c36a0b8bb1426015599ef33853a89b7932a57a" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.640188 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2xzsf"] Nov 25 15:06:01 crc kubenswrapper[4890]: W1125 15:06:01.646854 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4bbe88e6_b859_452a_baf1_8a963b475014.slice/crio-89d414ada7a78f155e2c2aeaae587d2d97d44344dd9cd013c8837d3bd6849428 WatchSource:0}: Error finding container 89d414ada7a78f155e2c2aeaae587d2d97d44344dd9cd013c8837d3bd6849428: Status 404 returned error can't find the container with id 89d414ada7a78f155e2c2aeaae587d2d97d44344dd9cd013c8837d3bd6849428 Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.647274 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmphq\" (UniqueName: \"kubernetes.io/projected/22b61785-937e-4562-8c48-5e3d390f3389-kube-api-access-vmphq\") on node \"crc\" DevicePath \"\"" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.647301 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f052eff-1a42-4c20-8644-d575fc557060-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.647311 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mlz6n\" (UniqueName: \"kubernetes.io/projected/8f052eff-1a42-4c20-8644-d575fc557060-kube-api-access-mlz6n\") on node \"crc\" DevicePath \"\"" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.647320 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f052eff-1a42-4c20-8644-d575fc557060-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.647328 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fbb3062-51d8-4fc0-9b25-5336de75d692-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.647337 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22b61785-937e-4562-8c48-5e3d390f3389-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.647346 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mkwzk\" (UniqueName: \"kubernetes.io/projected/7fbb3062-51d8-4fc0-9b25-5336de75d692-kube-api-access-mkwzk\") on node \"crc\" DevicePath \"\"" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.647355 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fbb3062-51d8-4fc0-9b25-5336de75d692-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.653715 4890 scope.go:117] "RemoveContainer" containerID="59347c6caa2d3aff1243c63d5b531762bea390ab80bcad117cda132ae2a5f867" Nov 25 15:06:01 crc kubenswrapper[4890]: E1125 15:06:01.654174 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59347c6caa2d3aff1243c63d5b531762bea390ab80bcad117cda132ae2a5f867\": container with ID starting with 59347c6caa2d3aff1243c63d5b531762bea390ab80bcad117cda132ae2a5f867 not found: ID does not exist" containerID="59347c6caa2d3aff1243c63d5b531762bea390ab80bcad117cda132ae2a5f867" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.654223 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59347c6caa2d3aff1243c63d5b531762bea390ab80bcad117cda132ae2a5f867"} err="failed to get container status \"59347c6caa2d3aff1243c63d5b531762bea390ab80bcad117cda132ae2a5f867\": rpc error: code = NotFound desc = could not find container \"59347c6caa2d3aff1243c63d5b531762bea390ab80bcad117cda132ae2a5f867\": container with ID starting with 59347c6caa2d3aff1243c63d5b531762bea390ab80bcad117cda132ae2a5f867 not found: ID does not exist" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.654279 4890 scope.go:117] "RemoveContainer" containerID="317401cc98b8211f11251559a4e5bf87764469a03304557c3931d7c9d3f1c13f" Nov 25 15:06:01 crc kubenswrapper[4890]: E1125 15:06:01.654964 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"317401cc98b8211f11251559a4e5bf87764469a03304557c3931d7c9d3f1c13f\": container with ID starting with 317401cc98b8211f11251559a4e5bf87764469a03304557c3931d7c9d3f1c13f not found: ID does not exist" containerID="317401cc98b8211f11251559a4e5bf87764469a03304557c3931d7c9d3f1c13f" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.654994 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"317401cc98b8211f11251559a4e5bf87764469a03304557c3931d7c9d3f1c13f"} err="failed to get container status \"317401cc98b8211f11251559a4e5bf87764469a03304557c3931d7c9d3f1c13f\": rpc error: code = NotFound desc = could not find container \"317401cc98b8211f11251559a4e5bf87764469a03304557c3931d7c9d3f1c13f\": container with ID starting with 317401cc98b8211f11251559a4e5bf87764469a03304557c3931d7c9d3f1c13f not found: ID does not exist" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.655016 4890 scope.go:117] "RemoveContainer" containerID="0ed3bd68e3b6f9f33a2a5c10e9c36a0b8bb1426015599ef33853a89b7932a57a" Nov 25 15:06:01 crc kubenswrapper[4890]: E1125 15:06:01.659522 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ed3bd68e3b6f9f33a2a5c10e9c36a0b8bb1426015599ef33853a89b7932a57a\": container with ID starting with 0ed3bd68e3b6f9f33a2a5c10e9c36a0b8bb1426015599ef33853a89b7932a57a not found: ID does not exist" containerID="0ed3bd68e3b6f9f33a2a5c10e9c36a0b8bb1426015599ef33853a89b7932a57a" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.659569 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ed3bd68e3b6f9f33a2a5c10e9c36a0b8bb1426015599ef33853a89b7932a57a"} err="failed to get container status \"0ed3bd68e3b6f9f33a2a5c10e9c36a0b8bb1426015599ef33853a89b7932a57a\": rpc error: code = NotFound desc = could not find container \"0ed3bd68e3b6f9f33a2a5c10e9c36a0b8bb1426015599ef33853a89b7932a57a\": container with ID starting with 0ed3bd68e3b6f9f33a2a5c10e9c36a0b8bb1426015599ef33853a89b7932a57a not found: ID does not exist" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.659600 4890 scope.go:117] "RemoveContainer" containerID="e1591e78261809d129766d029d60ca310164e994d53c83ec1c55c3ca7e263a85" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.664708 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22b61785-937e-4562-8c48-5e3d390f3389-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "22b61785-937e-4562-8c48-5e3d390f3389" (UID: "22b61785-937e-4562-8c48-5e3d390f3389"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.678579 4890 scope.go:117] "RemoveContainer" containerID="97823a3f2d4dcc99f7b1980839e5e94fc7b18d857a99e776c43aa11d1c9e7c36" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.692327 4890 scope.go:117] "RemoveContainer" containerID="9c61813686fa358cdacad333b64a1c4530ac05fa7b48bcdd7491621c9e449f32" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.714515 4890 scope.go:117] "RemoveContainer" containerID="e1591e78261809d129766d029d60ca310164e994d53c83ec1c55c3ca7e263a85" Nov 25 15:06:01 crc kubenswrapper[4890]: E1125 15:06:01.715131 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1591e78261809d129766d029d60ca310164e994d53c83ec1c55c3ca7e263a85\": container with ID starting with e1591e78261809d129766d029d60ca310164e994d53c83ec1c55c3ca7e263a85 not found: ID does not exist" containerID="e1591e78261809d129766d029d60ca310164e994d53c83ec1c55c3ca7e263a85" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.715195 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1591e78261809d129766d029d60ca310164e994d53c83ec1c55c3ca7e263a85"} err="failed to get container status \"e1591e78261809d129766d029d60ca310164e994d53c83ec1c55c3ca7e263a85\": rpc error: code = NotFound desc = could not find container \"e1591e78261809d129766d029d60ca310164e994d53c83ec1c55c3ca7e263a85\": container with ID starting with e1591e78261809d129766d029d60ca310164e994d53c83ec1c55c3ca7e263a85 not found: ID does not exist" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.715237 4890 scope.go:117] "RemoveContainer" containerID="97823a3f2d4dcc99f7b1980839e5e94fc7b18d857a99e776c43aa11d1c9e7c36" Nov 25 15:06:01 crc kubenswrapper[4890]: E1125 15:06:01.715649 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97823a3f2d4dcc99f7b1980839e5e94fc7b18d857a99e776c43aa11d1c9e7c36\": container with ID starting with 97823a3f2d4dcc99f7b1980839e5e94fc7b18d857a99e776c43aa11d1c9e7c36 not found: ID does not exist" containerID="97823a3f2d4dcc99f7b1980839e5e94fc7b18d857a99e776c43aa11d1c9e7c36" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.715689 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97823a3f2d4dcc99f7b1980839e5e94fc7b18d857a99e776c43aa11d1c9e7c36"} err="failed to get container status \"97823a3f2d4dcc99f7b1980839e5e94fc7b18d857a99e776c43aa11d1c9e7c36\": rpc error: code = NotFound desc = could not find container \"97823a3f2d4dcc99f7b1980839e5e94fc7b18d857a99e776c43aa11d1c9e7c36\": container with ID starting with 97823a3f2d4dcc99f7b1980839e5e94fc7b18d857a99e776c43aa11d1c9e7c36 not found: ID does not exist" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.715720 4890 scope.go:117] "RemoveContainer" containerID="9c61813686fa358cdacad333b64a1c4530ac05fa7b48bcdd7491621c9e449f32" Nov 25 15:06:01 crc kubenswrapper[4890]: E1125 15:06:01.716198 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c61813686fa358cdacad333b64a1c4530ac05fa7b48bcdd7491621c9e449f32\": container with ID starting with 9c61813686fa358cdacad333b64a1c4530ac05fa7b48bcdd7491621c9e449f32 not found: ID does not exist" containerID="9c61813686fa358cdacad333b64a1c4530ac05fa7b48bcdd7491621c9e449f32" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.716230 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c61813686fa358cdacad333b64a1c4530ac05fa7b48bcdd7491621c9e449f32"} err="failed to get container status \"9c61813686fa358cdacad333b64a1c4530ac05fa7b48bcdd7491621c9e449f32\": rpc error: code = NotFound desc = could not find container \"9c61813686fa358cdacad333b64a1c4530ac05fa7b48bcdd7491621c9e449f32\": container with ID starting with 9c61813686fa358cdacad333b64a1c4530ac05fa7b48bcdd7491621c9e449f32 not found: ID does not exist" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.716249 4890 scope.go:117] "RemoveContainer" containerID="1ad675f01e648f268e022e0e561a44eeac06d708344116d1e20e8f72fb0d5a82" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.743350 4890 scope.go:117] "RemoveContainer" containerID="3dbd44e5beadfc1768f2a52baf3ed9e523cadb8edb361a7cc36b6e0f71266084" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.748675 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22b61785-937e-4562-8c48-5e3d390f3389-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.758432 4890 scope.go:117] "RemoveContainer" containerID="ba7ced8954142349680e569794df37ccd3da2c9b7be0397aee6e5ac06a31693c" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.769371 4890 scope.go:117] "RemoveContainer" containerID="1ad675f01e648f268e022e0e561a44eeac06d708344116d1e20e8f72fb0d5a82" Nov 25 15:06:01 crc kubenswrapper[4890]: E1125 15:06:01.769757 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ad675f01e648f268e022e0e561a44eeac06d708344116d1e20e8f72fb0d5a82\": container with ID starting with 1ad675f01e648f268e022e0e561a44eeac06d708344116d1e20e8f72fb0d5a82 not found: ID does not exist" containerID="1ad675f01e648f268e022e0e561a44eeac06d708344116d1e20e8f72fb0d5a82" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.769800 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ad675f01e648f268e022e0e561a44eeac06d708344116d1e20e8f72fb0d5a82"} err="failed to get container status \"1ad675f01e648f268e022e0e561a44eeac06d708344116d1e20e8f72fb0d5a82\": rpc error: code = NotFound desc = could not find container \"1ad675f01e648f268e022e0e561a44eeac06d708344116d1e20e8f72fb0d5a82\": container with ID starting with 1ad675f01e648f268e022e0e561a44eeac06d708344116d1e20e8f72fb0d5a82 not found: ID does not exist" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.769831 4890 scope.go:117] "RemoveContainer" containerID="3dbd44e5beadfc1768f2a52baf3ed9e523cadb8edb361a7cc36b6e0f71266084" Nov 25 15:06:01 crc kubenswrapper[4890]: E1125 15:06:01.770457 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3dbd44e5beadfc1768f2a52baf3ed9e523cadb8edb361a7cc36b6e0f71266084\": container with ID starting with 3dbd44e5beadfc1768f2a52baf3ed9e523cadb8edb361a7cc36b6e0f71266084 not found: ID does not exist" containerID="3dbd44e5beadfc1768f2a52baf3ed9e523cadb8edb361a7cc36b6e0f71266084" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.770530 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3dbd44e5beadfc1768f2a52baf3ed9e523cadb8edb361a7cc36b6e0f71266084"} err="failed to get container status \"3dbd44e5beadfc1768f2a52baf3ed9e523cadb8edb361a7cc36b6e0f71266084\": rpc error: code = NotFound desc = could not find container \"3dbd44e5beadfc1768f2a52baf3ed9e523cadb8edb361a7cc36b6e0f71266084\": container with ID starting with 3dbd44e5beadfc1768f2a52baf3ed9e523cadb8edb361a7cc36b6e0f71266084 not found: ID does not exist" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.770570 4890 scope.go:117] "RemoveContainer" containerID="ba7ced8954142349680e569794df37ccd3da2c9b7be0397aee6e5ac06a31693c" Nov 25 15:06:01 crc kubenswrapper[4890]: E1125 15:06:01.770916 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba7ced8954142349680e569794df37ccd3da2c9b7be0397aee6e5ac06a31693c\": container with ID starting with ba7ced8954142349680e569794df37ccd3da2c9b7be0397aee6e5ac06a31693c not found: ID does not exist" containerID="ba7ced8954142349680e569794df37ccd3da2c9b7be0397aee6e5ac06a31693c" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.771017 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba7ced8954142349680e569794df37ccd3da2c9b7be0397aee6e5ac06a31693c"} err="failed to get container status \"ba7ced8954142349680e569794df37ccd3da2c9b7be0397aee6e5ac06a31693c\": rpc error: code = NotFound desc = could not find container \"ba7ced8954142349680e569794df37ccd3da2c9b7be0397aee6e5ac06a31693c\": container with ID starting with ba7ced8954142349680e569794df37ccd3da2c9b7be0397aee6e5ac06a31693c not found: ID does not exist" Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.813376 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hn24q"] Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.817729 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hn24q"] Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.828385 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-z4b8p"] Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.839112 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-z4b8p"] Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.854134 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vf44g"] Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.857240 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vf44g"] Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.871543 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tfnr6"] Nov 25 15:06:01 crc kubenswrapper[4890]: I1125 15:06:01.874657 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tfnr6"] Nov 25 15:06:02 crc kubenswrapper[4890]: I1125 15:06:02.181705 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22b61785-937e-4562-8c48-5e3d390f3389" path="/var/lib/kubelet/pods/22b61785-937e-4562-8c48-5e3d390f3389/volumes" Nov 25 15:06:02 crc kubenswrapper[4890]: I1125 15:06:02.182539 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fbb3062-51d8-4fc0-9b25-5336de75d692" path="/var/lib/kubelet/pods/7fbb3062-51d8-4fc0-9b25-5336de75d692/volumes" Nov 25 15:06:02 crc kubenswrapper[4890]: I1125 15:06:02.183352 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ef9a0ab-2131-42ff-823d-17d517929e77" path="/var/lib/kubelet/pods/8ef9a0ab-2131-42ff-823d-17d517929e77/volumes" Nov 25 15:06:02 crc kubenswrapper[4890]: I1125 15:06:02.185469 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f052eff-1a42-4c20-8644-d575fc557060" path="/var/lib/kubelet/pods/8f052eff-1a42-4c20-8644-d575fc557060/volumes" Nov 25 15:06:02 crc kubenswrapper[4890]: I1125 15:06:02.189044 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1983120-f62d-46b0-833e-db30a81973c1" path="/var/lib/kubelet/pods/b1983120-f62d-46b0-833e-db30a81973c1/volumes" Nov 25 15:06:02 crc kubenswrapper[4890]: I1125 15:06:02.529812 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2xzsf" event={"ID":"4bbe88e6-b859-452a-baf1-8a963b475014","Type":"ContainerStarted","Data":"17a73befb091b6d575f44bfa544a3b79cec0c8a8c327edbac49c7aebfe563aae"} Nov 25 15:06:02 crc kubenswrapper[4890]: I1125 15:06:02.530136 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-2xzsf" Nov 25 15:06:02 crc kubenswrapper[4890]: I1125 15:06:02.530150 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2xzsf" event={"ID":"4bbe88e6-b859-452a-baf1-8a963b475014","Type":"ContainerStarted","Data":"89d414ada7a78f155e2c2aeaae587d2d97d44344dd9cd013c8837d3bd6849428"} Nov 25 15:06:02 crc kubenswrapper[4890]: I1125 15:06:02.535987 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-2xzsf" Nov 25 15:06:02 crc kubenswrapper[4890]: I1125 15:06:02.548290 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-2xzsf" podStartSLOduration=2.548271001 podStartE2EDuration="2.548271001s" podCreationTimestamp="2025-11-25 15:06:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:06:02.546546403 +0000 UTC m=+220.989009013" watchObservedRunningTime="2025-11-25 15:06:02.548271001 +0000 UTC m=+220.990733601" Nov 25 15:06:02 crc kubenswrapper[4890]: I1125 15:06:02.734497 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fksn8"] Nov 25 15:06:02 crc kubenswrapper[4890]: E1125 15:06:02.734699 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22b61785-937e-4562-8c48-5e3d390f3389" containerName="extract-content" Nov 25 15:06:02 crc kubenswrapper[4890]: I1125 15:06:02.734710 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="22b61785-937e-4562-8c48-5e3d390f3389" containerName="extract-content" Nov 25 15:06:02 crc kubenswrapper[4890]: E1125 15:06:02.734721 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f052eff-1a42-4c20-8644-d575fc557060" containerName="registry-server" Nov 25 15:06:02 crc kubenswrapper[4890]: I1125 15:06:02.734727 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f052eff-1a42-4c20-8644-d575fc557060" containerName="registry-server" Nov 25 15:06:02 crc kubenswrapper[4890]: E1125 15:06:02.734737 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fbb3062-51d8-4fc0-9b25-5336de75d692" containerName="extract-utilities" Nov 25 15:06:02 crc kubenswrapper[4890]: I1125 15:06:02.734743 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fbb3062-51d8-4fc0-9b25-5336de75d692" containerName="extract-utilities" Nov 25 15:06:02 crc kubenswrapper[4890]: E1125 15:06:02.734751 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ef9a0ab-2131-42ff-823d-17d517929e77" containerName="marketplace-operator" Nov 25 15:06:02 crc kubenswrapper[4890]: I1125 15:06:02.734757 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ef9a0ab-2131-42ff-823d-17d517929e77" containerName="marketplace-operator" Nov 25 15:06:02 crc kubenswrapper[4890]: E1125 15:06:02.734766 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1983120-f62d-46b0-833e-db30a81973c1" containerName="extract-utilities" Nov 25 15:06:02 crc kubenswrapper[4890]: I1125 15:06:02.734771 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1983120-f62d-46b0-833e-db30a81973c1" containerName="extract-utilities" Nov 25 15:06:02 crc kubenswrapper[4890]: E1125 15:06:02.734778 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f052eff-1a42-4c20-8644-d575fc557060" containerName="extract-utilities" Nov 25 15:06:02 crc kubenswrapper[4890]: I1125 15:06:02.734783 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f052eff-1a42-4c20-8644-d575fc557060" containerName="extract-utilities" Nov 25 15:06:02 crc kubenswrapper[4890]: E1125 15:06:02.734793 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22b61785-937e-4562-8c48-5e3d390f3389" containerName="extract-utilities" Nov 25 15:06:02 crc kubenswrapper[4890]: I1125 15:06:02.734799 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="22b61785-937e-4562-8c48-5e3d390f3389" containerName="extract-utilities" Nov 25 15:06:02 crc kubenswrapper[4890]: E1125 15:06:02.734807 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1983120-f62d-46b0-833e-db30a81973c1" containerName="registry-server" Nov 25 15:06:02 crc kubenswrapper[4890]: I1125 15:06:02.734812 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1983120-f62d-46b0-833e-db30a81973c1" containerName="registry-server" Nov 25 15:06:02 crc kubenswrapper[4890]: E1125 15:06:02.734820 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fbb3062-51d8-4fc0-9b25-5336de75d692" containerName="registry-server" Nov 25 15:06:02 crc kubenswrapper[4890]: I1125 15:06:02.734825 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fbb3062-51d8-4fc0-9b25-5336de75d692" containerName="registry-server" Nov 25 15:06:02 crc kubenswrapper[4890]: E1125 15:06:02.734832 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f052eff-1a42-4c20-8644-d575fc557060" containerName="extract-content" Nov 25 15:06:02 crc kubenswrapper[4890]: I1125 15:06:02.734837 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f052eff-1a42-4c20-8644-d575fc557060" containerName="extract-content" Nov 25 15:06:02 crc kubenswrapper[4890]: E1125 15:06:02.734847 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fbb3062-51d8-4fc0-9b25-5336de75d692" containerName="extract-content" Nov 25 15:06:02 crc kubenswrapper[4890]: I1125 15:06:02.734852 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fbb3062-51d8-4fc0-9b25-5336de75d692" containerName="extract-content" Nov 25 15:06:02 crc kubenswrapper[4890]: E1125 15:06:02.734865 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22b61785-937e-4562-8c48-5e3d390f3389" containerName="registry-server" Nov 25 15:06:02 crc kubenswrapper[4890]: I1125 15:06:02.734870 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="22b61785-937e-4562-8c48-5e3d390f3389" containerName="registry-server" Nov 25 15:06:02 crc kubenswrapper[4890]: E1125 15:06:02.734877 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1983120-f62d-46b0-833e-db30a81973c1" containerName="extract-content" Nov 25 15:06:02 crc kubenswrapper[4890]: I1125 15:06:02.734882 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1983120-f62d-46b0-833e-db30a81973c1" containerName="extract-content" Nov 25 15:06:02 crc kubenswrapper[4890]: I1125 15:06:02.734960 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f052eff-1a42-4c20-8644-d575fc557060" containerName="registry-server" Nov 25 15:06:02 crc kubenswrapper[4890]: I1125 15:06:02.734974 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fbb3062-51d8-4fc0-9b25-5336de75d692" containerName="registry-server" Nov 25 15:06:02 crc kubenswrapper[4890]: I1125 15:06:02.734982 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ef9a0ab-2131-42ff-823d-17d517929e77" containerName="marketplace-operator" Nov 25 15:06:02 crc kubenswrapper[4890]: I1125 15:06:02.734990 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="22b61785-937e-4562-8c48-5e3d390f3389" containerName="registry-server" Nov 25 15:06:02 crc kubenswrapper[4890]: I1125 15:06:02.734998 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1983120-f62d-46b0-833e-db30a81973c1" containerName="registry-server" Nov 25 15:06:02 crc kubenswrapper[4890]: I1125 15:06:02.736090 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fksn8" Nov 25 15:06:02 crc kubenswrapper[4890]: I1125 15:06:02.738877 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 25 15:06:02 crc kubenswrapper[4890]: I1125 15:06:02.743371 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fksn8"] Nov 25 15:06:02 crc kubenswrapper[4890]: I1125 15:06:02.863302 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6afbc261-80f8-4f4a-814f-5df9ab43502e-utilities\") pod \"certified-operators-fksn8\" (UID: \"6afbc261-80f8-4f4a-814f-5df9ab43502e\") " pod="openshift-marketplace/certified-operators-fksn8" Nov 25 15:06:02 crc kubenswrapper[4890]: I1125 15:06:02.863347 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djbgn\" (UniqueName: \"kubernetes.io/projected/6afbc261-80f8-4f4a-814f-5df9ab43502e-kube-api-access-djbgn\") pod \"certified-operators-fksn8\" (UID: \"6afbc261-80f8-4f4a-814f-5df9ab43502e\") " pod="openshift-marketplace/certified-operators-fksn8" Nov 25 15:06:02 crc kubenswrapper[4890]: I1125 15:06:02.863387 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6afbc261-80f8-4f4a-814f-5df9ab43502e-catalog-content\") pod \"certified-operators-fksn8\" (UID: \"6afbc261-80f8-4f4a-814f-5df9ab43502e\") " pod="openshift-marketplace/certified-operators-fksn8" Nov 25 15:06:02 crc kubenswrapper[4890]: I1125 15:06:02.967790 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6afbc261-80f8-4f4a-814f-5df9ab43502e-utilities\") pod \"certified-operators-fksn8\" (UID: \"6afbc261-80f8-4f4a-814f-5df9ab43502e\") " pod="openshift-marketplace/certified-operators-fksn8" Nov 25 15:06:02 crc kubenswrapper[4890]: I1125 15:06:02.967844 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djbgn\" (UniqueName: \"kubernetes.io/projected/6afbc261-80f8-4f4a-814f-5df9ab43502e-kube-api-access-djbgn\") pod \"certified-operators-fksn8\" (UID: \"6afbc261-80f8-4f4a-814f-5df9ab43502e\") " pod="openshift-marketplace/certified-operators-fksn8" Nov 25 15:06:02 crc kubenswrapper[4890]: I1125 15:06:02.967883 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6afbc261-80f8-4f4a-814f-5df9ab43502e-catalog-content\") pod \"certified-operators-fksn8\" (UID: \"6afbc261-80f8-4f4a-814f-5df9ab43502e\") " pod="openshift-marketplace/certified-operators-fksn8" Nov 25 15:06:02 crc kubenswrapper[4890]: I1125 15:06:02.968429 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6afbc261-80f8-4f4a-814f-5df9ab43502e-catalog-content\") pod \"certified-operators-fksn8\" (UID: \"6afbc261-80f8-4f4a-814f-5df9ab43502e\") " pod="openshift-marketplace/certified-operators-fksn8" Nov 25 15:06:02 crc kubenswrapper[4890]: I1125 15:06:02.968748 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6afbc261-80f8-4f4a-814f-5df9ab43502e-utilities\") pod \"certified-operators-fksn8\" (UID: \"6afbc261-80f8-4f4a-814f-5df9ab43502e\") " pod="openshift-marketplace/certified-operators-fksn8" Nov 25 15:06:02 crc kubenswrapper[4890]: I1125 15:06:02.984719 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djbgn\" (UniqueName: \"kubernetes.io/projected/6afbc261-80f8-4f4a-814f-5df9ab43502e-kube-api-access-djbgn\") pod \"certified-operators-fksn8\" (UID: \"6afbc261-80f8-4f4a-814f-5df9ab43502e\") " pod="openshift-marketplace/certified-operators-fksn8" Nov 25 15:06:03 crc kubenswrapper[4890]: I1125 15:06:03.065409 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fksn8" Nov 25 15:06:03 crc kubenswrapper[4890]: I1125 15:06:03.507066 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fksn8"] Nov 25 15:06:03 crc kubenswrapper[4890]: I1125 15:06:03.542394 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fksn8" event={"ID":"6afbc261-80f8-4f4a-814f-5df9ab43502e","Type":"ContainerStarted","Data":"7db92a2cb259f676521102ec388a2c861105b5708b3bf7c4799e50d034cfafd2"} Nov 25 15:06:03 crc kubenswrapper[4890]: I1125 15:06:03.736259 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tjd2v"] Nov 25 15:06:03 crc kubenswrapper[4890]: I1125 15:06:03.737670 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tjd2v" Nov 25 15:06:03 crc kubenswrapper[4890]: I1125 15:06:03.739937 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 25 15:06:03 crc kubenswrapper[4890]: I1125 15:06:03.742751 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tjd2v"] Nov 25 15:06:03 crc kubenswrapper[4890]: I1125 15:06:03.781547 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f444e25c-b3bf-4046-a1ff-eba1a422a88e-utilities\") pod \"redhat-marketplace-tjd2v\" (UID: \"f444e25c-b3bf-4046-a1ff-eba1a422a88e\") " pod="openshift-marketplace/redhat-marketplace-tjd2v" Nov 25 15:06:03 crc kubenswrapper[4890]: I1125 15:06:03.781624 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nh4c\" (UniqueName: \"kubernetes.io/projected/f444e25c-b3bf-4046-a1ff-eba1a422a88e-kube-api-access-2nh4c\") pod \"redhat-marketplace-tjd2v\" (UID: \"f444e25c-b3bf-4046-a1ff-eba1a422a88e\") " pod="openshift-marketplace/redhat-marketplace-tjd2v" Nov 25 15:06:03 crc kubenswrapper[4890]: I1125 15:06:03.781656 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f444e25c-b3bf-4046-a1ff-eba1a422a88e-catalog-content\") pod \"redhat-marketplace-tjd2v\" (UID: \"f444e25c-b3bf-4046-a1ff-eba1a422a88e\") " pod="openshift-marketplace/redhat-marketplace-tjd2v" Nov 25 15:06:03 crc kubenswrapper[4890]: I1125 15:06:03.883429 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nh4c\" (UniqueName: \"kubernetes.io/projected/f444e25c-b3bf-4046-a1ff-eba1a422a88e-kube-api-access-2nh4c\") pod \"redhat-marketplace-tjd2v\" (UID: \"f444e25c-b3bf-4046-a1ff-eba1a422a88e\") " pod="openshift-marketplace/redhat-marketplace-tjd2v" Nov 25 15:06:03 crc kubenswrapper[4890]: I1125 15:06:03.883472 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f444e25c-b3bf-4046-a1ff-eba1a422a88e-catalog-content\") pod \"redhat-marketplace-tjd2v\" (UID: \"f444e25c-b3bf-4046-a1ff-eba1a422a88e\") " pod="openshift-marketplace/redhat-marketplace-tjd2v" Nov 25 15:06:03 crc kubenswrapper[4890]: I1125 15:06:03.883539 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f444e25c-b3bf-4046-a1ff-eba1a422a88e-utilities\") pod \"redhat-marketplace-tjd2v\" (UID: \"f444e25c-b3bf-4046-a1ff-eba1a422a88e\") " pod="openshift-marketplace/redhat-marketplace-tjd2v" Nov 25 15:06:03 crc kubenswrapper[4890]: I1125 15:06:03.883928 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f444e25c-b3bf-4046-a1ff-eba1a422a88e-utilities\") pod \"redhat-marketplace-tjd2v\" (UID: \"f444e25c-b3bf-4046-a1ff-eba1a422a88e\") " pod="openshift-marketplace/redhat-marketplace-tjd2v" Nov 25 15:06:03 crc kubenswrapper[4890]: I1125 15:06:03.883969 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f444e25c-b3bf-4046-a1ff-eba1a422a88e-catalog-content\") pod \"redhat-marketplace-tjd2v\" (UID: \"f444e25c-b3bf-4046-a1ff-eba1a422a88e\") " pod="openshift-marketplace/redhat-marketplace-tjd2v" Nov 25 15:06:03 crc kubenswrapper[4890]: I1125 15:06:03.902307 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nh4c\" (UniqueName: \"kubernetes.io/projected/f444e25c-b3bf-4046-a1ff-eba1a422a88e-kube-api-access-2nh4c\") pod \"redhat-marketplace-tjd2v\" (UID: \"f444e25c-b3bf-4046-a1ff-eba1a422a88e\") " pod="openshift-marketplace/redhat-marketplace-tjd2v" Nov 25 15:06:04 crc kubenswrapper[4890]: I1125 15:06:04.076864 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tjd2v" Nov 25 15:06:04 crc kubenswrapper[4890]: I1125 15:06:04.261027 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tjd2v"] Nov 25 15:06:04 crc kubenswrapper[4890]: W1125 15:06:04.266354 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf444e25c_b3bf_4046_a1ff_eba1a422a88e.slice/crio-b0711e1b8304ceb267e43817c08dc8818f1e07ffee4391a0c1c8b57a12659371 WatchSource:0}: Error finding container b0711e1b8304ceb267e43817c08dc8818f1e07ffee4391a0c1c8b57a12659371: Status 404 returned error can't find the container with id b0711e1b8304ceb267e43817c08dc8818f1e07ffee4391a0c1c8b57a12659371 Nov 25 15:06:04 crc kubenswrapper[4890]: I1125 15:06:04.552515 4890 generic.go:334] "Generic (PLEG): container finished" podID="6afbc261-80f8-4f4a-814f-5df9ab43502e" containerID="9971c2c8c3279154bc07c778e76adf8403bde0d3265d6c6981dbfc75f511563f" exitCode=0 Nov 25 15:06:04 crc kubenswrapper[4890]: I1125 15:06:04.553269 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fksn8" event={"ID":"6afbc261-80f8-4f4a-814f-5df9ab43502e","Type":"ContainerDied","Data":"9971c2c8c3279154bc07c778e76adf8403bde0d3265d6c6981dbfc75f511563f"} Nov 25 15:06:04 crc kubenswrapper[4890]: I1125 15:06:04.556265 4890 generic.go:334] "Generic (PLEG): container finished" podID="f444e25c-b3bf-4046-a1ff-eba1a422a88e" containerID="8d6f882acbc910a79cccc990cdd9dc22b5d39bf151fea4304f6bb260e82716b2" exitCode=0 Nov 25 15:06:04 crc kubenswrapper[4890]: I1125 15:06:04.556942 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tjd2v" event={"ID":"f444e25c-b3bf-4046-a1ff-eba1a422a88e","Type":"ContainerDied","Data":"8d6f882acbc910a79cccc990cdd9dc22b5d39bf151fea4304f6bb260e82716b2"} Nov 25 15:06:04 crc kubenswrapper[4890]: I1125 15:06:04.556985 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tjd2v" event={"ID":"f444e25c-b3bf-4046-a1ff-eba1a422a88e","Type":"ContainerStarted","Data":"b0711e1b8304ceb267e43817c08dc8818f1e07ffee4391a0c1c8b57a12659371"} Nov 25 15:06:05 crc kubenswrapper[4890]: I1125 15:06:05.138372 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jcmgp"] Nov 25 15:06:05 crc kubenswrapper[4890]: I1125 15:06:05.140358 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jcmgp" Nov 25 15:06:05 crc kubenswrapper[4890]: I1125 15:06:05.143832 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 25 15:06:05 crc kubenswrapper[4890]: I1125 15:06:05.154607 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jcmgp"] Nov 25 15:06:05 crc kubenswrapper[4890]: I1125 15:06:05.298910 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c83811ae-c669-4268-a58d-e4b44d18f26c-utilities\") pod \"redhat-operators-jcmgp\" (UID: \"c83811ae-c669-4268-a58d-e4b44d18f26c\") " pod="openshift-marketplace/redhat-operators-jcmgp" Nov 25 15:06:05 crc kubenswrapper[4890]: I1125 15:06:05.298984 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwds2\" (UniqueName: \"kubernetes.io/projected/c83811ae-c669-4268-a58d-e4b44d18f26c-kube-api-access-dwds2\") pod \"redhat-operators-jcmgp\" (UID: \"c83811ae-c669-4268-a58d-e4b44d18f26c\") " pod="openshift-marketplace/redhat-operators-jcmgp" Nov 25 15:06:05 crc kubenswrapper[4890]: I1125 15:06:05.299003 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c83811ae-c669-4268-a58d-e4b44d18f26c-catalog-content\") pod \"redhat-operators-jcmgp\" (UID: \"c83811ae-c669-4268-a58d-e4b44d18f26c\") " pod="openshift-marketplace/redhat-operators-jcmgp" Nov 25 15:06:05 crc kubenswrapper[4890]: I1125 15:06:05.400561 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c83811ae-c669-4268-a58d-e4b44d18f26c-catalog-content\") pod \"redhat-operators-jcmgp\" (UID: \"c83811ae-c669-4268-a58d-e4b44d18f26c\") " pod="openshift-marketplace/redhat-operators-jcmgp" Nov 25 15:06:05 crc kubenswrapper[4890]: I1125 15:06:05.400854 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwds2\" (UniqueName: \"kubernetes.io/projected/c83811ae-c669-4268-a58d-e4b44d18f26c-kube-api-access-dwds2\") pod \"redhat-operators-jcmgp\" (UID: \"c83811ae-c669-4268-a58d-e4b44d18f26c\") " pod="openshift-marketplace/redhat-operators-jcmgp" Nov 25 15:06:05 crc kubenswrapper[4890]: I1125 15:06:05.400974 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c83811ae-c669-4268-a58d-e4b44d18f26c-utilities\") pod \"redhat-operators-jcmgp\" (UID: \"c83811ae-c669-4268-a58d-e4b44d18f26c\") " pod="openshift-marketplace/redhat-operators-jcmgp" Nov 25 15:06:05 crc kubenswrapper[4890]: I1125 15:06:05.401000 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c83811ae-c669-4268-a58d-e4b44d18f26c-catalog-content\") pod \"redhat-operators-jcmgp\" (UID: \"c83811ae-c669-4268-a58d-e4b44d18f26c\") " pod="openshift-marketplace/redhat-operators-jcmgp" Nov 25 15:06:05 crc kubenswrapper[4890]: I1125 15:06:05.401450 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c83811ae-c669-4268-a58d-e4b44d18f26c-utilities\") pod \"redhat-operators-jcmgp\" (UID: \"c83811ae-c669-4268-a58d-e4b44d18f26c\") " pod="openshift-marketplace/redhat-operators-jcmgp" Nov 25 15:06:05 crc kubenswrapper[4890]: I1125 15:06:05.425346 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwds2\" (UniqueName: \"kubernetes.io/projected/c83811ae-c669-4268-a58d-e4b44d18f26c-kube-api-access-dwds2\") pod \"redhat-operators-jcmgp\" (UID: \"c83811ae-c669-4268-a58d-e4b44d18f26c\") " pod="openshift-marketplace/redhat-operators-jcmgp" Nov 25 15:06:05 crc kubenswrapper[4890]: I1125 15:06:05.467109 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jcmgp" Nov 25 15:06:05 crc kubenswrapper[4890]: I1125 15:06:05.704597 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jcmgp"] Nov 25 15:06:06 crc kubenswrapper[4890]: I1125 15:06:06.140391 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-wv9pw"] Nov 25 15:06:06 crc kubenswrapper[4890]: I1125 15:06:06.142823 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wv9pw" Nov 25 15:06:06 crc kubenswrapper[4890]: I1125 15:06:06.149649 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 25 15:06:06 crc kubenswrapper[4890]: I1125 15:06:06.151856 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wv9pw"] Nov 25 15:06:06 crc kubenswrapper[4890]: I1125 15:06:06.220092 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9f6v\" (UniqueName: \"kubernetes.io/projected/6febf84d-85dd-45d8-a32e-bc978311ed98-kube-api-access-j9f6v\") pod \"community-operators-wv9pw\" (UID: \"6febf84d-85dd-45d8-a32e-bc978311ed98\") " pod="openshift-marketplace/community-operators-wv9pw" Nov 25 15:06:06 crc kubenswrapper[4890]: I1125 15:06:06.220144 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6febf84d-85dd-45d8-a32e-bc978311ed98-catalog-content\") pod \"community-operators-wv9pw\" (UID: \"6febf84d-85dd-45d8-a32e-bc978311ed98\") " pod="openshift-marketplace/community-operators-wv9pw" Nov 25 15:06:06 crc kubenswrapper[4890]: I1125 15:06:06.220189 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6febf84d-85dd-45d8-a32e-bc978311ed98-utilities\") pod \"community-operators-wv9pw\" (UID: \"6febf84d-85dd-45d8-a32e-bc978311ed98\") " pod="openshift-marketplace/community-operators-wv9pw" Nov 25 15:06:06 crc kubenswrapper[4890]: I1125 15:06:06.321120 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9f6v\" (UniqueName: \"kubernetes.io/projected/6febf84d-85dd-45d8-a32e-bc978311ed98-kube-api-access-j9f6v\") pod \"community-operators-wv9pw\" (UID: \"6febf84d-85dd-45d8-a32e-bc978311ed98\") " pod="openshift-marketplace/community-operators-wv9pw" Nov 25 15:06:06 crc kubenswrapper[4890]: I1125 15:06:06.321226 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6febf84d-85dd-45d8-a32e-bc978311ed98-catalog-content\") pod \"community-operators-wv9pw\" (UID: \"6febf84d-85dd-45d8-a32e-bc978311ed98\") " pod="openshift-marketplace/community-operators-wv9pw" Nov 25 15:06:06 crc kubenswrapper[4890]: I1125 15:06:06.321263 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6febf84d-85dd-45d8-a32e-bc978311ed98-utilities\") pod \"community-operators-wv9pw\" (UID: \"6febf84d-85dd-45d8-a32e-bc978311ed98\") " pod="openshift-marketplace/community-operators-wv9pw" Nov 25 15:06:06 crc kubenswrapper[4890]: I1125 15:06:06.321836 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6febf84d-85dd-45d8-a32e-bc978311ed98-utilities\") pod \"community-operators-wv9pw\" (UID: \"6febf84d-85dd-45d8-a32e-bc978311ed98\") " pod="openshift-marketplace/community-operators-wv9pw" Nov 25 15:06:06 crc kubenswrapper[4890]: I1125 15:06:06.321836 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6febf84d-85dd-45d8-a32e-bc978311ed98-catalog-content\") pod \"community-operators-wv9pw\" (UID: \"6febf84d-85dd-45d8-a32e-bc978311ed98\") " pod="openshift-marketplace/community-operators-wv9pw" Nov 25 15:06:06 crc kubenswrapper[4890]: I1125 15:06:06.341418 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9f6v\" (UniqueName: \"kubernetes.io/projected/6febf84d-85dd-45d8-a32e-bc978311ed98-kube-api-access-j9f6v\") pod \"community-operators-wv9pw\" (UID: \"6febf84d-85dd-45d8-a32e-bc978311ed98\") " pod="openshift-marketplace/community-operators-wv9pw" Nov 25 15:06:06 crc kubenswrapper[4890]: I1125 15:06:06.462467 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wv9pw" Nov 25 15:06:06 crc kubenswrapper[4890]: I1125 15:06:06.585942 4890 generic.go:334] "Generic (PLEG): container finished" podID="6afbc261-80f8-4f4a-814f-5df9ab43502e" containerID="e47a2f0344175bbad15beca2430397705cfd1719b00fd0feceb796a7618b5322" exitCode=0 Nov 25 15:06:06 crc kubenswrapper[4890]: I1125 15:06:06.586120 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fksn8" event={"ID":"6afbc261-80f8-4f4a-814f-5df9ab43502e","Type":"ContainerDied","Data":"e47a2f0344175bbad15beca2430397705cfd1719b00fd0feceb796a7618b5322"} Nov 25 15:06:06 crc kubenswrapper[4890]: I1125 15:06:06.589945 4890 generic.go:334] "Generic (PLEG): container finished" podID="f444e25c-b3bf-4046-a1ff-eba1a422a88e" containerID="4a6c8b5c3a33fa314cc7b7de1442051bef3fae480473145fe2235935c3ce181f" exitCode=0 Nov 25 15:06:06 crc kubenswrapper[4890]: I1125 15:06:06.590758 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tjd2v" event={"ID":"f444e25c-b3bf-4046-a1ff-eba1a422a88e","Type":"ContainerDied","Data":"4a6c8b5c3a33fa314cc7b7de1442051bef3fae480473145fe2235935c3ce181f"} Nov 25 15:06:06 crc kubenswrapper[4890]: I1125 15:06:06.593512 4890 generic.go:334] "Generic (PLEG): container finished" podID="c83811ae-c669-4268-a58d-e4b44d18f26c" containerID="effc7d2ae8c5c7f241bd7dba65e7157341d6a01e93eef9b0a8de8e3388d16526" exitCode=0 Nov 25 15:06:06 crc kubenswrapper[4890]: I1125 15:06:06.593559 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jcmgp" event={"ID":"c83811ae-c669-4268-a58d-e4b44d18f26c","Type":"ContainerDied","Data":"effc7d2ae8c5c7f241bd7dba65e7157341d6a01e93eef9b0a8de8e3388d16526"} Nov 25 15:06:06 crc kubenswrapper[4890]: I1125 15:06:06.593592 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jcmgp" event={"ID":"c83811ae-c669-4268-a58d-e4b44d18f26c","Type":"ContainerStarted","Data":"014923c13e015d321b917e83b24689fbde406fbdca2bbac86ad571a093c32346"} Nov 25 15:06:06 crc kubenswrapper[4890]: I1125 15:06:06.641709 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wv9pw"] Nov 25 15:06:06 crc kubenswrapper[4890]: W1125 15:06:06.653664 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6febf84d_85dd_45d8_a32e_bc978311ed98.slice/crio-f7f3ac4b168bff400e533780771eea8c855e98984c4dc3dc6fdc76a899ae27e6 WatchSource:0}: Error finding container f7f3ac4b168bff400e533780771eea8c855e98984c4dc3dc6fdc76a899ae27e6: Status 404 returned error can't find the container with id f7f3ac4b168bff400e533780771eea8c855e98984c4dc3dc6fdc76a899ae27e6 Nov 25 15:06:07 crc kubenswrapper[4890]: I1125 15:06:07.599044 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fksn8" event={"ID":"6afbc261-80f8-4f4a-814f-5df9ab43502e","Type":"ContainerStarted","Data":"845be98a484b350449672a6bb0b1798bebe18d1d802f3d59f7b7e4f9352100e0"} Nov 25 15:06:07 crc kubenswrapper[4890]: I1125 15:06:07.601107 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tjd2v" event={"ID":"f444e25c-b3bf-4046-a1ff-eba1a422a88e","Type":"ContainerStarted","Data":"04c409416394f8abb8174bca2fceb58b1c8afb415ea77c94d56e49da60134d37"} Nov 25 15:06:07 crc kubenswrapper[4890]: I1125 15:06:07.602632 4890 generic.go:334] "Generic (PLEG): container finished" podID="6febf84d-85dd-45d8-a32e-bc978311ed98" containerID="2fc8c56c15eaf81ef26fdcab5ab1bf53eb820dc3116e4e2336449c2d1654c39e" exitCode=0 Nov 25 15:06:07 crc kubenswrapper[4890]: I1125 15:06:07.602664 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wv9pw" event={"ID":"6febf84d-85dd-45d8-a32e-bc978311ed98","Type":"ContainerDied","Data":"2fc8c56c15eaf81ef26fdcab5ab1bf53eb820dc3116e4e2336449c2d1654c39e"} Nov 25 15:06:07 crc kubenswrapper[4890]: I1125 15:06:07.602682 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wv9pw" event={"ID":"6febf84d-85dd-45d8-a32e-bc978311ed98","Type":"ContainerStarted","Data":"f7f3ac4b168bff400e533780771eea8c855e98984c4dc3dc6fdc76a899ae27e6"} Nov 25 15:06:07 crc kubenswrapper[4890]: I1125 15:06:07.647250 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fksn8" podStartSLOduration=3.085057416 podStartE2EDuration="5.647229694s" podCreationTimestamp="2025-11-25 15:06:02 +0000 UTC" firstStartedPulling="2025-11-25 15:06:04.555027037 +0000 UTC m=+222.997489647" lastFinishedPulling="2025-11-25 15:06:07.117199315 +0000 UTC m=+225.559661925" observedRunningTime="2025-11-25 15:06:07.623959448 +0000 UTC m=+226.066422068" watchObservedRunningTime="2025-11-25 15:06:07.647229694 +0000 UTC m=+226.089692304" Nov 25 15:06:07 crc kubenswrapper[4890]: I1125 15:06:07.660173 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tjd2v" podStartSLOduration=2.11582958 podStartE2EDuration="4.660142192s" podCreationTimestamp="2025-11-25 15:06:03 +0000 UTC" firstStartedPulling="2025-11-25 15:06:04.558942306 +0000 UTC m=+223.001404916" lastFinishedPulling="2025-11-25 15:06:07.103254918 +0000 UTC m=+225.545717528" observedRunningTime="2025-11-25 15:06:07.659171995 +0000 UTC m=+226.101634615" watchObservedRunningTime="2025-11-25 15:06:07.660142192 +0000 UTC m=+226.102604802" Nov 25 15:06:09 crc kubenswrapper[4890]: I1125 15:06:09.615555 4890 generic.go:334] "Generic (PLEG): container finished" podID="6febf84d-85dd-45d8-a32e-bc978311ed98" containerID="2acfef7b9ba50e368fad34e67848e20758f2619222bcac74ca7cefa436c0ba87" exitCode=0 Nov 25 15:06:09 crc kubenswrapper[4890]: I1125 15:06:09.615627 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wv9pw" event={"ID":"6febf84d-85dd-45d8-a32e-bc978311ed98","Type":"ContainerDied","Data":"2acfef7b9ba50e368fad34e67848e20758f2619222bcac74ca7cefa436c0ba87"} Nov 25 15:06:10 crc kubenswrapper[4890]: I1125 15:06:10.624950 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wv9pw" event={"ID":"6febf84d-85dd-45d8-a32e-bc978311ed98","Type":"ContainerStarted","Data":"2b6d1256478ad7877a0259e89b82b8b3c6ecb95cb1a0b7f86c266cb04d479d08"} Nov 25 15:06:10 crc kubenswrapper[4890]: I1125 15:06:10.646223 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-wv9pw" podStartSLOduration=2.197293779 podStartE2EDuration="4.646204052s" podCreationTimestamp="2025-11-25 15:06:06 +0000 UTC" firstStartedPulling="2025-11-25 15:06:07.611702347 +0000 UTC m=+226.054164947" lastFinishedPulling="2025-11-25 15:06:10.06061261 +0000 UTC m=+228.503075220" observedRunningTime="2025-11-25 15:06:10.641937233 +0000 UTC m=+229.084399853" watchObservedRunningTime="2025-11-25 15:06:10.646204052 +0000 UTC m=+229.088666662" Nov 25 15:06:13 crc kubenswrapper[4890]: I1125 15:06:13.065757 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fksn8" Nov 25 15:06:13 crc kubenswrapper[4890]: I1125 15:06:13.066315 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fksn8" Nov 25 15:06:13 crc kubenswrapper[4890]: I1125 15:06:13.109737 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fksn8" Nov 25 15:06:13 crc kubenswrapper[4890]: I1125 15:06:13.679704 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fksn8" Nov 25 15:06:14 crc kubenswrapper[4890]: I1125 15:06:14.077240 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tjd2v" Nov 25 15:06:14 crc kubenswrapper[4890]: I1125 15:06:14.077568 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tjd2v" Nov 25 15:06:14 crc kubenswrapper[4890]: I1125 15:06:14.114964 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tjd2v" Nov 25 15:06:14 crc kubenswrapper[4890]: I1125 15:06:14.684793 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tjd2v" Nov 25 15:06:16 crc kubenswrapper[4890]: I1125 15:06:16.463743 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-wv9pw" Nov 25 15:06:16 crc kubenswrapper[4890]: I1125 15:06:16.464351 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-wv9pw" Nov 25 15:06:16 crc kubenswrapper[4890]: I1125 15:06:16.502822 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-wv9pw" Nov 25 15:06:16 crc kubenswrapper[4890]: I1125 15:06:16.658135 4890 generic.go:334] "Generic (PLEG): container finished" podID="c83811ae-c669-4268-a58d-e4b44d18f26c" containerID="19a4b227c3e578ef5498279ae0d69818bce7f437d7a88f8f13922726125e3291" exitCode=0 Nov 25 15:06:16 crc kubenswrapper[4890]: I1125 15:06:16.658257 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jcmgp" event={"ID":"c83811ae-c669-4268-a58d-e4b44d18f26c","Type":"ContainerDied","Data":"19a4b227c3e578ef5498279ae0d69818bce7f437d7a88f8f13922726125e3291"} Nov 25 15:06:16 crc kubenswrapper[4890]: I1125 15:06:16.695074 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-wv9pw" Nov 25 15:06:18 crc kubenswrapper[4890]: I1125 15:06:18.670225 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jcmgp" event={"ID":"c83811ae-c669-4268-a58d-e4b44d18f26c","Type":"ContainerStarted","Data":"6266f2e088446c8531805d972af03bc3530ba558c1bf6f37e4c6bdb8883c2340"} Nov 25 15:06:25 crc kubenswrapper[4890]: I1125 15:06:25.467449 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jcmgp" Nov 25 15:06:25 crc kubenswrapper[4890]: I1125 15:06:25.467948 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jcmgp" Nov 25 15:06:25 crc kubenswrapper[4890]: I1125 15:06:25.505502 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jcmgp" Nov 25 15:06:25 crc kubenswrapper[4890]: I1125 15:06:25.520694 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jcmgp" podStartSLOduration=10.01616768 podStartE2EDuration="20.520674529s" podCreationTimestamp="2025-11-25 15:06:05 +0000 UTC" firstStartedPulling="2025-11-25 15:06:06.595429976 +0000 UTC m=+225.037892586" lastFinishedPulling="2025-11-25 15:06:17.099936825 +0000 UTC m=+235.542399435" observedRunningTime="2025-11-25 15:06:18.691058409 +0000 UTC m=+237.133521029" watchObservedRunningTime="2025-11-25 15:06:25.520674529 +0000 UTC m=+243.963137149" Nov 25 15:06:25 crc kubenswrapper[4890]: I1125 15:06:25.748112 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jcmgp" Nov 25 15:07:56 crc kubenswrapper[4890]: I1125 15:07:56.448413 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 15:07:56 crc kubenswrapper[4890]: I1125 15:07:56.449469 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 15:08:26 crc kubenswrapper[4890]: I1125 15:08:26.447440 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 15:08:26 crc kubenswrapper[4890]: I1125 15:08:26.448099 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 15:08:56 crc kubenswrapper[4890]: I1125 15:08:56.446901 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 15:08:56 crc kubenswrapper[4890]: I1125 15:08:56.447431 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 15:08:56 crc kubenswrapper[4890]: I1125 15:08:56.447484 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" Nov 25 15:08:56 crc kubenswrapper[4890]: I1125 15:08:56.448032 4890 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2f432adbd81e026db39d6c93118322aceca51fcd04a65e4110d7e407d22f0ff0"} pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 15:08:56 crc kubenswrapper[4890]: I1125 15:08:56.448092 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" containerID="cri-o://2f432adbd81e026db39d6c93118322aceca51fcd04a65e4110d7e407d22f0ff0" gracePeriod=600 Nov 25 15:08:56 crc kubenswrapper[4890]: E1125 15:08:56.539689 4890 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4e4f849d_f239_4727_a73e_18327856929a.slice/crio-conmon-2f432adbd81e026db39d6c93118322aceca51fcd04a65e4110d7e407d22f0ff0.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4e4f849d_f239_4727_a73e_18327856929a.slice/crio-2f432adbd81e026db39d6c93118322aceca51fcd04a65e4110d7e407d22f0ff0.scope\": RecentStats: unable to find data in memory cache]" Nov 25 15:08:56 crc kubenswrapper[4890]: I1125 15:08:56.624929 4890 generic.go:334] "Generic (PLEG): container finished" podID="4e4f849d-f239-4727-a73e-18327856929a" containerID="2f432adbd81e026db39d6c93118322aceca51fcd04a65e4110d7e407d22f0ff0" exitCode=0 Nov 25 15:08:56 crc kubenswrapper[4890]: I1125 15:08:56.624994 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" event={"ID":"4e4f849d-f239-4727-a73e-18327856929a","Type":"ContainerDied","Data":"2f432adbd81e026db39d6c93118322aceca51fcd04a65e4110d7e407d22f0ff0"} Nov 25 15:08:56 crc kubenswrapper[4890]: I1125 15:08:56.625358 4890 scope.go:117] "RemoveContainer" containerID="0902d6f8e56aecc5a2a13a7148f94a8a00ad139027b3c036aba0215ba0a67f59" Nov 25 15:08:57 crc kubenswrapper[4890]: I1125 15:08:57.634190 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" event={"ID":"4e4f849d-f239-4727-a73e-18327856929a","Type":"ContainerStarted","Data":"1a14507b222e3efb8d4e5a5a36210ae167e2bc170367e62ff6f845c2d500391c"} Nov 25 15:09:49 crc kubenswrapper[4890]: I1125 15:09:49.174738 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-t7bf9"] Nov 25 15:09:49 crc kubenswrapper[4890]: I1125 15:09:49.176174 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-t7bf9" Nov 25 15:09:49 crc kubenswrapper[4890]: I1125 15:09:49.194109 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-t7bf9"] Nov 25 15:09:49 crc kubenswrapper[4890]: I1125 15:09:49.302315 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d25a6e07-bd9d-4148-818c-80e2b7c40be8-ca-trust-extracted\") pod \"image-registry-66df7c8f76-t7bf9\" (UID: \"d25a6e07-bd9d-4148-818c-80e2b7c40be8\") " pod="openshift-image-registry/image-registry-66df7c8f76-t7bf9" Nov 25 15:09:49 crc kubenswrapper[4890]: I1125 15:09:49.302485 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-t7bf9\" (UID: \"d25a6e07-bd9d-4148-818c-80e2b7c40be8\") " pod="openshift-image-registry/image-registry-66df7c8f76-t7bf9" Nov 25 15:09:49 crc kubenswrapper[4890]: I1125 15:09:49.302561 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d25a6e07-bd9d-4148-818c-80e2b7c40be8-bound-sa-token\") pod \"image-registry-66df7c8f76-t7bf9\" (UID: \"d25a6e07-bd9d-4148-818c-80e2b7c40be8\") " pod="openshift-image-registry/image-registry-66df7c8f76-t7bf9" Nov 25 15:09:49 crc kubenswrapper[4890]: I1125 15:09:49.302609 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d25a6e07-bd9d-4148-818c-80e2b7c40be8-registry-tls\") pod \"image-registry-66df7c8f76-t7bf9\" (UID: \"d25a6e07-bd9d-4148-818c-80e2b7c40be8\") " pod="openshift-image-registry/image-registry-66df7c8f76-t7bf9" Nov 25 15:09:49 crc kubenswrapper[4890]: I1125 15:09:49.302652 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d25a6e07-bd9d-4148-818c-80e2b7c40be8-trusted-ca\") pod \"image-registry-66df7c8f76-t7bf9\" (UID: \"d25a6e07-bd9d-4148-818c-80e2b7c40be8\") " pod="openshift-image-registry/image-registry-66df7c8f76-t7bf9" Nov 25 15:09:49 crc kubenswrapper[4890]: I1125 15:09:49.302798 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d25a6e07-bd9d-4148-818c-80e2b7c40be8-installation-pull-secrets\") pod \"image-registry-66df7c8f76-t7bf9\" (UID: \"d25a6e07-bd9d-4148-818c-80e2b7c40be8\") " pod="openshift-image-registry/image-registry-66df7c8f76-t7bf9" Nov 25 15:09:49 crc kubenswrapper[4890]: I1125 15:09:49.302862 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-478sf\" (UniqueName: \"kubernetes.io/projected/d25a6e07-bd9d-4148-818c-80e2b7c40be8-kube-api-access-478sf\") pod \"image-registry-66df7c8f76-t7bf9\" (UID: \"d25a6e07-bd9d-4148-818c-80e2b7c40be8\") " pod="openshift-image-registry/image-registry-66df7c8f76-t7bf9" Nov 25 15:09:49 crc kubenswrapper[4890]: I1125 15:09:49.302928 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d25a6e07-bd9d-4148-818c-80e2b7c40be8-registry-certificates\") pod \"image-registry-66df7c8f76-t7bf9\" (UID: \"d25a6e07-bd9d-4148-818c-80e2b7c40be8\") " pod="openshift-image-registry/image-registry-66df7c8f76-t7bf9" Nov 25 15:09:49 crc kubenswrapper[4890]: I1125 15:09:49.331709 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-t7bf9\" (UID: \"d25a6e07-bd9d-4148-818c-80e2b7c40be8\") " pod="openshift-image-registry/image-registry-66df7c8f76-t7bf9" Nov 25 15:09:49 crc kubenswrapper[4890]: I1125 15:09:49.404670 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d25a6e07-bd9d-4148-818c-80e2b7c40be8-bound-sa-token\") pod \"image-registry-66df7c8f76-t7bf9\" (UID: \"d25a6e07-bd9d-4148-818c-80e2b7c40be8\") " pod="openshift-image-registry/image-registry-66df7c8f76-t7bf9" Nov 25 15:09:49 crc kubenswrapper[4890]: I1125 15:09:49.404724 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d25a6e07-bd9d-4148-818c-80e2b7c40be8-registry-tls\") pod \"image-registry-66df7c8f76-t7bf9\" (UID: \"d25a6e07-bd9d-4148-818c-80e2b7c40be8\") " pod="openshift-image-registry/image-registry-66df7c8f76-t7bf9" Nov 25 15:09:49 crc kubenswrapper[4890]: I1125 15:09:49.404757 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d25a6e07-bd9d-4148-818c-80e2b7c40be8-trusted-ca\") pod \"image-registry-66df7c8f76-t7bf9\" (UID: \"d25a6e07-bd9d-4148-818c-80e2b7c40be8\") " pod="openshift-image-registry/image-registry-66df7c8f76-t7bf9" Nov 25 15:09:49 crc kubenswrapper[4890]: I1125 15:09:49.404803 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d25a6e07-bd9d-4148-818c-80e2b7c40be8-installation-pull-secrets\") pod \"image-registry-66df7c8f76-t7bf9\" (UID: \"d25a6e07-bd9d-4148-818c-80e2b7c40be8\") " pod="openshift-image-registry/image-registry-66df7c8f76-t7bf9" Nov 25 15:09:49 crc kubenswrapper[4890]: I1125 15:09:49.404829 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-478sf\" (UniqueName: \"kubernetes.io/projected/d25a6e07-bd9d-4148-818c-80e2b7c40be8-kube-api-access-478sf\") pod \"image-registry-66df7c8f76-t7bf9\" (UID: \"d25a6e07-bd9d-4148-818c-80e2b7c40be8\") " pod="openshift-image-registry/image-registry-66df7c8f76-t7bf9" Nov 25 15:09:49 crc kubenswrapper[4890]: I1125 15:09:49.404854 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d25a6e07-bd9d-4148-818c-80e2b7c40be8-registry-certificates\") pod \"image-registry-66df7c8f76-t7bf9\" (UID: \"d25a6e07-bd9d-4148-818c-80e2b7c40be8\") " pod="openshift-image-registry/image-registry-66df7c8f76-t7bf9" Nov 25 15:09:49 crc kubenswrapper[4890]: I1125 15:09:49.404880 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d25a6e07-bd9d-4148-818c-80e2b7c40be8-ca-trust-extracted\") pod \"image-registry-66df7c8f76-t7bf9\" (UID: \"d25a6e07-bd9d-4148-818c-80e2b7c40be8\") " pod="openshift-image-registry/image-registry-66df7c8f76-t7bf9" Nov 25 15:09:49 crc kubenswrapper[4890]: I1125 15:09:49.405696 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d25a6e07-bd9d-4148-818c-80e2b7c40be8-ca-trust-extracted\") pod \"image-registry-66df7c8f76-t7bf9\" (UID: \"d25a6e07-bd9d-4148-818c-80e2b7c40be8\") " pod="openshift-image-registry/image-registry-66df7c8f76-t7bf9" Nov 25 15:09:49 crc kubenswrapper[4890]: I1125 15:09:49.406439 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d25a6e07-bd9d-4148-818c-80e2b7c40be8-registry-certificates\") pod \"image-registry-66df7c8f76-t7bf9\" (UID: \"d25a6e07-bd9d-4148-818c-80e2b7c40be8\") " pod="openshift-image-registry/image-registry-66df7c8f76-t7bf9" Nov 25 15:09:49 crc kubenswrapper[4890]: I1125 15:09:49.406690 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d25a6e07-bd9d-4148-818c-80e2b7c40be8-trusted-ca\") pod \"image-registry-66df7c8f76-t7bf9\" (UID: \"d25a6e07-bd9d-4148-818c-80e2b7c40be8\") " pod="openshift-image-registry/image-registry-66df7c8f76-t7bf9" Nov 25 15:09:49 crc kubenswrapper[4890]: I1125 15:09:49.410495 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d25a6e07-bd9d-4148-818c-80e2b7c40be8-installation-pull-secrets\") pod \"image-registry-66df7c8f76-t7bf9\" (UID: \"d25a6e07-bd9d-4148-818c-80e2b7c40be8\") " pod="openshift-image-registry/image-registry-66df7c8f76-t7bf9" Nov 25 15:09:49 crc kubenswrapper[4890]: I1125 15:09:49.415685 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d25a6e07-bd9d-4148-818c-80e2b7c40be8-registry-tls\") pod \"image-registry-66df7c8f76-t7bf9\" (UID: \"d25a6e07-bd9d-4148-818c-80e2b7c40be8\") " pod="openshift-image-registry/image-registry-66df7c8f76-t7bf9" Nov 25 15:09:49 crc kubenswrapper[4890]: I1125 15:09:49.421220 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-478sf\" (UniqueName: \"kubernetes.io/projected/d25a6e07-bd9d-4148-818c-80e2b7c40be8-kube-api-access-478sf\") pod \"image-registry-66df7c8f76-t7bf9\" (UID: \"d25a6e07-bd9d-4148-818c-80e2b7c40be8\") " pod="openshift-image-registry/image-registry-66df7c8f76-t7bf9" Nov 25 15:09:49 crc kubenswrapper[4890]: I1125 15:09:49.421515 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d25a6e07-bd9d-4148-818c-80e2b7c40be8-bound-sa-token\") pod \"image-registry-66df7c8f76-t7bf9\" (UID: \"d25a6e07-bd9d-4148-818c-80e2b7c40be8\") " pod="openshift-image-registry/image-registry-66df7c8f76-t7bf9" Nov 25 15:09:49 crc kubenswrapper[4890]: I1125 15:09:49.491719 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-t7bf9" Nov 25 15:09:49 crc kubenswrapper[4890]: I1125 15:09:49.708640 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-t7bf9"] Nov 25 15:09:49 crc kubenswrapper[4890]: I1125 15:09:49.953141 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-t7bf9" event={"ID":"d25a6e07-bd9d-4148-818c-80e2b7c40be8","Type":"ContainerStarted","Data":"9107490ea9a642b9827de07907dcb8906c05f0c3457842e87ce229cb6af94a1a"} Nov 25 15:09:49 crc kubenswrapper[4890]: I1125 15:09:49.953198 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-t7bf9" event={"ID":"d25a6e07-bd9d-4148-818c-80e2b7c40be8","Type":"ContainerStarted","Data":"03d5419fe8211ae6040cc095116a45bfc04754d6e65a4d1677991251da8b5598"} Nov 25 15:09:49 crc kubenswrapper[4890]: I1125 15:09:49.953291 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-t7bf9" Nov 25 15:09:49 crc kubenswrapper[4890]: I1125 15:09:49.980297 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-t7bf9" podStartSLOduration=0.980279638 podStartE2EDuration="980.279638ms" podCreationTimestamp="2025-11-25 15:09:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:09:49.97807745 +0000 UTC m=+448.420540090" watchObservedRunningTime="2025-11-25 15:09:49.980279638 +0000 UTC m=+448.422742248" Nov 25 15:10:09 crc kubenswrapper[4890]: I1125 15:10:09.497931 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-t7bf9" Nov 25 15:10:09 crc kubenswrapper[4890]: I1125 15:10:09.540834 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-vkrx2"] Nov 25 15:10:34 crc kubenswrapper[4890]: I1125 15:10:34.579803 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" podUID="96800ba3-0ee4-42e0-b8f7-316ef57a6173" containerName="registry" containerID="cri-o://7eae6ff20f6d7f1191e0f629ea1c2463746285c08eb996c6c7ba18f5a4856b72" gracePeriod=30 Nov 25 15:10:35 crc kubenswrapper[4890]: I1125 15:10:35.317058 4890 generic.go:334] "Generic (PLEG): container finished" podID="96800ba3-0ee4-42e0-b8f7-316ef57a6173" containerID="7eae6ff20f6d7f1191e0f629ea1c2463746285c08eb996c6c7ba18f5a4856b72" exitCode=0 Nov 25 15:10:35 crc kubenswrapper[4890]: I1125 15:10:35.317201 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" event={"ID":"96800ba3-0ee4-42e0-b8f7-316ef57a6173","Type":"ContainerDied","Data":"7eae6ff20f6d7f1191e0f629ea1c2463746285c08eb996c6c7ba18f5a4856b72"} Nov 25 15:10:35 crc kubenswrapper[4890]: I1125 15:10:35.481630 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:10:35 crc kubenswrapper[4890]: I1125 15:10:35.681953 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p5ssx\" (UniqueName: \"kubernetes.io/projected/96800ba3-0ee4-42e0-b8f7-316ef57a6173-kube-api-access-p5ssx\") pod \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " Nov 25 15:10:35 crc kubenswrapper[4890]: I1125 15:10:35.682046 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/96800ba3-0ee4-42e0-b8f7-316ef57a6173-registry-certificates\") pod \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " Nov 25 15:10:35 crc kubenswrapper[4890]: I1125 15:10:35.682087 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/96800ba3-0ee4-42e0-b8f7-316ef57a6173-ca-trust-extracted\") pod \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " Nov 25 15:10:35 crc kubenswrapper[4890]: I1125 15:10:35.682155 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/96800ba3-0ee4-42e0-b8f7-316ef57a6173-registry-tls\") pod \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " Nov 25 15:10:35 crc kubenswrapper[4890]: I1125 15:10:35.682507 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " Nov 25 15:10:35 crc kubenswrapper[4890]: I1125 15:10:35.682590 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/96800ba3-0ee4-42e0-b8f7-316ef57a6173-installation-pull-secrets\") pod \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " Nov 25 15:10:35 crc kubenswrapper[4890]: I1125 15:10:35.682651 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/96800ba3-0ee4-42e0-b8f7-316ef57a6173-trusted-ca\") pod \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " Nov 25 15:10:35 crc kubenswrapper[4890]: I1125 15:10:35.682717 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/96800ba3-0ee4-42e0-b8f7-316ef57a6173-bound-sa-token\") pod \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\" (UID: \"96800ba3-0ee4-42e0-b8f7-316ef57a6173\") " Nov 25 15:10:35 crc kubenswrapper[4890]: I1125 15:10:35.684643 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96800ba3-0ee4-42e0-b8f7-316ef57a6173-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "96800ba3-0ee4-42e0-b8f7-316ef57a6173" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:10:35 crc kubenswrapper[4890]: I1125 15:10:35.684696 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96800ba3-0ee4-42e0-b8f7-316ef57a6173-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "96800ba3-0ee4-42e0-b8f7-316ef57a6173" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:10:35 crc kubenswrapper[4890]: I1125 15:10:35.689507 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96800ba3-0ee4-42e0-b8f7-316ef57a6173-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "96800ba3-0ee4-42e0-b8f7-316ef57a6173" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:10:35 crc kubenswrapper[4890]: I1125 15:10:35.690377 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96800ba3-0ee4-42e0-b8f7-316ef57a6173-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "96800ba3-0ee4-42e0-b8f7-316ef57a6173" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:10:35 crc kubenswrapper[4890]: I1125 15:10:35.690784 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96800ba3-0ee4-42e0-b8f7-316ef57a6173-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "96800ba3-0ee4-42e0-b8f7-316ef57a6173" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:10:35 crc kubenswrapper[4890]: I1125 15:10:35.692943 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96800ba3-0ee4-42e0-b8f7-316ef57a6173-kube-api-access-p5ssx" (OuterVolumeSpecName: "kube-api-access-p5ssx") pod "96800ba3-0ee4-42e0-b8f7-316ef57a6173" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173"). InnerVolumeSpecName "kube-api-access-p5ssx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:10:35 crc kubenswrapper[4890]: I1125 15:10:35.706733 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96800ba3-0ee4-42e0-b8f7-316ef57a6173-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "96800ba3-0ee4-42e0-b8f7-316ef57a6173" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:10:35 crc kubenswrapper[4890]: I1125 15:10:35.712640 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "96800ba3-0ee4-42e0-b8f7-316ef57a6173" (UID: "96800ba3-0ee4-42e0-b8f7-316ef57a6173"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 25 15:10:35 crc kubenswrapper[4890]: I1125 15:10:35.784151 4890 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/96800ba3-0ee4-42e0-b8f7-316ef57a6173-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 25 15:10:35 crc kubenswrapper[4890]: I1125 15:10:35.784219 4890 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/96800ba3-0ee4-42e0-b8f7-316ef57a6173-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 25 15:10:35 crc kubenswrapper[4890]: I1125 15:10:35.784235 4890 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/96800ba3-0ee4-42e0-b8f7-316ef57a6173-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 15:10:35 crc kubenswrapper[4890]: I1125 15:10:35.784245 4890 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/96800ba3-0ee4-42e0-b8f7-316ef57a6173-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 25 15:10:35 crc kubenswrapper[4890]: I1125 15:10:35.784256 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p5ssx\" (UniqueName: \"kubernetes.io/projected/96800ba3-0ee4-42e0-b8f7-316ef57a6173-kube-api-access-p5ssx\") on node \"crc\" DevicePath \"\"" Nov 25 15:10:35 crc kubenswrapper[4890]: I1125 15:10:35.784266 4890 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/96800ba3-0ee4-42e0-b8f7-316ef57a6173-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 25 15:10:35 crc kubenswrapper[4890]: I1125 15:10:35.784275 4890 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/96800ba3-0ee4-42e0-b8f7-316ef57a6173-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 25 15:10:36 crc kubenswrapper[4890]: I1125 15:10:36.325381 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" Nov 25 15:10:36 crc kubenswrapper[4890]: I1125 15:10:36.325402 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-vkrx2" event={"ID":"96800ba3-0ee4-42e0-b8f7-316ef57a6173","Type":"ContainerDied","Data":"23eb54f5a723ab1789e822b18a86de7fdb10770fc00dcbacfd094343feabd298"} Nov 25 15:10:36 crc kubenswrapper[4890]: I1125 15:10:36.325480 4890 scope.go:117] "RemoveContainer" containerID="7eae6ff20f6d7f1191e0f629ea1c2463746285c08eb996c6c7ba18f5a4856b72" Nov 25 15:10:36 crc kubenswrapper[4890]: I1125 15:10:36.344150 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-vkrx2"] Nov 25 15:10:36 crc kubenswrapper[4890]: I1125 15:10:36.352928 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-vkrx2"] Nov 25 15:10:38 crc kubenswrapper[4890]: I1125 15:10:38.182882 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96800ba3-0ee4-42e0-b8f7-316ef57a6173" path="/var/lib/kubelet/pods/96800ba3-0ee4-42e0-b8f7-316ef57a6173/volumes" Nov 25 15:10:56 crc kubenswrapper[4890]: I1125 15:10:56.447996 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 15:10:56 crc kubenswrapper[4890]: I1125 15:10:56.448686 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 15:11:26 crc kubenswrapper[4890]: I1125 15:11:26.448313 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 15:11:26 crc kubenswrapper[4890]: I1125 15:11:26.450295 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 15:11:39 crc kubenswrapper[4890]: I1125 15:11:39.756851 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-x7dwr"] Nov 25 15:11:39 crc kubenswrapper[4890]: E1125 15:11:39.757898 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96800ba3-0ee4-42e0-b8f7-316ef57a6173" containerName="registry" Nov 25 15:11:39 crc kubenswrapper[4890]: I1125 15:11:39.757912 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="96800ba3-0ee4-42e0-b8f7-316ef57a6173" containerName="registry" Nov 25 15:11:39 crc kubenswrapper[4890]: I1125 15:11:39.758022 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="96800ba3-0ee4-42e0-b8f7-316ef57a6173" containerName="registry" Nov 25 15:11:39 crc kubenswrapper[4890]: I1125 15:11:39.758509 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-x7dwr" Nov 25 15:11:39 crc kubenswrapper[4890]: I1125 15:11:39.761732 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Nov 25 15:11:39 crc kubenswrapper[4890]: I1125 15:11:39.762455 4890 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-n9lgx" Nov 25 15:11:39 crc kubenswrapper[4890]: I1125 15:11:39.763181 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Nov 25 15:11:39 crc kubenswrapper[4890]: I1125 15:11:39.765210 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-9c9wk"] Nov 25 15:11:39 crc kubenswrapper[4890]: I1125 15:11:39.765881 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-9c9wk" Nov 25 15:11:39 crc kubenswrapper[4890]: I1125 15:11:39.768101 4890 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-hm6p5" Nov 25 15:11:39 crc kubenswrapper[4890]: I1125 15:11:39.770045 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-x7dwr"] Nov 25 15:11:39 crc kubenswrapper[4890]: I1125 15:11:39.780148 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-9c9wk"] Nov 25 15:11:39 crc kubenswrapper[4890]: I1125 15:11:39.815792 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-r5kd7"] Nov 25 15:11:39 crc kubenswrapper[4890]: I1125 15:11:39.816514 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-r5kd7" Nov 25 15:11:39 crc kubenswrapper[4890]: I1125 15:11:39.818106 4890 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-26pqz" Nov 25 15:11:39 crc kubenswrapper[4890]: I1125 15:11:39.829474 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-r5kd7"] Nov 25 15:11:39 crc kubenswrapper[4890]: I1125 15:11:39.853493 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nt8l2\" (UniqueName: \"kubernetes.io/projected/d0f0970d-430a-4ea0-b18c-6af23eb40c71-kube-api-access-nt8l2\") pod \"cert-manager-cainjector-7f985d654d-x7dwr\" (UID: \"d0f0970d-430a-4ea0-b18c-6af23eb40c71\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-x7dwr" Nov 25 15:11:39 crc kubenswrapper[4890]: I1125 15:11:39.853548 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkhjh\" (UniqueName: \"kubernetes.io/projected/e14d1071-6d9a-4543-b593-8e999b7dcad2-kube-api-access-dkhjh\") pod \"cert-manager-5b446d88c5-9c9wk\" (UID: \"e14d1071-6d9a-4543-b593-8e999b7dcad2\") " pod="cert-manager/cert-manager-5b446d88c5-9c9wk" Nov 25 15:11:39 crc kubenswrapper[4890]: I1125 15:11:39.955667 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7dd8\" (UniqueName: \"kubernetes.io/projected/3e0ccee5-4080-4973-99df-e4d6e929e471-kube-api-access-g7dd8\") pod \"cert-manager-webhook-5655c58dd6-r5kd7\" (UID: \"3e0ccee5-4080-4973-99df-e4d6e929e471\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-r5kd7" Nov 25 15:11:39 crc kubenswrapper[4890]: I1125 15:11:39.955731 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nt8l2\" (UniqueName: \"kubernetes.io/projected/d0f0970d-430a-4ea0-b18c-6af23eb40c71-kube-api-access-nt8l2\") pod \"cert-manager-cainjector-7f985d654d-x7dwr\" (UID: \"d0f0970d-430a-4ea0-b18c-6af23eb40c71\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-x7dwr" Nov 25 15:11:39 crc kubenswrapper[4890]: I1125 15:11:39.955758 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkhjh\" (UniqueName: \"kubernetes.io/projected/e14d1071-6d9a-4543-b593-8e999b7dcad2-kube-api-access-dkhjh\") pod \"cert-manager-5b446d88c5-9c9wk\" (UID: \"e14d1071-6d9a-4543-b593-8e999b7dcad2\") " pod="cert-manager/cert-manager-5b446d88c5-9c9wk" Nov 25 15:11:39 crc kubenswrapper[4890]: I1125 15:11:39.975315 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkhjh\" (UniqueName: \"kubernetes.io/projected/e14d1071-6d9a-4543-b593-8e999b7dcad2-kube-api-access-dkhjh\") pod \"cert-manager-5b446d88c5-9c9wk\" (UID: \"e14d1071-6d9a-4543-b593-8e999b7dcad2\") " pod="cert-manager/cert-manager-5b446d88c5-9c9wk" Nov 25 15:11:39 crc kubenswrapper[4890]: I1125 15:11:39.975315 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nt8l2\" (UniqueName: \"kubernetes.io/projected/d0f0970d-430a-4ea0-b18c-6af23eb40c71-kube-api-access-nt8l2\") pod \"cert-manager-cainjector-7f985d654d-x7dwr\" (UID: \"d0f0970d-430a-4ea0-b18c-6af23eb40c71\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-x7dwr" Nov 25 15:11:40 crc kubenswrapper[4890]: I1125 15:11:40.057075 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7dd8\" (UniqueName: \"kubernetes.io/projected/3e0ccee5-4080-4973-99df-e4d6e929e471-kube-api-access-g7dd8\") pod \"cert-manager-webhook-5655c58dd6-r5kd7\" (UID: \"3e0ccee5-4080-4973-99df-e4d6e929e471\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-r5kd7" Nov 25 15:11:40 crc kubenswrapper[4890]: I1125 15:11:40.080724 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7dd8\" (UniqueName: \"kubernetes.io/projected/3e0ccee5-4080-4973-99df-e4d6e929e471-kube-api-access-g7dd8\") pod \"cert-manager-webhook-5655c58dd6-r5kd7\" (UID: \"3e0ccee5-4080-4973-99df-e4d6e929e471\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-r5kd7" Nov 25 15:11:40 crc kubenswrapper[4890]: I1125 15:11:40.080853 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-x7dwr" Nov 25 15:11:40 crc kubenswrapper[4890]: I1125 15:11:40.090924 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-9c9wk" Nov 25 15:11:40 crc kubenswrapper[4890]: I1125 15:11:40.141121 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-r5kd7" Nov 25 15:11:40 crc kubenswrapper[4890]: I1125 15:11:40.381896 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-r5kd7"] Nov 25 15:11:40 crc kubenswrapper[4890]: I1125 15:11:40.389379 4890 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 15:11:40 crc kubenswrapper[4890]: I1125 15:11:40.488582 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-9c9wk"] Nov 25 15:11:40 crc kubenswrapper[4890]: W1125 15:11:40.495010 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode14d1071_6d9a_4543_b593_8e999b7dcad2.slice/crio-36f683e077f94e7a8fc48a37bc417fd96f2eb3686b30cedc2c062f958e7a66ea WatchSource:0}: Error finding container 36f683e077f94e7a8fc48a37bc417fd96f2eb3686b30cedc2c062f958e7a66ea: Status 404 returned error can't find the container with id 36f683e077f94e7a8fc48a37bc417fd96f2eb3686b30cedc2c062f958e7a66ea Nov 25 15:11:40 crc kubenswrapper[4890]: W1125 15:11:40.543878 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd0f0970d_430a_4ea0_b18c_6af23eb40c71.slice/crio-75c914574b30a5a7f2319e126d3a4cceae977165fd88648e016f6effab049ff7 WatchSource:0}: Error finding container 75c914574b30a5a7f2319e126d3a4cceae977165fd88648e016f6effab049ff7: Status 404 returned error can't find the container with id 75c914574b30a5a7f2319e126d3a4cceae977165fd88648e016f6effab049ff7 Nov 25 15:11:40 crc kubenswrapper[4890]: I1125 15:11:40.546832 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-x7dwr"] Nov 25 15:11:40 crc kubenswrapper[4890]: I1125 15:11:40.664946 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-9c9wk" event={"ID":"e14d1071-6d9a-4543-b593-8e999b7dcad2","Type":"ContainerStarted","Data":"36f683e077f94e7a8fc48a37bc417fd96f2eb3686b30cedc2c062f958e7a66ea"} Nov 25 15:11:40 crc kubenswrapper[4890]: I1125 15:11:40.666011 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-x7dwr" event={"ID":"d0f0970d-430a-4ea0-b18c-6af23eb40c71","Type":"ContainerStarted","Data":"75c914574b30a5a7f2319e126d3a4cceae977165fd88648e016f6effab049ff7"} Nov 25 15:11:40 crc kubenswrapper[4890]: I1125 15:11:40.668137 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-r5kd7" event={"ID":"3e0ccee5-4080-4973-99df-e4d6e929e471","Type":"ContainerStarted","Data":"2f6e5dc5778a0490494ed65c65486d060d45692582314c7528e3c19501126467"} Nov 25 15:11:44 crc kubenswrapper[4890]: I1125 15:11:44.697273 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-9c9wk" event={"ID":"e14d1071-6d9a-4543-b593-8e999b7dcad2","Type":"ContainerStarted","Data":"b8f2d09a82996fa87b1da07726363e7fa38591a8f4f4bdd13b0fdbbb73bb5c28"} Nov 25 15:11:44 crc kubenswrapper[4890]: I1125 15:11:44.699190 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-x7dwr" event={"ID":"d0f0970d-430a-4ea0-b18c-6af23eb40c71","Type":"ContainerStarted","Data":"91bc437d2b50db9aa97f6cabfe86828900be0cbc84357d63d8f0d1f5d780d628"} Nov 25 15:11:44 crc kubenswrapper[4890]: I1125 15:11:44.700957 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-r5kd7" event={"ID":"3e0ccee5-4080-4973-99df-e4d6e929e471","Type":"ContainerStarted","Data":"8d900728b78be69b41e7bdf1f8ef5670c6bfb80fb5e5d9801dead7aa2b04a4e6"} Nov 25 15:11:44 crc kubenswrapper[4890]: I1125 15:11:44.701145 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-r5kd7" Nov 25 15:11:44 crc kubenswrapper[4890]: I1125 15:11:44.717076 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-9c9wk" podStartSLOduration=2.383462681 podStartE2EDuration="5.717052054s" podCreationTimestamp="2025-11-25 15:11:39 +0000 UTC" firstStartedPulling="2025-11-25 15:11:40.497926623 +0000 UTC m=+558.940389243" lastFinishedPulling="2025-11-25 15:11:43.831516006 +0000 UTC m=+562.273978616" observedRunningTime="2025-11-25 15:11:44.712807378 +0000 UTC m=+563.155269998" watchObservedRunningTime="2025-11-25 15:11:44.717052054 +0000 UTC m=+563.159514664" Nov 25 15:11:44 crc kubenswrapper[4890]: I1125 15:11:44.738581 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-r5kd7" podStartSLOduration=2.317793157 podStartE2EDuration="5.738551732s" podCreationTimestamp="2025-11-25 15:11:39 +0000 UTC" firstStartedPulling="2025-11-25 15:11:40.38911133 +0000 UTC m=+558.831573940" lastFinishedPulling="2025-11-25 15:11:43.809869905 +0000 UTC m=+562.252332515" observedRunningTime="2025-11-25 15:11:44.738360818 +0000 UTC m=+563.180823428" watchObservedRunningTime="2025-11-25 15:11:44.738551732 +0000 UTC m=+563.181014342" Nov 25 15:11:44 crc kubenswrapper[4890]: I1125 15:11:44.764948 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-x7dwr" podStartSLOduration=2.500978602 podStartE2EDuration="5.764913352s" podCreationTimestamp="2025-11-25 15:11:39 +0000 UTC" firstStartedPulling="2025-11-25 15:11:40.545901374 +0000 UTC m=+558.988363984" lastFinishedPulling="2025-11-25 15:11:43.809836114 +0000 UTC m=+562.252298734" observedRunningTime="2025-11-25 15:11:44.75964149 +0000 UTC m=+563.202104090" watchObservedRunningTime="2025-11-25 15:11:44.764913352 +0000 UTC m=+563.207375982" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.147419 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7zdk8"] Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.148546 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" podUID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerName="ovn-controller" containerID="cri-o://9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100" gracePeriod=30 Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.148632 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" podUID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerName="nbdb" containerID="cri-o://7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a" gracePeriod=30 Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.148682 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" podUID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerName="northd" containerID="cri-o://89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91" gracePeriod=30 Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.148850 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" podUID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9" gracePeriod=30 Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.148916 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" podUID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerName="kube-rbac-proxy-node" containerID="cri-o://f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48" gracePeriod=30 Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.148958 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" podUID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerName="ovn-acl-logging" containerID="cri-o://bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a" gracePeriod=30 Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.149218 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" podUID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerName="sbdb" containerID="cri-o://c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611" gracePeriod=30 Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.149753 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-r5kd7" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.181226 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" podUID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerName="ovnkube-controller" containerID="cri-o://10d43df056314c8846655bab7a1e130886141cb09aee7b55ffd6845e3e982a19" gracePeriod=30 Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.472576 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7zdk8_9a404958-dffa-4a6a-99fc-3e4950614bd4/ovnkube-controller/2.log" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.476154 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7zdk8_9a404958-dffa-4a6a-99fc-3e4950614bd4/ovn-acl-logging/0.log" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.476839 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7zdk8_9a404958-dffa-4a6a-99fc-3e4950614bd4/ovn-controller/0.log" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.477470 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.535863 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-6vnct"] Nov 25 15:11:50 crc kubenswrapper[4890]: E1125 15:11:50.536114 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerName="ovnkube-controller" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.536135 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerName="ovnkube-controller" Nov 25 15:11:50 crc kubenswrapper[4890]: E1125 15:11:50.536146 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerName="ovnkube-controller" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.536153 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerName="ovnkube-controller" Nov 25 15:11:50 crc kubenswrapper[4890]: E1125 15:11:50.536197 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerName="kube-rbac-proxy-ovn-metrics" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.536206 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerName="kube-rbac-proxy-ovn-metrics" Nov 25 15:11:50 crc kubenswrapper[4890]: E1125 15:11:50.536215 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerName="sbdb" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.536223 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerName="sbdb" Nov 25 15:11:50 crc kubenswrapper[4890]: E1125 15:11:50.536230 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerName="ovn-acl-logging" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.536238 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerName="ovn-acl-logging" Nov 25 15:11:50 crc kubenswrapper[4890]: E1125 15:11:50.536246 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerName="ovnkube-controller" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.536253 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerName="ovnkube-controller" Nov 25 15:11:50 crc kubenswrapper[4890]: E1125 15:11:50.536265 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerName="kube-rbac-proxy-node" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.536272 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerName="kube-rbac-proxy-node" Nov 25 15:11:50 crc kubenswrapper[4890]: E1125 15:11:50.536280 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerName="ovn-controller" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.536287 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerName="ovn-controller" Nov 25 15:11:50 crc kubenswrapper[4890]: E1125 15:11:50.536302 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerName="nbdb" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.536309 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerName="nbdb" Nov 25 15:11:50 crc kubenswrapper[4890]: E1125 15:11:50.536322 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerName="kubecfg-setup" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.536329 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerName="kubecfg-setup" Nov 25 15:11:50 crc kubenswrapper[4890]: E1125 15:11:50.536342 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerName="northd" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.536349 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerName="northd" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.536462 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerName="nbdb" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.536475 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerName="ovnkube-controller" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.536483 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerName="ovn-acl-logging" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.536494 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerName="northd" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.536507 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerName="kube-rbac-proxy-node" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.536517 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerName="ovnkube-controller" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.536526 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerName="ovn-controller" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.536537 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerName="kube-rbac-proxy-ovn-metrics" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.536547 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerName="sbdb" Nov 25 15:11:50 crc kubenswrapper[4890]: E1125 15:11:50.536630 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerName="ovnkube-controller" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.536636 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerName="ovnkube-controller" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.536720 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerName="ovnkube-controller" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.536732 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerName="ovnkube-controller" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.538313 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.594494 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9a404958-dffa-4a6a-99fc-3e4950614bd4-ovnkube-config\") pod \"9a404958-dffa-4a6a-99fc-3e4950614bd4\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.594537 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-node-log\") pod \"9a404958-dffa-4a6a-99fc-3e4950614bd4\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.594563 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9a404958-dffa-4a6a-99fc-3e4950614bd4-ovn-node-metrics-cert\") pod \"9a404958-dffa-4a6a-99fc-3e4950614bd4\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.594580 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-log-socket\") pod \"9a404958-dffa-4a6a-99fc-3e4950614bd4\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.594595 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-run-systemd\") pod \"9a404958-dffa-4a6a-99fc-3e4950614bd4\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.594609 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-host-run-ovn-kubernetes\") pod \"9a404958-dffa-4a6a-99fc-3e4950614bd4\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.594639 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-host-run-netns\") pod \"9a404958-dffa-4a6a-99fc-3e4950614bd4\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.594656 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-var-lib-openvswitch\") pod \"9a404958-dffa-4a6a-99fc-3e4950614bd4\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.594684 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-node-log" (OuterVolumeSpecName: "node-log") pod "9a404958-dffa-4a6a-99fc-3e4950614bd4" (UID: "9a404958-dffa-4a6a-99fc-3e4950614bd4"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.594720 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "9a404958-dffa-4a6a-99fc-3e4950614bd4" (UID: "9a404958-dffa-4a6a-99fc-3e4950614bd4"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.594733 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "9a404958-dffa-4a6a-99fc-3e4950614bd4" (UID: "9a404958-dffa-4a6a-99fc-3e4950614bd4"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.594740 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "9a404958-dffa-4a6a-99fc-3e4950614bd4" (UID: "9a404958-dffa-4a6a-99fc-3e4950614bd4"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.594751 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-log-socket" (OuterVolumeSpecName: "log-socket") pod "9a404958-dffa-4a6a-99fc-3e4950614bd4" (UID: "9a404958-dffa-4a6a-99fc-3e4950614bd4"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.594757 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "9a404958-dffa-4a6a-99fc-3e4950614bd4" (UID: "9a404958-dffa-4a6a-99fc-3e4950614bd4"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.594691 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-host-cni-netd\") pod \"9a404958-dffa-4a6a-99fc-3e4950614bd4\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.594851 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-host-kubelet\") pod \"9a404958-dffa-4a6a-99fc-3e4950614bd4\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.594895 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92hnd\" (UniqueName: \"kubernetes.io/projected/9a404958-dffa-4a6a-99fc-3e4950614bd4-kube-api-access-92hnd\") pod \"9a404958-dffa-4a6a-99fc-3e4950614bd4\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.594919 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-systemd-units\") pod \"9a404958-dffa-4a6a-99fc-3e4950614bd4\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.595001 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "9a404958-dffa-4a6a-99fc-3e4950614bd4" (UID: "9a404958-dffa-4a6a-99fc-3e4950614bd4"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.594965 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-run-ovn\") pod \"9a404958-dffa-4a6a-99fc-3e4950614bd4\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.595046 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9a404958-dffa-4a6a-99fc-3e4950614bd4-env-overrides\") pod \"9a404958-dffa-4a6a-99fc-3e4950614bd4\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.595039 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "9a404958-dffa-4a6a-99fc-3e4950614bd4" (UID: "9a404958-dffa-4a6a-99fc-3e4950614bd4"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.595058 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "9a404958-dffa-4a6a-99fc-3e4950614bd4" (UID: "9a404958-dffa-4a6a-99fc-3e4950614bd4"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.595103 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a404958-dffa-4a6a-99fc-3e4950614bd4-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "9a404958-dffa-4a6a-99fc-3e4950614bd4" (UID: "9a404958-dffa-4a6a-99fc-3e4950614bd4"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.595193 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "9a404958-dffa-4a6a-99fc-3e4950614bd4" (UID: "9a404958-dffa-4a6a-99fc-3e4950614bd4"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.595418 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a404958-dffa-4a6a-99fc-3e4950614bd4-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "9a404958-dffa-4a6a-99fc-3e4950614bd4" (UID: "9a404958-dffa-4a6a-99fc-3e4950614bd4"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.595451 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-etc-openvswitch\") pod \"9a404958-dffa-4a6a-99fc-3e4950614bd4\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.595503 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9a404958-dffa-4a6a-99fc-3e4950614bd4-ovnkube-script-lib\") pod \"9a404958-dffa-4a6a-99fc-3e4950614bd4\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.595525 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-host-slash\") pod \"9a404958-dffa-4a6a-99fc-3e4950614bd4\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.595779 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-host-slash" (OuterVolumeSpecName: "host-slash") pod "9a404958-dffa-4a6a-99fc-3e4950614bd4" (UID: "9a404958-dffa-4a6a-99fc-3e4950614bd4"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.595879 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a404958-dffa-4a6a-99fc-3e4950614bd4-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "9a404958-dffa-4a6a-99fc-3e4950614bd4" (UID: "9a404958-dffa-4a6a-99fc-3e4950614bd4"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.595949 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-host-cni-bin\") pod \"9a404958-dffa-4a6a-99fc-3e4950614bd4\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.595972 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-host-var-lib-cni-networks-ovn-kubernetes\") pod \"9a404958-dffa-4a6a-99fc-3e4950614bd4\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.596023 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "9a404958-dffa-4a6a-99fc-3e4950614bd4" (UID: "9a404958-dffa-4a6a-99fc-3e4950614bd4"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.596056 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-run-openvswitch\") pod \"9a404958-dffa-4a6a-99fc-3e4950614bd4\" (UID: \"9a404958-dffa-4a6a-99fc-3e4950614bd4\") " Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.596110 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "9a404958-dffa-4a6a-99fc-3e4950614bd4" (UID: "9a404958-dffa-4a6a-99fc-3e4950614bd4"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.596185 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "9a404958-dffa-4a6a-99fc-3e4950614bd4" (UID: "9a404958-dffa-4a6a-99fc-3e4950614bd4"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.596365 4890 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.596387 4890 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9a404958-dffa-4a6a-99fc-3e4950614bd4-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.596398 4890 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-host-slash\") on node \"crc\" DevicePath \"\"" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.596409 4890 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-host-cni-bin\") on node \"crc\" DevicePath \"\"" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.596420 4890 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.596431 4890 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-run-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.596442 4890 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9a404958-dffa-4a6a-99fc-3e4950614bd4-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.596452 4890 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-node-log\") on node \"crc\" DevicePath \"\"" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.596463 4890 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-log-socket\") on node \"crc\" DevicePath \"\"" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.596473 4890 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.596482 4890 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-host-run-netns\") on node \"crc\" DevicePath \"\"" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.596492 4890 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.596501 4890 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-host-cni-netd\") on node \"crc\" DevicePath \"\"" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.596510 4890 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-host-kubelet\") on node \"crc\" DevicePath \"\"" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.596520 4890 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-systemd-units\") on node \"crc\" DevicePath \"\"" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.596531 4890 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.596541 4890 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9a404958-dffa-4a6a-99fc-3e4950614bd4-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.599750 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a404958-dffa-4a6a-99fc-3e4950614bd4-kube-api-access-92hnd" (OuterVolumeSpecName: "kube-api-access-92hnd") pod "9a404958-dffa-4a6a-99fc-3e4950614bd4" (UID: "9a404958-dffa-4a6a-99fc-3e4950614bd4"). InnerVolumeSpecName "kube-api-access-92hnd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.599921 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a404958-dffa-4a6a-99fc-3e4950614bd4-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "9a404958-dffa-4a6a-99fc-3e4950614bd4" (UID: "9a404958-dffa-4a6a-99fc-3e4950614bd4"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.607009 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "9a404958-dffa-4a6a-99fc-3e4950614bd4" (UID: "9a404958-dffa-4a6a-99fc-3e4950614bd4"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.698076 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e395c87e-30ef-495f-a38f-d3bf5fdcac14-host-cni-bin\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.698185 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e395c87e-30ef-495f-a38f-d3bf5fdcac14-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.698223 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e395c87e-30ef-495f-a38f-d3bf5fdcac14-var-lib-openvswitch\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.698269 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e395c87e-30ef-495f-a38f-d3bf5fdcac14-host-kubelet\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.698361 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e395c87e-30ef-495f-a38f-d3bf5fdcac14-ovnkube-config\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.698420 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e395c87e-30ef-495f-a38f-d3bf5fdcac14-host-run-ovn-kubernetes\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.698474 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e395c87e-30ef-495f-a38f-d3bf5fdcac14-host-run-netns\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.698518 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e395c87e-30ef-495f-a38f-d3bf5fdcac14-log-socket\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.698560 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e395c87e-30ef-495f-a38f-d3bf5fdcac14-ovn-node-metrics-cert\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.698624 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e395c87e-30ef-495f-a38f-d3bf5fdcac14-systemd-units\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.698656 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e395c87e-30ef-495f-a38f-d3bf5fdcac14-etc-openvswitch\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.698679 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9st7x\" (UniqueName: \"kubernetes.io/projected/e395c87e-30ef-495f-a38f-d3bf5fdcac14-kube-api-access-9st7x\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.698698 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e395c87e-30ef-495f-a38f-d3bf5fdcac14-run-systemd\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.698716 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e395c87e-30ef-495f-a38f-d3bf5fdcac14-host-cni-netd\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.698819 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e395c87e-30ef-495f-a38f-d3bf5fdcac14-ovnkube-script-lib\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.698876 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e395c87e-30ef-495f-a38f-d3bf5fdcac14-host-slash\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.698918 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e395c87e-30ef-495f-a38f-d3bf5fdcac14-run-ovn\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.698963 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e395c87e-30ef-495f-a38f-d3bf5fdcac14-run-openvswitch\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.699056 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e395c87e-30ef-495f-a38f-d3bf5fdcac14-env-overrides\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.699152 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e395c87e-30ef-495f-a38f-d3bf5fdcac14-node-log\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.699286 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92hnd\" (UniqueName: \"kubernetes.io/projected/9a404958-dffa-4a6a-99fc-3e4950614bd4-kube-api-access-92hnd\") on node \"crc\" DevicePath \"\"" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.699311 4890 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9a404958-dffa-4a6a-99fc-3e4950614bd4-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.699359 4890 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9a404958-dffa-4a6a-99fc-3e4950614bd4-run-systemd\") on node \"crc\" DevicePath \"\"" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.742506 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4txd2_0fcdeaed-8701-42ad-8534-a77b7f3f93f6/kube-multus/1.log" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.742887 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4txd2_0fcdeaed-8701-42ad-8534-a77b7f3f93f6/kube-multus/0.log" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.742919 4890 generic.go:334] "Generic (PLEG): container finished" podID="0fcdeaed-8701-42ad-8534-a77b7f3f93f6" containerID="4168b721ad650800e545d62906f69c441da51f6d8f11dd2537c8f2b2301afc2b" exitCode=2 Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.742972 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4txd2" event={"ID":"0fcdeaed-8701-42ad-8534-a77b7f3f93f6","Type":"ContainerDied","Data":"4168b721ad650800e545d62906f69c441da51f6d8f11dd2537c8f2b2301afc2b"} Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.743008 4890 scope.go:117] "RemoveContainer" containerID="b4322dbd78237bf443a985ce3c79bca4cba2103862d81177f5474a76f4905eed" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.743412 4890 scope.go:117] "RemoveContainer" containerID="4168b721ad650800e545d62906f69c441da51f6d8f11dd2537c8f2b2301afc2b" Nov 25 15:11:50 crc kubenswrapper[4890]: E1125 15:11:50.743580 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-4txd2_openshift-multus(0fcdeaed-8701-42ad-8534-a77b7f3f93f6)\"" pod="openshift-multus/multus-4txd2" podUID="0fcdeaed-8701-42ad-8534-a77b7f3f93f6" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.751074 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7zdk8_9a404958-dffa-4a6a-99fc-3e4950614bd4/ovnkube-controller/2.log" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.754228 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7zdk8_9a404958-dffa-4a6a-99fc-3e4950614bd4/ovn-acl-logging/0.log" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.754690 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7zdk8_9a404958-dffa-4a6a-99fc-3e4950614bd4/ovn-controller/0.log" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755083 4890 generic.go:334] "Generic (PLEG): container finished" podID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerID="10d43df056314c8846655bab7a1e130886141cb09aee7b55ffd6845e3e982a19" exitCode=0 Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755106 4890 generic.go:334] "Generic (PLEG): container finished" podID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerID="c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611" exitCode=0 Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755113 4890 generic.go:334] "Generic (PLEG): container finished" podID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerID="7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a" exitCode=0 Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755119 4890 generic.go:334] "Generic (PLEG): container finished" podID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerID="89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91" exitCode=0 Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755125 4890 generic.go:334] "Generic (PLEG): container finished" podID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerID="cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9" exitCode=0 Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755131 4890 generic.go:334] "Generic (PLEG): container finished" podID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerID="f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48" exitCode=0 Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755137 4890 generic.go:334] "Generic (PLEG): container finished" podID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerID="bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a" exitCode=143 Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755144 4890 generic.go:334] "Generic (PLEG): container finished" podID="9a404958-dffa-4a6a-99fc-3e4950614bd4" containerID="9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100" exitCode=143 Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755179 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755183 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" event={"ID":"9a404958-dffa-4a6a-99fc-3e4950614bd4","Type":"ContainerDied","Data":"10d43df056314c8846655bab7a1e130886141cb09aee7b55ffd6845e3e982a19"} Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755213 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" event={"ID":"9a404958-dffa-4a6a-99fc-3e4950614bd4","Type":"ContainerDied","Data":"c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611"} Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755224 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" event={"ID":"9a404958-dffa-4a6a-99fc-3e4950614bd4","Type":"ContainerDied","Data":"7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a"} Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755233 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" event={"ID":"9a404958-dffa-4a6a-99fc-3e4950614bd4","Type":"ContainerDied","Data":"89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91"} Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755243 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" event={"ID":"9a404958-dffa-4a6a-99fc-3e4950614bd4","Type":"ContainerDied","Data":"cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9"} Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755252 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" event={"ID":"9a404958-dffa-4a6a-99fc-3e4950614bd4","Type":"ContainerDied","Data":"f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48"} Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755262 4890 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"10d43df056314c8846655bab7a1e130886141cb09aee7b55ffd6845e3e982a19"} Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755271 4890 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fc71e729066adb3b4b41639d6028364bc27ae5a938d23ff9b00f7c813f207d8d"} Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755277 4890 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611"} Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755282 4890 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a"} Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755287 4890 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91"} Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755292 4890 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9"} Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755296 4890 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48"} Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755301 4890 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a"} Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755306 4890 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100"} Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755311 4890 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0"} Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755318 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" event={"ID":"9a404958-dffa-4a6a-99fc-3e4950614bd4","Type":"ContainerDied","Data":"bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a"} Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755325 4890 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"10d43df056314c8846655bab7a1e130886141cb09aee7b55ffd6845e3e982a19"} Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755331 4890 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fc71e729066adb3b4b41639d6028364bc27ae5a938d23ff9b00f7c813f207d8d"} Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755336 4890 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611"} Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755341 4890 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a"} Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755346 4890 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91"} Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755350 4890 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9"} Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755355 4890 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48"} Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755360 4890 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a"} Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755364 4890 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100"} Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755369 4890 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0"} Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755377 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" event={"ID":"9a404958-dffa-4a6a-99fc-3e4950614bd4","Type":"ContainerDied","Data":"9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100"} Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755384 4890 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"10d43df056314c8846655bab7a1e130886141cb09aee7b55ffd6845e3e982a19"} Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755389 4890 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fc71e729066adb3b4b41639d6028364bc27ae5a938d23ff9b00f7c813f207d8d"} Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755395 4890 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611"} Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755400 4890 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a"} Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755404 4890 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91"} Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755409 4890 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9"} Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755414 4890 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48"} Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755420 4890 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a"} Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755425 4890 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100"} Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755431 4890 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0"} Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755440 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7zdk8" event={"ID":"9a404958-dffa-4a6a-99fc-3e4950614bd4","Type":"ContainerDied","Data":"006be7dbdea159ba7aed943d4fdf527a2dfec84e73e9e193e2f96a4465006a9b"} Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755448 4890 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"10d43df056314c8846655bab7a1e130886141cb09aee7b55ffd6845e3e982a19"} Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755455 4890 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fc71e729066adb3b4b41639d6028364bc27ae5a938d23ff9b00f7c813f207d8d"} Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755462 4890 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611"} Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755469 4890 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a"} Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755476 4890 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91"} Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755483 4890 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9"} Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755489 4890 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48"} Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755496 4890 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a"} Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755502 4890 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100"} Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.755509 4890 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0"} Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.793613 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7zdk8"] Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.800504 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e395c87e-30ef-495f-a38f-d3bf5fdcac14-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.800545 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e395c87e-30ef-495f-a38f-d3bf5fdcac14-var-lib-openvswitch\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.800568 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e395c87e-30ef-495f-a38f-d3bf5fdcac14-host-kubelet\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.800597 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e395c87e-30ef-495f-a38f-d3bf5fdcac14-ovnkube-config\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.800615 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e395c87e-30ef-495f-a38f-d3bf5fdcac14-host-run-ovn-kubernetes\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.800637 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e395c87e-30ef-495f-a38f-d3bf5fdcac14-log-socket\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.800652 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e395c87e-30ef-495f-a38f-d3bf5fdcac14-ovn-node-metrics-cert\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.800657 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e395c87e-30ef-495f-a38f-d3bf5fdcac14-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.800666 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e395c87e-30ef-495f-a38f-d3bf5fdcac14-host-run-netns\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.800706 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e395c87e-30ef-495f-a38f-d3bf5fdcac14-host-run-netns\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.800742 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e395c87e-30ef-495f-a38f-d3bf5fdcac14-var-lib-openvswitch\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.800734 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e395c87e-30ef-495f-a38f-d3bf5fdcac14-systemd-units\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.800766 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e395c87e-30ef-495f-a38f-d3bf5fdcac14-host-kubelet\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.800774 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e395c87e-30ef-495f-a38f-d3bf5fdcac14-etc-openvswitch\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.800811 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9st7x\" (UniqueName: \"kubernetes.io/projected/e395c87e-30ef-495f-a38f-d3bf5fdcac14-kube-api-access-9st7x\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.800840 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e395c87e-30ef-495f-a38f-d3bf5fdcac14-run-systemd\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.800858 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e395c87e-30ef-495f-a38f-d3bf5fdcac14-host-cni-netd\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.800902 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e395c87e-30ef-495f-a38f-d3bf5fdcac14-ovnkube-script-lib\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.800916 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e395c87e-30ef-495f-a38f-d3bf5fdcac14-host-slash\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.800930 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e395c87e-30ef-495f-a38f-d3bf5fdcac14-run-ovn\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.800944 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e395c87e-30ef-495f-a38f-d3bf5fdcac14-run-openvswitch\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.800961 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e395c87e-30ef-495f-a38f-d3bf5fdcac14-env-overrides\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.800993 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e395c87e-30ef-495f-a38f-d3bf5fdcac14-node-log\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.801017 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e395c87e-30ef-495f-a38f-d3bf5fdcac14-host-cni-bin\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.801142 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e395c87e-30ef-495f-a38f-d3bf5fdcac14-host-cni-bin\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.801179 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e395c87e-30ef-495f-a38f-d3bf5fdcac14-systemd-units\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.801202 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e395c87e-30ef-495f-a38f-d3bf5fdcac14-etc-openvswitch\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.801446 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e395c87e-30ef-495f-a38f-d3bf5fdcac14-ovnkube-config\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.801454 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e395c87e-30ef-495f-a38f-d3bf5fdcac14-run-systemd\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.801481 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e395c87e-30ef-495f-a38f-d3bf5fdcac14-host-cni-netd\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.801488 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e395c87e-30ef-495f-a38f-d3bf5fdcac14-host-run-ovn-kubernetes\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.801515 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e395c87e-30ef-495f-a38f-d3bf5fdcac14-host-slash\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.801539 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e395c87e-30ef-495f-a38f-d3bf5fdcac14-run-ovn\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.801559 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e395c87e-30ef-495f-a38f-d3bf5fdcac14-run-openvswitch\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.801906 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e395c87e-30ef-495f-a38f-d3bf5fdcac14-ovnkube-script-lib\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.801939 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e395c87e-30ef-495f-a38f-d3bf5fdcac14-log-socket\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.801945 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e395c87e-30ef-495f-a38f-d3bf5fdcac14-env-overrides\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.801980 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e395c87e-30ef-495f-a38f-d3bf5fdcac14-node-log\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.802986 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7zdk8"] Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.804831 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e395c87e-30ef-495f-a38f-d3bf5fdcac14-ovn-node-metrics-cert\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.808523 4890 scope.go:117] "RemoveContainer" containerID="10d43df056314c8846655bab7a1e130886141cb09aee7b55ffd6845e3e982a19" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.816880 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9st7x\" (UniqueName: \"kubernetes.io/projected/e395c87e-30ef-495f-a38f-d3bf5fdcac14-kube-api-access-9st7x\") pod \"ovnkube-node-6vnct\" (UID: \"e395c87e-30ef-495f-a38f-d3bf5fdcac14\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.827063 4890 scope.go:117] "RemoveContainer" containerID="fc71e729066adb3b4b41639d6028364bc27ae5a938d23ff9b00f7c813f207d8d" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.842353 4890 scope.go:117] "RemoveContainer" containerID="c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.854580 4890 scope.go:117] "RemoveContainer" containerID="7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.854951 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.868722 4890 scope.go:117] "RemoveContainer" containerID="89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.881275 4890 scope.go:117] "RemoveContainer" containerID="cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.897314 4890 scope.go:117] "RemoveContainer" containerID="f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.909318 4890 scope.go:117] "RemoveContainer" containerID="bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.928478 4890 scope.go:117] "RemoveContainer" containerID="9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.950432 4890 scope.go:117] "RemoveContainer" containerID="50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.966324 4890 scope.go:117] "RemoveContainer" containerID="10d43df056314c8846655bab7a1e130886141cb09aee7b55ffd6845e3e982a19" Nov 25 15:11:50 crc kubenswrapper[4890]: E1125 15:11:50.966698 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10d43df056314c8846655bab7a1e130886141cb09aee7b55ffd6845e3e982a19\": container with ID starting with 10d43df056314c8846655bab7a1e130886141cb09aee7b55ffd6845e3e982a19 not found: ID does not exist" containerID="10d43df056314c8846655bab7a1e130886141cb09aee7b55ffd6845e3e982a19" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.966733 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10d43df056314c8846655bab7a1e130886141cb09aee7b55ffd6845e3e982a19"} err="failed to get container status \"10d43df056314c8846655bab7a1e130886141cb09aee7b55ffd6845e3e982a19\": rpc error: code = NotFound desc = could not find container \"10d43df056314c8846655bab7a1e130886141cb09aee7b55ffd6845e3e982a19\": container with ID starting with 10d43df056314c8846655bab7a1e130886141cb09aee7b55ffd6845e3e982a19 not found: ID does not exist" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.966760 4890 scope.go:117] "RemoveContainer" containerID="fc71e729066adb3b4b41639d6028364bc27ae5a938d23ff9b00f7c813f207d8d" Nov 25 15:11:50 crc kubenswrapper[4890]: E1125 15:11:50.966959 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc71e729066adb3b4b41639d6028364bc27ae5a938d23ff9b00f7c813f207d8d\": container with ID starting with fc71e729066adb3b4b41639d6028364bc27ae5a938d23ff9b00f7c813f207d8d not found: ID does not exist" containerID="fc71e729066adb3b4b41639d6028364bc27ae5a938d23ff9b00f7c813f207d8d" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.966986 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc71e729066adb3b4b41639d6028364bc27ae5a938d23ff9b00f7c813f207d8d"} err="failed to get container status \"fc71e729066adb3b4b41639d6028364bc27ae5a938d23ff9b00f7c813f207d8d\": rpc error: code = NotFound desc = could not find container \"fc71e729066adb3b4b41639d6028364bc27ae5a938d23ff9b00f7c813f207d8d\": container with ID starting with fc71e729066adb3b4b41639d6028364bc27ae5a938d23ff9b00f7c813f207d8d not found: ID does not exist" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.967003 4890 scope.go:117] "RemoveContainer" containerID="c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611" Nov 25 15:11:50 crc kubenswrapper[4890]: E1125 15:11:50.967369 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611\": container with ID starting with c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611 not found: ID does not exist" containerID="c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.967400 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611"} err="failed to get container status \"c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611\": rpc error: code = NotFound desc = could not find container \"c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611\": container with ID starting with c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611 not found: ID does not exist" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.967423 4890 scope.go:117] "RemoveContainer" containerID="7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a" Nov 25 15:11:50 crc kubenswrapper[4890]: E1125 15:11:50.967659 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a\": container with ID starting with 7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a not found: ID does not exist" containerID="7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.967690 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a"} err="failed to get container status \"7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a\": rpc error: code = NotFound desc = could not find container \"7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a\": container with ID starting with 7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a not found: ID does not exist" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.967708 4890 scope.go:117] "RemoveContainer" containerID="89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91" Nov 25 15:11:50 crc kubenswrapper[4890]: E1125 15:11:50.967949 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91\": container with ID starting with 89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91 not found: ID does not exist" containerID="89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.967974 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91"} err="failed to get container status \"89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91\": rpc error: code = NotFound desc = could not find container \"89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91\": container with ID starting with 89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91 not found: ID does not exist" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.967985 4890 scope.go:117] "RemoveContainer" containerID="cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9" Nov 25 15:11:50 crc kubenswrapper[4890]: E1125 15:11:50.968213 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9\": container with ID starting with cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9 not found: ID does not exist" containerID="cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.968236 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9"} err="failed to get container status \"cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9\": rpc error: code = NotFound desc = could not find container \"cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9\": container with ID starting with cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9 not found: ID does not exist" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.968248 4890 scope.go:117] "RemoveContainer" containerID="f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48" Nov 25 15:11:50 crc kubenswrapper[4890]: E1125 15:11:50.968662 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48\": container with ID starting with f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48 not found: ID does not exist" containerID="f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.968691 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48"} err="failed to get container status \"f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48\": rpc error: code = NotFound desc = could not find container \"f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48\": container with ID starting with f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48 not found: ID does not exist" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.968708 4890 scope.go:117] "RemoveContainer" containerID="bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a" Nov 25 15:11:50 crc kubenswrapper[4890]: E1125 15:11:50.968931 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a\": container with ID starting with bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a not found: ID does not exist" containerID="bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.968958 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a"} err="failed to get container status \"bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a\": rpc error: code = NotFound desc = could not find container \"bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a\": container with ID starting with bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a not found: ID does not exist" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.968975 4890 scope.go:117] "RemoveContainer" containerID="9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100" Nov 25 15:11:50 crc kubenswrapper[4890]: E1125 15:11:50.969216 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100\": container with ID starting with 9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100 not found: ID does not exist" containerID="9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.969248 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100"} err="failed to get container status \"9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100\": rpc error: code = NotFound desc = could not find container \"9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100\": container with ID starting with 9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100 not found: ID does not exist" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.969270 4890 scope.go:117] "RemoveContainer" containerID="50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0" Nov 25 15:11:50 crc kubenswrapper[4890]: E1125 15:11:50.969523 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\": container with ID starting with 50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0 not found: ID does not exist" containerID="50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.969572 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0"} err="failed to get container status \"50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\": rpc error: code = NotFound desc = could not find container \"50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\": container with ID starting with 50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0 not found: ID does not exist" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.969588 4890 scope.go:117] "RemoveContainer" containerID="10d43df056314c8846655bab7a1e130886141cb09aee7b55ffd6845e3e982a19" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.969838 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10d43df056314c8846655bab7a1e130886141cb09aee7b55ffd6845e3e982a19"} err="failed to get container status \"10d43df056314c8846655bab7a1e130886141cb09aee7b55ffd6845e3e982a19\": rpc error: code = NotFound desc = could not find container \"10d43df056314c8846655bab7a1e130886141cb09aee7b55ffd6845e3e982a19\": container with ID starting with 10d43df056314c8846655bab7a1e130886141cb09aee7b55ffd6845e3e982a19 not found: ID does not exist" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.969871 4890 scope.go:117] "RemoveContainer" containerID="fc71e729066adb3b4b41639d6028364bc27ae5a938d23ff9b00f7c813f207d8d" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.970110 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc71e729066adb3b4b41639d6028364bc27ae5a938d23ff9b00f7c813f207d8d"} err="failed to get container status \"fc71e729066adb3b4b41639d6028364bc27ae5a938d23ff9b00f7c813f207d8d\": rpc error: code = NotFound desc = could not find container \"fc71e729066adb3b4b41639d6028364bc27ae5a938d23ff9b00f7c813f207d8d\": container with ID starting with fc71e729066adb3b4b41639d6028364bc27ae5a938d23ff9b00f7c813f207d8d not found: ID does not exist" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.970138 4890 scope.go:117] "RemoveContainer" containerID="c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.970395 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611"} err="failed to get container status \"c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611\": rpc error: code = NotFound desc = could not find container \"c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611\": container with ID starting with c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611 not found: ID does not exist" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.970421 4890 scope.go:117] "RemoveContainer" containerID="7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.970709 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a"} err="failed to get container status \"7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a\": rpc error: code = NotFound desc = could not find container \"7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a\": container with ID starting with 7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a not found: ID does not exist" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.970732 4890 scope.go:117] "RemoveContainer" containerID="89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.970949 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91"} err="failed to get container status \"89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91\": rpc error: code = NotFound desc = could not find container \"89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91\": container with ID starting with 89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91 not found: ID does not exist" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.970973 4890 scope.go:117] "RemoveContainer" containerID="cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.971197 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9"} err="failed to get container status \"cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9\": rpc error: code = NotFound desc = could not find container \"cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9\": container with ID starting with cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9 not found: ID does not exist" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.971222 4890 scope.go:117] "RemoveContainer" containerID="f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.971470 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48"} err="failed to get container status \"f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48\": rpc error: code = NotFound desc = could not find container \"f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48\": container with ID starting with f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48 not found: ID does not exist" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.971489 4890 scope.go:117] "RemoveContainer" containerID="bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.971681 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a"} err="failed to get container status \"bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a\": rpc error: code = NotFound desc = could not find container \"bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a\": container with ID starting with bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a not found: ID does not exist" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.971701 4890 scope.go:117] "RemoveContainer" containerID="9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.971902 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100"} err="failed to get container status \"9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100\": rpc error: code = NotFound desc = could not find container \"9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100\": container with ID starting with 9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100 not found: ID does not exist" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.971922 4890 scope.go:117] "RemoveContainer" containerID="50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.972112 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0"} err="failed to get container status \"50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\": rpc error: code = NotFound desc = could not find container \"50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\": container with ID starting with 50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0 not found: ID does not exist" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.972137 4890 scope.go:117] "RemoveContainer" containerID="10d43df056314c8846655bab7a1e130886141cb09aee7b55ffd6845e3e982a19" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.972339 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10d43df056314c8846655bab7a1e130886141cb09aee7b55ffd6845e3e982a19"} err="failed to get container status \"10d43df056314c8846655bab7a1e130886141cb09aee7b55ffd6845e3e982a19\": rpc error: code = NotFound desc = could not find container \"10d43df056314c8846655bab7a1e130886141cb09aee7b55ffd6845e3e982a19\": container with ID starting with 10d43df056314c8846655bab7a1e130886141cb09aee7b55ffd6845e3e982a19 not found: ID does not exist" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.972363 4890 scope.go:117] "RemoveContainer" containerID="fc71e729066adb3b4b41639d6028364bc27ae5a938d23ff9b00f7c813f207d8d" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.972545 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc71e729066adb3b4b41639d6028364bc27ae5a938d23ff9b00f7c813f207d8d"} err="failed to get container status \"fc71e729066adb3b4b41639d6028364bc27ae5a938d23ff9b00f7c813f207d8d\": rpc error: code = NotFound desc = could not find container \"fc71e729066adb3b4b41639d6028364bc27ae5a938d23ff9b00f7c813f207d8d\": container with ID starting with fc71e729066adb3b4b41639d6028364bc27ae5a938d23ff9b00f7c813f207d8d not found: ID does not exist" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.972568 4890 scope.go:117] "RemoveContainer" containerID="c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.972772 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611"} err="failed to get container status \"c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611\": rpc error: code = NotFound desc = could not find container \"c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611\": container with ID starting with c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611 not found: ID does not exist" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.972857 4890 scope.go:117] "RemoveContainer" containerID="7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.973068 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a"} err="failed to get container status \"7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a\": rpc error: code = NotFound desc = could not find container \"7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a\": container with ID starting with 7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a not found: ID does not exist" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.973092 4890 scope.go:117] "RemoveContainer" containerID="89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.973304 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91"} err="failed to get container status \"89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91\": rpc error: code = NotFound desc = could not find container \"89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91\": container with ID starting with 89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91 not found: ID does not exist" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.973331 4890 scope.go:117] "RemoveContainer" containerID="cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.973537 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9"} err="failed to get container status \"cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9\": rpc error: code = NotFound desc = could not find container \"cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9\": container with ID starting with cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9 not found: ID does not exist" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.973585 4890 scope.go:117] "RemoveContainer" containerID="f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.973793 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48"} err="failed to get container status \"f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48\": rpc error: code = NotFound desc = could not find container \"f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48\": container with ID starting with f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48 not found: ID does not exist" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.973820 4890 scope.go:117] "RemoveContainer" containerID="bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.974006 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a"} err="failed to get container status \"bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a\": rpc error: code = NotFound desc = could not find container \"bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a\": container with ID starting with bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a not found: ID does not exist" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.974030 4890 scope.go:117] "RemoveContainer" containerID="9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.974230 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100"} err="failed to get container status \"9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100\": rpc error: code = NotFound desc = could not find container \"9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100\": container with ID starting with 9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100 not found: ID does not exist" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.974254 4890 scope.go:117] "RemoveContainer" containerID="50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.974434 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0"} err="failed to get container status \"50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\": rpc error: code = NotFound desc = could not find container \"50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\": container with ID starting with 50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0 not found: ID does not exist" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.974457 4890 scope.go:117] "RemoveContainer" containerID="10d43df056314c8846655bab7a1e130886141cb09aee7b55ffd6845e3e982a19" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.974642 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10d43df056314c8846655bab7a1e130886141cb09aee7b55ffd6845e3e982a19"} err="failed to get container status \"10d43df056314c8846655bab7a1e130886141cb09aee7b55ffd6845e3e982a19\": rpc error: code = NotFound desc = could not find container \"10d43df056314c8846655bab7a1e130886141cb09aee7b55ffd6845e3e982a19\": container with ID starting with 10d43df056314c8846655bab7a1e130886141cb09aee7b55ffd6845e3e982a19 not found: ID does not exist" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.974666 4890 scope.go:117] "RemoveContainer" containerID="fc71e729066adb3b4b41639d6028364bc27ae5a938d23ff9b00f7c813f207d8d" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.974872 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc71e729066adb3b4b41639d6028364bc27ae5a938d23ff9b00f7c813f207d8d"} err="failed to get container status \"fc71e729066adb3b4b41639d6028364bc27ae5a938d23ff9b00f7c813f207d8d\": rpc error: code = NotFound desc = could not find container \"fc71e729066adb3b4b41639d6028364bc27ae5a938d23ff9b00f7c813f207d8d\": container with ID starting with fc71e729066adb3b4b41639d6028364bc27ae5a938d23ff9b00f7c813f207d8d not found: ID does not exist" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.974899 4890 scope.go:117] "RemoveContainer" containerID="c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.975093 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611"} err="failed to get container status \"c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611\": rpc error: code = NotFound desc = could not find container \"c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611\": container with ID starting with c09fc204b78b38a760b4715476afed5ffee1b7bf0e4d9c000d304e65dee66611 not found: ID does not exist" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.975119 4890 scope.go:117] "RemoveContainer" containerID="7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.975330 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a"} err="failed to get container status \"7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a\": rpc error: code = NotFound desc = could not find container \"7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a\": container with ID starting with 7cef96a0b4254e0d3d7e4dd1627bb547c0ca2131a39f1e7acc7de6b3c584c77a not found: ID does not exist" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.975356 4890 scope.go:117] "RemoveContainer" containerID="89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.975554 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91"} err="failed to get container status \"89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91\": rpc error: code = NotFound desc = could not find container \"89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91\": container with ID starting with 89b55057a99cc2edd0c0067d8c49d9f9cc3da35020fcf463fd1942e752372a91 not found: ID does not exist" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.975578 4890 scope.go:117] "RemoveContainer" containerID="cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.975780 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9"} err="failed to get container status \"cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9\": rpc error: code = NotFound desc = could not find container \"cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9\": container with ID starting with cc0b68404c2463b1de69f18c0a9d4d8ee32c84f0a941a529d070941156756eb9 not found: ID does not exist" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.975804 4890 scope.go:117] "RemoveContainer" containerID="f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.976019 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48"} err="failed to get container status \"f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48\": rpc error: code = NotFound desc = could not find container \"f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48\": container with ID starting with f48f3dd38b05025d637de538875ffe3d5ad44b06f217465e6f57fcbbdf052e48 not found: ID does not exist" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.976042 4890 scope.go:117] "RemoveContainer" containerID="bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.976303 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a"} err="failed to get container status \"bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a\": rpc error: code = NotFound desc = could not find container \"bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a\": container with ID starting with bd616baba3920bfb3445e3ece6812ad734922b7eac579848860a421873a7f45a not found: ID does not exist" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.976325 4890 scope.go:117] "RemoveContainer" containerID="9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.976510 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100"} err="failed to get container status \"9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100\": rpc error: code = NotFound desc = could not find container \"9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100\": container with ID starting with 9c8c4563d0d72ea7e3ed0d90da97a566b59d60c05882d15ff15f718ec619c100 not found: ID does not exist" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.976534 4890 scope.go:117] "RemoveContainer" containerID="50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0" Nov 25 15:11:50 crc kubenswrapper[4890]: I1125 15:11:50.976714 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0"} err="failed to get container status \"50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\": rpc error: code = NotFound desc = could not find container \"50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0\": container with ID starting with 50f4ec0b509350bc9054d3ee859211cc80b92ccb75ce3ae257794695eeb3fed0 not found: ID does not exist" Nov 25 15:11:51 crc kubenswrapper[4890]: I1125 15:11:51.762290 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4txd2_0fcdeaed-8701-42ad-8534-a77b7f3f93f6/kube-multus/1.log" Nov 25 15:11:51 crc kubenswrapper[4890]: I1125 15:11:51.764947 4890 generic.go:334] "Generic (PLEG): container finished" podID="e395c87e-30ef-495f-a38f-d3bf5fdcac14" containerID="9a8e5708c45b714ac177b04a96208228366b062c5eaeae966163b187714d4d55" exitCode=0 Nov 25 15:11:51 crc kubenswrapper[4890]: I1125 15:11:51.764974 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" event={"ID":"e395c87e-30ef-495f-a38f-d3bf5fdcac14","Type":"ContainerDied","Data":"9a8e5708c45b714ac177b04a96208228366b062c5eaeae966163b187714d4d55"} Nov 25 15:11:51 crc kubenswrapper[4890]: I1125 15:11:51.765012 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" event={"ID":"e395c87e-30ef-495f-a38f-d3bf5fdcac14","Type":"ContainerStarted","Data":"c4729f10c919fa0c70715c91a2de6e02f7e0751a604aeb29bf57cfd30d4aadc0"} Nov 25 15:11:52 crc kubenswrapper[4890]: I1125 15:11:52.179150 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a404958-dffa-4a6a-99fc-3e4950614bd4" path="/var/lib/kubelet/pods/9a404958-dffa-4a6a-99fc-3e4950614bd4/volumes" Nov 25 15:11:52 crc kubenswrapper[4890]: I1125 15:11:52.773632 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" event={"ID":"e395c87e-30ef-495f-a38f-d3bf5fdcac14","Type":"ContainerStarted","Data":"b82f07a9e180e665dcf2a456d224d21d834d3c5f9754b0a585352b8f22af94d3"} Nov 25 15:11:52 crc kubenswrapper[4890]: I1125 15:11:52.773955 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" event={"ID":"e395c87e-30ef-495f-a38f-d3bf5fdcac14","Type":"ContainerStarted","Data":"aa8f4182e3225d960aa2e6cea7ed9b5e68bf480b0f89e3e8427b39720d9db6ff"} Nov 25 15:11:52 crc kubenswrapper[4890]: I1125 15:11:52.773966 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" event={"ID":"e395c87e-30ef-495f-a38f-d3bf5fdcac14","Type":"ContainerStarted","Data":"f7665b850bbed4dea5e8f85dc12f7e9c36b4f8bbf2248dd229c9bd105dd5c3d8"} Nov 25 15:11:52 crc kubenswrapper[4890]: I1125 15:11:52.773976 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" event={"ID":"e395c87e-30ef-495f-a38f-d3bf5fdcac14","Type":"ContainerStarted","Data":"4702a27159a9a728ad5e5abe4fb3bba66e07d9dada2b59399b98fae3cad8ba99"} Nov 25 15:11:52 crc kubenswrapper[4890]: I1125 15:11:52.773984 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" event={"ID":"e395c87e-30ef-495f-a38f-d3bf5fdcac14","Type":"ContainerStarted","Data":"237ae6351f03479690c3234ac4bb60ec1b182ffddce257aaa8a1ede09447e923"} Nov 25 15:11:52 crc kubenswrapper[4890]: I1125 15:11:52.773993 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" event={"ID":"e395c87e-30ef-495f-a38f-d3bf5fdcac14","Type":"ContainerStarted","Data":"5f793974b8322fe563fd9d6b9f69896328fd6f29e79589fa4fabe19b9b6404e1"} Nov 25 15:11:55 crc kubenswrapper[4890]: I1125 15:11:55.796662 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" event={"ID":"e395c87e-30ef-495f-a38f-d3bf5fdcac14","Type":"ContainerStarted","Data":"96416828d9e6863b0d11b3d568a7953009bbdbfecf632395efd3f547b2e4b9f7"} Nov 25 15:11:56 crc kubenswrapper[4890]: I1125 15:11:56.448068 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 15:11:56 crc kubenswrapper[4890]: I1125 15:11:56.448498 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 15:11:56 crc kubenswrapper[4890]: I1125 15:11:56.448550 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" Nov 25 15:11:56 crc kubenswrapper[4890]: I1125 15:11:56.449221 4890 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1a14507b222e3efb8d4e5a5a36210ae167e2bc170367e62ff6f845c2d500391c"} pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 15:11:56 crc kubenswrapper[4890]: I1125 15:11:56.449282 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" containerID="cri-o://1a14507b222e3efb8d4e5a5a36210ae167e2bc170367e62ff6f845c2d500391c" gracePeriod=600 Nov 25 15:11:56 crc kubenswrapper[4890]: I1125 15:11:56.804834 4890 generic.go:334] "Generic (PLEG): container finished" podID="4e4f849d-f239-4727-a73e-18327856929a" containerID="1a14507b222e3efb8d4e5a5a36210ae167e2bc170367e62ff6f845c2d500391c" exitCode=0 Nov 25 15:11:56 crc kubenswrapper[4890]: I1125 15:11:56.804879 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" event={"ID":"4e4f849d-f239-4727-a73e-18327856929a","Type":"ContainerDied","Data":"1a14507b222e3efb8d4e5a5a36210ae167e2bc170367e62ff6f845c2d500391c"} Nov 25 15:11:56 crc kubenswrapper[4890]: I1125 15:11:56.804907 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" event={"ID":"4e4f849d-f239-4727-a73e-18327856929a","Type":"ContainerStarted","Data":"361ef05d7f3bcf13e88a1632cd64dd7b6044add1c3fc4ac44bf729fc819bec71"} Nov 25 15:11:56 crc kubenswrapper[4890]: I1125 15:11:56.804934 4890 scope.go:117] "RemoveContainer" containerID="2f432adbd81e026db39d6c93118322aceca51fcd04a65e4110d7e407d22f0ff0" Nov 25 15:11:57 crc kubenswrapper[4890]: I1125 15:11:57.814882 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" event={"ID":"e395c87e-30ef-495f-a38f-d3bf5fdcac14","Type":"ContainerStarted","Data":"c3b694fe191a9483fa0492ee5ecaa110b79538c65a9901da86d350d7e03430f7"} Nov 25 15:11:57 crc kubenswrapper[4890]: I1125 15:11:57.815455 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:57 crc kubenswrapper[4890]: I1125 15:11:57.815469 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:57 crc kubenswrapper[4890]: I1125 15:11:57.839982 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" podStartSLOduration=7.839958833 podStartE2EDuration="7.839958833s" podCreationTimestamp="2025-11-25 15:11:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:11:57.838713131 +0000 UTC m=+576.281175771" watchObservedRunningTime="2025-11-25 15:11:57.839958833 +0000 UTC m=+576.282421463" Nov 25 15:11:57 crc kubenswrapper[4890]: I1125 15:11:57.846005 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:58 crc kubenswrapper[4890]: I1125 15:11:58.819341 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:11:58 crc kubenswrapper[4890]: I1125 15:11:58.852143 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:12:06 crc kubenswrapper[4890]: I1125 15:12:06.173939 4890 scope.go:117] "RemoveContainer" containerID="4168b721ad650800e545d62906f69c441da51f6d8f11dd2537c8f2b2301afc2b" Nov 25 15:12:06 crc kubenswrapper[4890]: I1125 15:12:06.876747 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4txd2_0fcdeaed-8701-42ad-8534-a77b7f3f93f6/kube-multus/1.log" Nov 25 15:12:06 crc kubenswrapper[4890]: I1125 15:12:06.877101 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4txd2" event={"ID":"0fcdeaed-8701-42ad-8534-a77b7f3f93f6","Type":"ContainerStarted","Data":"3932124bc7e17dc55f2beafe1e473126fe4f68871cb3a630fba2c21055d49b70"} Nov 25 15:12:20 crc kubenswrapper[4890]: I1125 15:12:20.883950 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6vnct" Nov 25 15:12:29 crc kubenswrapper[4890]: I1125 15:12:29.994502 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772er4qx8"] Nov 25 15:12:29 crc kubenswrapper[4890]: I1125 15:12:29.997381 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772er4qx8" Nov 25 15:12:29 crc kubenswrapper[4890]: I1125 15:12:29.999932 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 25 15:12:30 crc kubenswrapper[4890]: I1125 15:12:30.019033 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772er4qx8"] Nov 25 15:12:30 crc kubenswrapper[4890]: I1125 15:12:30.051276 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d89e1065-7553-4d5b-aef6-38108a5cb8ae-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772er4qx8\" (UID: \"d89e1065-7553-4d5b-aef6-38108a5cb8ae\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772er4qx8" Nov 25 15:12:30 crc kubenswrapper[4890]: I1125 15:12:30.051325 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d89e1065-7553-4d5b-aef6-38108a5cb8ae-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772er4qx8\" (UID: \"d89e1065-7553-4d5b-aef6-38108a5cb8ae\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772er4qx8" Nov 25 15:12:30 crc kubenswrapper[4890]: I1125 15:12:30.051392 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95722\" (UniqueName: \"kubernetes.io/projected/d89e1065-7553-4d5b-aef6-38108a5cb8ae-kube-api-access-95722\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772er4qx8\" (UID: \"d89e1065-7553-4d5b-aef6-38108a5cb8ae\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772er4qx8" Nov 25 15:12:30 crc kubenswrapper[4890]: I1125 15:12:30.152521 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d89e1065-7553-4d5b-aef6-38108a5cb8ae-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772er4qx8\" (UID: \"d89e1065-7553-4d5b-aef6-38108a5cb8ae\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772er4qx8" Nov 25 15:12:30 crc kubenswrapper[4890]: I1125 15:12:30.152653 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d89e1065-7553-4d5b-aef6-38108a5cb8ae-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772er4qx8\" (UID: \"d89e1065-7553-4d5b-aef6-38108a5cb8ae\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772er4qx8" Nov 25 15:12:30 crc kubenswrapper[4890]: I1125 15:12:30.152758 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95722\" (UniqueName: \"kubernetes.io/projected/d89e1065-7553-4d5b-aef6-38108a5cb8ae-kube-api-access-95722\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772er4qx8\" (UID: \"d89e1065-7553-4d5b-aef6-38108a5cb8ae\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772er4qx8" Nov 25 15:12:30 crc kubenswrapper[4890]: I1125 15:12:30.153616 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d89e1065-7553-4d5b-aef6-38108a5cb8ae-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772er4qx8\" (UID: \"d89e1065-7553-4d5b-aef6-38108a5cb8ae\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772er4qx8" Nov 25 15:12:30 crc kubenswrapper[4890]: I1125 15:12:30.153616 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d89e1065-7553-4d5b-aef6-38108a5cb8ae-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772er4qx8\" (UID: \"d89e1065-7553-4d5b-aef6-38108a5cb8ae\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772er4qx8" Nov 25 15:12:30 crc kubenswrapper[4890]: I1125 15:12:30.175950 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95722\" (UniqueName: \"kubernetes.io/projected/d89e1065-7553-4d5b-aef6-38108a5cb8ae-kube-api-access-95722\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772er4qx8\" (UID: \"d89e1065-7553-4d5b-aef6-38108a5cb8ae\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772er4qx8" Nov 25 15:12:30 crc kubenswrapper[4890]: I1125 15:12:30.328325 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772er4qx8" Nov 25 15:12:30 crc kubenswrapper[4890]: I1125 15:12:30.565104 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772er4qx8"] Nov 25 15:12:31 crc kubenswrapper[4890]: I1125 15:12:31.014026 4890 generic.go:334] "Generic (PLEG): container finished" podID="d89e1065-7553-4d5b-aef6-38108a5cb8ae" containerID="ec317e74d545e888dcafbfa3a629a1bf0aa234a970198eb8d803387680292ea4" exitCode=0 Nov 25 15:12:31 crc kubenswrapper[4890]: I1125 15:12:31.014097 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772er4qx8" event={"ID":"d89e1065-7553-4d5b-aef6-38108a5cb8ae","Type":"ContainerDied","Data":"ec317e74d545e888dcafbfa3a629a1bf0aa234a970198eb8d803387680292ea4"} Nov 25 15:12:31 crc kubenswrapper[4890]: I1125 15:12:31.014148 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772er4qx8" event={"ID":"d89e1065-7553-4d5b-aef6-38108a5cb8ae","Type":"ContainerStarted","Data":"00df1ecd782216da59db723daf35dc4bda80cbde86c8078a48bfbe8ff818bcc8"} Nov 25 15:12:33 crc kubenswrapper[4890]: I1125 15:12:33.028184 4890 generic.go:334] "Generic (PLEG): container finished" podID="d89e1065-7553-4d5b-aef6-38108a5cb8ae" containerID="f212193084a5142803bbab7b409dbef3af1cf748df3062555e38c1b36bb0891b" exitCode=0 Nov 25 15:12:33 crc kubenswrapper[4890]: I1125 15:12:33.028243 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772er4qx8" event={"ID":"d89e1065-7553-4d5b-aef6-38108a5cb8ae","Type":"ContainerDied","Data":"f212193084a5142803bbab7b409dbef3af1cf748df3062555e38c1b36bb0891b"} Nov 25 15:12:34 crc kubenswrapper[4890]: I1125 15:12:34.034383 4890 generic.go:334] "Generic (PLEG): container finished" podID="d89e1065-7553-4d5b-aef6-38108a5cb8ae" containerID="7e1d2bff7aed52e78829f0c407187f92d20b0598d2a7d7434eef1455a1630b13" exitCode=0 Nov 25 15:12:34 crc kubenswrapper[4890]: I1125 15:12:34.034425 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772er4qx8" event={"ID":"d89e1065-7553-4d5b-aef6-38108a5cb8ae","Type":"ContainerDied","Data":"7e1d2bff7aed52e78829f0c407187f92d20b0598d2a7d7434eef1455a1630b13"} Nov 25 15:12:35 crc kubenswrapper[4890]: I1125 15:12:35.259918 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772er4qx8" Nov 25 15:12:35 crc kubenswrapper[4890]: I1125 15:12:35.420775 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d89e1065-7553-4d5b-aef6-38108a5cb8ae-bundle\") pod \"d89e1065-7553-4d5b-aef6-38108a5cb8ae\" (UID: \"d89e1065-7553-4d5b-aef6-38108a5cb8ae\") " Nov 25 15:12:35 crc kubenswrapper[4890]: I1125 15:12:35.420928 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95722\" (UniqueName: \"kubernetes.io/projected/d89e1065-7553-4d5b-aef6-38108a5cb8ae-kube-api-access-95722\") pod \"d89e1065-7553-4d5b-aef6-38108a5cb8ae\" (UID: \"d89e1065-7553-4d5b-aef6-38108a5cb8ae\") " Nov 25 15:12:35 crc kubenswrapper[4890]: I1125 15:12:35.420976 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d89e1065-7553-4d5b-aef6-38108a5cb8ae-util\") pod \"d89e1065-7553-4d5b-aef6-38108a5cb8ae\" (UID: \"d89e1065-7553-4d5b-aef6-38108a5cb8ae\") " Nov 25 15:12:35 crc kubenswrapper[4890]: I1125 15:12:35.422589 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d89e1065-7553-4d5b-aef6-38108a5cb8ae-bundle" (OuterVolumeSpecName: "bundle") pod "d89e1065-7553-4d5b-aef6-38108a5cb8ae" (UID: "d89e1065-7553-4d5b-aef6-38108a5cb8ae"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:12:35 crc kubenswrapper[4890]: I1125 15:12:35.427249 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d89e1065-7553-4d5b-aef6-38108a5cb8ae-kube-api-access-95722" (OuterVolumeSpecName: "kube-api-access-95722") pod "d89e1065-7553-4d5b-aef6-38108a5cb8ae" (UID: "d89e1065-7553-4d5b-aef6-38108a5cb8ae"). InnerVolumeSpecName "kube-api-access-95722". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:12:35 crc kubenswrapper[4890]: I1125 15:12:35.449915 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d89e1065-7553-4d5b-aef6-38108a5cb8ae-util" (OuterVolumeSpecName: "util") pod "d89e1065-7553-4d5b-aef6-38108a5cb8ae" (UID: "d89e1065-7553-4d5b-aef6-38108a5cb8ae"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:12:35 crc kubenswrapper[4890]: I1125 15:12:35.522884 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95722\" (UniqueName: \"kubernetes.io/projected/d89e1065-7553-4d5b-aef6-38108a5cb8ae-kube-api-access-95722\") on node \"crc\" DevicePath \"\"" Nov 25 15:12:35 crc kubenswrapper[4890]: I1125 15:12:35.522934 4890 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d89e1065-7553-4d5b-aef6-38108a5cb8ae-util\") on node \"crc\" DevicePath \"\"" Nov 25 15:12:35 crc kubenswrapper[4890]: I1125 15:12:35.522948 4890 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d89e1065-7553-4d5b-aef6-38108a5cb8ae-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:12:36 crc kubenswrapper[4890]: I1125 15:12:36.081048 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772er4qx8" event={"ID":"d89e1065-7553-4d5b-aef6-38108a5cb8ae","Type":"ContainerDied","Data":"00df1ecd782216da59db723daf35dc4bda80cbde86c8078a48bfbe8ff818bcc8"} Nov 25 15:12:36 crc kubenswrapper[4890]: I1125 15:12:36.081105 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="00df1ecd782216da59db723daf35dc4bda80cbde86c8078a48bfbe8ff818bcc8" Nov 25 15:12:36 crc kubenswrapper[4890]: I1125 15:12:36.081130 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772er4qx8" Nov 25 15:12:38 crc kubenswrapper[4890]: I1125 15:12:38.822823 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-v4blp"] Nov 25 15:12:38 crc kubenswrapper[4890]: E1125 15:12:38.823427 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d89e1065-7553-4d5b-aef6-38108a5cb8ae" containerName="pull" Nov 25 15:12:38 crc kubenswrapper[4890]: I1125 15:12:38.823442 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="d89e1065-7553-4d5b-aef6-38108a5cb8ae" containerName="pull" Nov 25 15:12:38 crc kubenswrapper[4890]: E1125 15:12:38.823462 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d89e1065-7553-4d5b-aef6-38108a5cb8ae" containerName="util" Nov 25 15:12:38 crc kubenswrapper[4890]: I1125 15:12:38.823469 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="d89e1065-7553-4d5b-aef6-38108a5cb8ae" containerName="util" Nov 25 15:12:38 crc kubenswrapper[4890]: E1125 15:12:38.823480 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d89e1065-7553-4d5b-aef6-38108a5cb8ae" containerName="extract" Nov 25 15:12:38 crc kubenswrapper[4890]: I1125 15:12:38.823488 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="d89e1065-7553-4d5b-aef6-38108a5cb8ae" containerName="extract" Nov 25 15:12:38 crc kubenswrapper[4890]: I1125 15:12:38.823603 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="d89e1065-7553-4d5b-aef6-38108a5cb8ae" containerName="extract" Nov 25 15:12:38 crc kubenswrapper[4890]: I1125 15:12:38.824083 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-v4blp" Nov 25 15:12:38 crc kubenswrapper[4890]: I1125 15:12:38.826403 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-snhn6" Nov 25 15:12:38 crc kubenswrapper[4890]: I1125 15:12:38.826461 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Nov 25 15:12:38 crc kubenswrapper[4890]: I1125 15:12:38.826597 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Nov 25 15:12:38 crc kubenswrapper[4890]: I1125 15:12:38.839721 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-v4blp"] Nov 25 15:12:38 crc kubenswrapper[4890]: I1125 15:12:38.963606 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpjb7\" (UniqueName: \"kubernetes.io/projected/213afbd7-4c85-41e6-94c9-dac6db9da0d0-kube-api-access-xpjb7\") pod \"nmstate-operator-557fdffb88-v4blp\" (UID: \"213afbd7-4c85-41e6-94c9-dac6db9da0d0\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-v4blp" Nov 25 15:12:39 crc kubenswrapper[4890]: I1125 15:12:39.065448 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpjb7\" (UniqueName: \"kubernetes.io/projected/213afbd7-4c85-41e6-94c9-dac6db9da0d0-kube-api-access-xpjb7\") pod \"nmstate-operator-557fdffb88-v4blp\" (UID: \"213afbd7-4c85-41e6-94c9-dac6db9da0d0\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-v4blp" Nov 25 15:12:39 crc kubenswrapper[4890]: I1125 15:12:39.091886 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpjb7\" (UniqueName: \"kubernetes.io/projected/213afbd7-4c85-41e6-94c9-dac6db9da0d0-kube-api-access-xpjb7\") pod \"nmstate-operator-557fdffb88-v4blp\" (UID: \"213afbd7-4c85-41e6-94c9-dac6db9da0d0\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-v4blp" Nov 25 15:12:39 crc kubenswrapper[4890]: I1125 15:12:39.143214 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-v4blp" Nov 25 15:12:39 crc kubenswrapper[4890]: I1125 15:12:39.373803 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-v4blp"] Nov 25 15:12:40 crc kubenswrapper[4890]: I1125 15:12:40.104676 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-v4blp" event={"ID":"213afbd7-4c85-41e6-94c9-dac6db9da0d0","Type":"ContainerStarted","Data":"3be06a3519932a213d029d636db88049f3dba3f9c12f9b2d00888e65f48da8e1"} Nov 25 15:12:42 crc kubenswrapper[4890]: I1125 15:12:42.125415 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-v4blp" event={"ID":"213afbd7-4c85-41e6-94c9-dac6db9da0d0","Type":"ContainerStarted","Data":"8a45b04472eeb33b7a08db641720d5afd51cdbe7b62ce6682332865ffbb30609"} Nov 25 15:12:42 crc kubenswrapper[4890]: I1125 15:12:42.160783 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-557fdffb88-v4blp" podStartSLOduration=1.823345637 podStartE2EDuration="4.160754038s" podCreationTimestamp="2025-11-25 15:12:38 +0000 UTC" firstStartedPulling="2025-11-25 15:12:39.402286801 +0000 UTC m=+617.844749411" lastFinishedPulling="2025-11-25 15:12:41.739695162 +0000 UTC m=+620.182157812" observedRunningTime="2025-11-25 15:12:42.160737698 +0000 UTC m=+620.603200308" watchObservedRunningTime="2025-11-25 15:12:42.160754038 +0000 UTC m=+620.603216688" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.183033 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-8mxxw"] Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.184360 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-8mxxw" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.190399 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-kr52s" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.191816 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-nhh4b"] Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.192643 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-nhh4b" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.194090 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.205034 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-8mxxw"] Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.213406 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-sk8fz"] Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.214703 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-sk8fz" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.217312 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-nhh4b"] Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.313592 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-mrnt9"] Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.314285 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-mrnt9" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.316337 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-ctl55" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.316795 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.316842 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.331254 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-mrnt9"] Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.337807 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/1e35f1ea-5729-463f-b875-6b897b4d5944-ovs-socket\") pod \"nmstate-handler-sk8fz\" (UID: \"1e35f1ea-5729-463f-b875-6b897b4d5944\") " pod="openshift-nmstate/nmstate-handler-sk8fz" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.337844 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kv6v\" (UniqueName: \"kubernetes.io/projected/1e35f1ea-5729-463f-b875-6b897b4d5944-kube-api-access-4kv6v\") pod \"nmstate-handler-sk8fz\" (UID: \"1e35f1ea-5729-463f-b875-6b897b4d5944\") " pod="openshift-nmstate/nmstate-handler-sk8fz" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.337873 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/1e35f1ea-5729-463f-b875-6b897b4d5944-nmstate-lock\") pod \"nmstate-handler-sk8fz\" (UID: \"1e35f1ea-5729-463f-b875-6b897b4d5944\") " pod="openshift-nmstate/nmstate-handler-sk8fz" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.337972 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjcxl\" (UniqueName: \"kubernetes.io/projected/f6edfb37-872a-4515-ada7-07d552977de0-kube-api-access-kjcxl\") pod \"nmstate-webhook-6b89b748d8-nhh4b\" (UID: \"f6edfb37-872a-4515-ada7-07d552977de0\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-nhh4b" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.337988 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/1e35f1ea-5729-463f-b875-6b897b4d5944-dbus-socket\") pod \"nmstate-handler-sk8fz\" (UID: \"1e35f1ea-5729-463f-b875-6b897b4d5944\") " pod="openshift-nmstate/nmstate-handler-sk8fz" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.338029 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dp5q\" (UniqueName: \"kubernetes.io/projected/9d98268c-c10a-4157-9737-e2df6cbdcfe5-kube-api-access-2dp5q\") pod \"nmstate-metrics-5dcf9c57c5-8mxxw\" (UID: \"9d98268c-c10a-4157-9737-e2df6cbdcfe5\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-8mxxw" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.338048 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/f6edfb37-872a-4515-ada7-07d552977de0-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-nhh4b\" (UID: \"f6edfb37-872a-4515-ada7-07d552977de0\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-nhh4b" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.439572 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/1e35f1ea-5729-463f-b875-6b897b4d5944-ovs-socket\") pod \"nmstate-handler-sk8fz\" (UID: \"1e35f1ea-5729-463f-b875-6b897b4d5944\") " pod="openshift-nmstate/nmstate-handler-sk8fz" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.439628 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kv6v\" (UniqueName: \"kubernetes.io/projected/1e35f1ea-5729-463f-b875-6b897b4d5944-kube-api-access-4kv6v\") pod \"nmstate-handler-sk8fz\" (UID: \"1e35f1ea-5729-463f-b875-6b897b4d5944\") " pod="openshift-nmstate/nmstate-handler-sk8fz" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.439675 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/85206fda-8548-4b94-82ce-df958e8d4045-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-mrnt9\" (UID: \"85206fda-8548-4b94-82ce-df958e8d4045\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-mrnt9" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.439704 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/1e35f1ea-5729-463f-b875-6b897b4d5944-ovs-socket\") pod \"nmstate-handler-sk8fz\" (UID: \"1e35f1ea-5729-463f-b875-6b897b4d5944\") " pod="openshift-nmstate/nmstate-handler-sk8fz" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.439715 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/1e35f1ea-5729-463f-b875-6b897b4d5944-nmstate-lock\") pod \"nmstate-handler-sk8fz\" (UID: \"1e35f1ea-5729-463f-b875-6b897b4d5944\") " pod="openshift-nmstate/nmstate-handler-sk8fz" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.439774 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/1e35f1ea-5729-463f-b875-6b897b4d5944-nmstate-lock\") pod \"nmstate-handler-sk8fz\" (UID: \"1e35f1ea-5729-463f-b875-6b897b4d5944\") " pod="openshift-nmstate/nmstate-handler-sk8fz" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.439969 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/85206fda-8548-4b94-82ce-df958e8d4045-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-mrnt9\" (UID: \"85206fda-8548-4b94-82ce-df958e8d4045\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-mrnt9" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.440099 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9l9p\" (UniqueName: \"kubernetes.io/projected/85206fda-8548-4b94-82ce-df958e8d4045-kube-api-access-t9l9p\") pod \"nmstate-console-plugin-5874bd7bc5-mrnt9\" (UID: \"85206fda-8548-4b94-82ce-df958e8d4045\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-mrnt9" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.440175 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/1e35f1ea-5729-463f-b875-6b897b4d5944-dbus-socket\") pod \"nmstate-handler-sk8fz\" (UID: \"1e35f1ea-5729-463f-b875-6b897b4d5944\") " pod="openshift-nmstate/nmstate-handler-sk8fz" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.440206 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjcxl\" (UniqueName: \"kubernetes.io/projected/f6edfb37-872a-4515-ada7-07d552977de0-kube-api-access-kjcxl\") pod \"nmstate-webhook-6b89b748d8-nhh4b\" (UID: \"f6edfb37-872a-4515-ada7-07d552977de0\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-nhh4b" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.440524 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dp5q\" (UniqueName: \"kubernetes.io/projected/9d98268c-c10a-4157-9737-e2df6cbdcfe5-kube-api-access-2dp5q\") pod \"nmstate-metrics-5dcf9c57c5-8mxxw\" (UID: \"9d98268c-c10a-4157-9737-e2df6cbdcfe5\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-8mxxw" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.440574 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/f6edfb37-872a-4515-ada7-07d552977de0-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-nhh4b\" (UID: \"f6edfb37-872a-4515-ada7-07d552977de0\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-nhh4b" Nov 25 15:12:50 crc kubenswrapper[4890]: E1125 15:12:50.440730 4890 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Nov 25 15:12:50 crc kubenswrapper[4890]: E1125 15:12:50.440794 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f6edfb37-872a-4515-ada7-07d552977de0-tls-key-pair podName:f6edfb37-872a-4515-ada7-07d552977de0 nodeName:}" failed. No retries permitted until 2025-11-25 15:12:50.940769938 +0000 UTC m=+629.383232558 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/f6edfb37-872a-4515-ada7-07d552977de0-tls-key-pair") pod "nmstate-webhook-6b89b748d8-nhh4b" (UID: "f6edfb37-872a-4515-ada7-07d552977de0") : secret "openshift-nmstate-webhook" not found Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.441076 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/1e35f1ea-5729-463f-b875-6b897b4d5944-dbus-socket\") pod \"nmstate-handler-sk8fz\" (UID: \"1e35f1ea-5729-463f-b875-6b897b4d5944\") " pod="openshift-nmstate/nmstate-handler-sk8fz" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.478147 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kv6v\" (UniqueName: \"kubernetes.io/projected/1e35f1ea-5729-463f-b875-6b897b4d5944-kube-api-access-4kv6v\") pod \"nmstate-handler-sk8fz\" (UID: \"1e35f1ea-5729-463f-b875-6b897b4d5944\") " pod="openshift-nmstate/nmstate-handler-sk8fz" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.478970 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dp5q\" (UniqueName: \"kubernetes.io/projected/9d98268c-c10a-4157-9737-e2df6cbdcfe5-kube-api-access-2dp5q\") pod \"nmstate-metrics-5dcf9c57c5-8mxxw\" (UID: \"9d98268c-c10a-4157-9737-e2df6cbdcfe5\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-8mxxw" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.495016 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjcxl\" (UniqueName: \"kubernetes.io/projected/f6edfb37-872a-4515-ada7-07d552977de0-kube-api-access-kjcxl\") pod \"nmstate-webhook-6b89b748d8-nhh4b\" (UID: \"f6edfb37-872a-4515-ada7-07d552977de0\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-nhh4b" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.511816 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-8mxxw" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.531872 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-cf9bfd646-gbdr8"] Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.532547 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-cf9bfd646-gbdr8" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.539531 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-sk8fz" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.543713 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/85206fda-8548-4b94-82ce-df958e8d4045-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-mrnt9\" (UID: \"85206fda-8548-4b94-82ce-df958e8d4045\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-mrnt9" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.543801 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/85206fda-8548-4b94-82ce-df958e8d4045-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-mrnt9\" (UID: \"85206fda-8548-4b94-82ce-df958e8d4045\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-mrnt9" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.543840 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9l9p\" (UniqueName: \"kubernetes.io/projected/85206fda-8548-4b94-82ce-df958e8d4045-kube-api-access-t9l9p\") pod \"nmstate-console-plugin-5874bd7bc5-mrnt9\" (UID: \"85206fda-8548-4b94-82ce-df958e8d4045\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-mrnt9" Nov 25 15:12:50 crc kubenswrapper[4890]: E1125 15:12:50.544726 4890 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Nov 25 15:12:50 crc kubenswrapper[4890]: E1125 15:12:50.544803 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/85206fda-8548-4b94-82ce-df958e8d4045-plugin-serving-cert podName:85206fda-8548-4b94-82ce-df958e8d4045 nodeName:}" failed. No retries permitted until 2025-11-25 15:12:51.044782246 +0000 UTC m=+629.487244856 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/85206fda-8548-4b94-82ce-df958e8d4045-plugin-serving-cert") pod "nmstate-console-plugin-5874bd7bc5-mrnt9" (UID: "85206fda-8548-4b94-82ce-df958e8d4045") : secret "plugin-serving-cert" not found Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.545557 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/85206fda-8548-4b94-82ce-df958e8d4045-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-mrnt9\" (UID: \"85206fda-8548-4b94-82ce-df958e8d4045\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-mrnt9" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.551108 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-cf9bfd646-gbdr8"] Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.575897 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9l9p\" (UniqueName: \"kubernetes.io/projected/85206fda-8548-4b94-82ce-df958e8d4045-kube-api-access-t9l9p\") pod \"nmstate-console-plugin-5874bd7bc5-mrnt9\" (UID: \"85206fda-8548-4b94-82ce-df958e8d4045\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-mrnt9" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.645065 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0b71ef6e-ce5d-4f68-9df2-c09659013e8b-console-serving-cert\") pod \"console-cf9bfd646-gbdr8\" (UID: \"0b71ef6e-ce5d-4f68-9df2-c09659013e8b\") " pod="openshift-console/console-cf9bfd646-gbdr8" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.645148 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgsbm\" (UniqueName: \"kubernetes.io/projected/0b71ef6e-ce5d-4f68-9df2-c09659013e8b-kube-api-access-vgsbm\") pod \"console-cf9bfd646-gbdr8\" (UID: \"0b71ef6e-ce5d-4f68-9df2-c09659013e8b\") " pod="openshift-console/console-cf9bfd646-gbdr8" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.645241 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0b71ef6e-ce5d-4f68-9df2-c09659013e8b-console-config\") pod \"console-cf9bfd646-gbdr8\" (UID: \"0b71ef6e-ce5d-4f68-9df2-c09659013e8b\") " pod="openshift-console/console-cf9bfd646-gbdr8" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.645273 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0b71ef6e-ce5d-4f68-9df2-c09659013e8b-console-oauth-config\") pod \"console-cf9bfd646-gbdr8\" (UID: \"0b71ef6e-ce5d-4f68-9df2-c09659013e8b\") " pod="openshift-console/console-cf9bfd646-gbdr8" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.645304 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0b71ef6e-ce5d-4f68-9df2-c09659013e8b-trusted-ca-bundle\") pod \"console-cf9bfd646-gbdr8\" (UID: \"0b71ef6e-ce5d-4f68-9df2-c09659013e8b\") " pod="openshift-console/console-cf9bfd646-gbdr8" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.645379 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b71ef6e-ce5d-4f68-9df2-c09659013e8b-service-ca\") pod \"console-cf9bfd646-gbdr8\" (UID: \"0b71ef6e-ce5d-4f68-9df2-c09659013e8b\") " pod="openshift-console/console-cf9bfd646-gbdr8" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.645414 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0b71ef6e-ce5d-4f68-9df2-c09659013e8b-oauth-serving-cert\") pod \"console-cf9bfd646-gbdr8\" (UID: \"0b71ef6e-ce5d-4f68-9df2-c09659013e8b\") " pod="openshift-console/console-cf9bfd646-gbdr8" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.747002 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0b71ef6e-ce5d-4f68-9df2-c09659013e8b-console-config\") pod \"console-cf9bfd646-gbdr8\" (UID: \"0b71ef6e-ce5d-4f68-9df2-c09659013e8b\") " pod="openshift-console/console-cf9bfd646-gbdr8" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.747062 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0b71ef6e-ce5d-4f68-9df2-c09659013e8b-console-oauth-config\") pod \"console-cf9bfd646-gbdr8\" (UID: \"0b71ef6e-ce5d-4f68-9df2-c09659013e8b\") " pod="openshift-console/console-cf9bfd646-gbdr8" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.747079 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0b71ef6e-ce5d-4f68-9df2-c09659013e8b-trusted-ca-bundle\") pod \"console-cf9bfd646-gbdr8\" (UID: \"0b71ef6e-ce5d-4f68-9df2-c09659013e8b\") " pod="openshift-console/console-cf9bfd646-gbdr8" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.747131 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b71ef6e-ce5d-4f68-9df2-c09659013e8b-service-ca\") pod \"console-cf9bfd646-gbdr8\" (UID: \"0b71ef6e-ce5d-4f68-9df2-c09659013e8b\") " pod="openshift-console/console-cf9bfd646-gbdr8" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.747152 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0b71ef6e-ce5d-4f68-9df2-c09659013e8b-oauth-serving-cert\") pod \"console-cf9bfd646-gbdr8\" (UID: \"0b71ef6e-ce5d-4f68-9df2-c09659013e8b\") " pod="openshift-console/console-cf9bfd646-gbdr8" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.747205 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0b71ef6e-ce5d-4f68-9df2-c09659013e8b-console-serving-cert\") pod \"console-cf9bfd646-gbdr8\" (UID: \"0b71ef6e-ce5d-4f68-9df2-c09659013e8b\") " pod="openshift-console/console-cf9bfd646-gbdr8" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.747234 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgsbm\" (UniqueName: \"kubernetes.io/projected/0b71ef6e-ce5d-4f68-9df2-c09659013e8b-kube-api-access-vgsbm\") pod \"console-cf9bfd646-gbdr8\" (UID: \"0b71ef6e-ce5d-4f68-9df2-c09659013e8b\") " pod="openshift-console/console-cf9bfd646-gbdr8" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.748410 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b71ef6e-ce5d-4f68-9df2-c09659013e8b-service-ca\") pod \"console-cf9bfd646-gbdr8\" (UID: \"0b71ef6e-ce5d-4f68-9df2-c09659013e8b\") " pod="openshift-console/console-cf9bfd646-gbdr8" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.748453 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0b71ef6e-ce5d-4f68-9df2-c09659013e8b-console-config\") pod \"console-cf9bfd646-gbdr8\" (UID: \"0b71ef6e-ce5d-4f68-9df2-c09659013e8b\") " pod="openshift-console/console-cf9bfd646-gbdr8" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.748689 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0b71ef6e-ce5d-4f68-9df2-c09659013e8b-oauth-serving-cert\") pod \"console-cf9bfd646-gbdr8\" (UID: \"0b71ef6e-ce5d-4f68-9df2-c09659013e8b\") " pod="openshift-console/console-cf9bfd646-gbdr8" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.749046 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0b71ef6e-ce5d-4f68-9df2-c09659013e8b-trusted-ca-bundle\") pod \"console-cf9bfd646-gbdr8\" (UID: \"0b71ef6e-ce5d-4f68-9df2-c09659013e8b\") " pod="openshift-console/console-cf9bfd646-gbdr8" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.750791 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0b71ef6e-ce5d-4f68-9df2-c09659013e8b-console-serving-cert\") pod \"console-cf9bfd646-gbdr8\" (UID: \"0b71ef6e-ce5d-4f68-9df2-c09659013e8b\") " pod="openshift-console/console-cf9bfd646-gbdr8" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.751747 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0b71ef6e-ce5d-4f68-9df2-c09659013e8b-console-oauth-config\") pod \"console-cf9bfd646-gbdr8\" (UID: \"0b71ef6e-ce5d-4f68-9df2-c09659013e8b\") " pod="openshift-console/console-cf9bfd646-gbdr8" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.765269 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgsbm\" (UniqueName: \"kubernetes.io/projected/0b71ef6e-ce5d-4f68-9df2-c09659013e8b-kube-api-access-vgsbm\") pod \"console-cf9bfd646-gbdr8\" (UID: \"0b71ef6e-ce5d-4f68-9df2-c09659013e8b\") " pod="openshift-console/console-cf9bfd646-gbdr8" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.918333 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-cf9bfd646-gbdr8" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.920979 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-8mxxw"] Nov 25 15:12:50 crc kubenswrapper[4890]: W1125 15:12:50.934525 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d98268c_c10a_4157_9737_e2df6cbdcfe5.slice/crio-26a90970586bf0388a0209251f979d82e035f4d7694060066c1109be8118cc3f WatchSource:0}: Error finding container 26a90970586bf0388a0209251f979d82e035f4d7694060066c1109be8118cc3f: Status 404 returned error can't find the container with id 26a90970586bf0388a0209251f979d82e035f4d7694060066c1109be8118cc3f Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.949814 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/f6edfb37-872a-4515-ada7-07d552977de0-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-nhh4b\" (UID: \"f6edfb37-872a-4515-ada7-07d552977de0\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-nhh4b" Nov 25 15:12:50 crc kubenswrapper[4890]: I1125 15:12:50.953741 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/f6edfb37-872a-4515-ada7-07d552977de0-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-nhh4b\" (UID: \"f6edfb37-872a-4515-ada7-07d552977de0\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-nhh4b" Nov 25 15:12:51 crc kubenswrapper[4890]: I1125 15:12:51.051105 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/85206fda-8548-4b94-82ce-df958e8d4045-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-mrnt9\" (UID: \"85206fda-8548-4b94-82ce-df958e8d4045\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-mrnt9" Nov 25 15:12:51 crc kubenswrapper[4890]: I1125 15:12:51.056904 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/85206fda-8548-4b94-82ce-df958e8d4045-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-mrnt9\" (UID: \"85206fda-8548-4b94-82ce-df958e8d4045\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-mrnt9" Nov 25 15:12:51 crc kubenswrapper[4890]: I1125 15:12:51.090232 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-cf9bfd646-gbdr8"] Nov 25 15:12:51 crc kubenswrapper[4890]: W1125 15:12:51.096521 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0b71ef6e_ce5d_4f68_9df2_c09659013e8b.slice/crio-76f56166bf72add598063e484d64b8144fc7c705691cf2be9d46e3d5e0bc506b WatchSource:0}: Error finding container 76f56166bf72add598063e484d64b8144fc7c705691cf2be9d46e3d5e0bc506b: Status 404 returned error can't find the container with id 76f56166bf72add598063e484d64b8144fc7c705691cf2be9d46e3d5e0bc506b Nov 25 15:12:51 crc kubenswrapper[4890]: I1125 15:12:51.126841 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-nhh4b" Nov 25 15:12:51 crc kubenswrapper[4890]: I1125 15:12:51.180515 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-sk8fz" event={"ID":"1e35f1ea-5729-463f-b875-6b897b4d5944","Type":"ContainerStarted","Data":"bc36ef3dc8d69b0685f7d60b86acf87facc9a9caec01452b11309c6f8abee145"} Nov 25 15:12:51 crc kubenswrapper[4890]: I1125 15:12:51.181711 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-8mxxw" event={"ID":"9d98268c-c10a-4157-9737-e2df6cbdcfe5","Type":"ContainerStarted","Data":"26a90970586bf0388a0209251f979d82e035f4d7694060066c1109be8118cc3f"} Nov 25 15:12:51 crc kubenswrapper[4890]: I1125 15:12:51.182721 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-cf9bfd646-gbdr8" event={"ID":"0b71ef6e-ce5d-4f68-9df2-c09659013e8b","Type":"ContainerStarted","Data":"76f56166bf72add598063e484d64b8144fc7c705691cf2be9d46e3d5e0bc506b"} Nov 25 15:12:51 crc kubenswrapper[4890]: I1125 15:12:51.226823 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-mrnt9" Nov 25 15:12:51 crc kubenswrapper[4890]: I1125 15:12:51.458792 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-mrnt9"] Nov 25 15:12:51 crc kubenswrapper[4890]: I1125 15:12:51.543174 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-nhh4b"] Nov 25 15:12:51 crc kubenswrapper[4890]: W1125 15:12:51.549629 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf6edfb37_872a_4515_ada7_07d552977de0.slice/crio-4447b8433d332eebf7b509fb55f414991739a0e5bdf37fe7793da851e0ace17e WatchSource:0}: Error finding container 4447b8433d332eebf7b509fb55f414991739a0e5bdf37fe7793da851e0ace17e: Status 404 returned error can't find the container with id 4447b8433d332eebf7b509fb55f414991739a0e5bdf37fe7793da851e0ace17e Nov 25 15:12:52 crc kubenswrapper[4890]: I1125 15:12:52.193058 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-nhh4b" event={"ID":"f6edfb37-872a-4515-ada7-07d552977de0","Type":"ContainerStarted","Data":"4447b8433d332eebf7b509fb55f414991739a0e5bdf37fe7793da851e0ace17e"} Nov 25 15:12:52 crc kubenswrapper[4890]: I1125 15:12:52.194802 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-cf9bfd646-gbdr8" event={"ID":"0b71ef6e-ce5d-4f68-9df2-c09659013e8b","Type":"ContainerStarted","Data":"b91237e0f915377a0f72889c95257a32b2c290b6cf978ab95cbe759e6e3ab29b"} Nov 25 15:12:52 crc kubenswrapper[4890]: I1125 15:12:52.196015 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-mrnt9" event={"ID":"85206fda-8548-4b94-82ce-df958e8d4045","Type":"ContainerStarted","Data":"da94ebb01939a0c19eeab8a81fe306bc2d49ae06aed8d6f81c101f72d91f8614"} Nov 25 15:12:52 crc kubenswrapper[4890]: I1125 15:12:52.254989 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-cf9bfd646-gbdr8" podStartSLOduration=2.25496919 podStartE2EDuration="2.25496919s" podCreationTimestamp="2025-11-25 15:12:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:12:52.254832667 +0000 UTC m=+630.697295277" watchObservedRunningTime="2025-11-25 15:12:52.25496919 +0000 UTC m=+630.697431800" Nov 25 15:12:54 crc kubenswrapper[4890]: I1125 15:12:54.208047 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-sk8fz" event={"ID":"1e35f1ea-5729-463f-b875-6b897b4d5944","Type":"ContainerStarted","Data":"d476a16cbd135c746bdef27402d6b59475c940fce726990faa7e02be7dc2dc10"} Nov 25 15:12:54 crc kubenswrapper[4890]: I1125 15:12:54.208757 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-sk8fz" Nov 25 15:12:54 crc kubenswrapper[4890]: I1125 15:12:54.211151 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-8mxxw" event={"ID":"9d98268c-c10a-4157-9737-e2df6cbdcfe5","Type":"ContainerStarted","Data":"21e9aaf4fd45f86001a7f51a049340087d1388c33392f05062cc230f987953d4"} Nov 25 15:12:54 crc kubenswrapper[4890]: I1125 15:12:54.214135 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-mrnt9" event={"ID":"85206fda-8548-4b94-82ce-df958e8d4045","Type":"ContainerStarted","Data":"4f7d709315673393c815ea3d364937755dcaa12653c9ea88188208235367a79e"} Nov 25 15:12:54 crc kubenswrapper[4890]: I1125 15:12:54.216845 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-nhh4b" event={"ID":"f6edfb37-872a-4515-ada7-07d552977de0","Type":"ContainerStarted","Data":"4005f01d3272e50a571f0ef8b29a46b1aea73c532de54fe1b0cb6a6a0579368b"} Nov 25 15:12:54 crc kubenswrapper[4890]: I1125 15:12:54.217366 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-nhh4b" Nov 25 15:12:54 crc kubenswrapper[4890]: I1125 15:12:54.227817 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-sk8fz" podStartSLOduration=1.580449773 podStartE2EDuration="4.227791818s" podCreationTimestamp="2025-11-25 15:12:50 +0000 UTC" firstStartedPulling="2025-11-25 15:12:50.600277415 +0000 UTC m=+629.042740025" lastFinishedPulling="2025-11-25 15:12:53.24761946 +0000 UTC m=+631.690082070" observedRunningTime="2025-11-25 15:12:54.222340535 +0000 UTC m=+632.664803175" watchObservedRunningTime="2025-11-25 15:12:54.227791818 +0000 UTC m=+632.670254448" Nov 25 15:12:54 crc kubenswrapper[4890]: I1125 15:12:54.242054 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-nhh4b" podStartSLOduration=2.544913979 podStartE2EDuration="4.242034364s" podCreationTimestamp="2025-11-25 15:12:50 +0000 UTC" firstStartedPulling="2025-11-25 15:12:51.553863707 +0000 UTC m=+629.996326317" lastFinishedPulling="2025-11-25 15:12:53.250984102 +0000 UTC m=+631.693446702" observedRunningTime="2025-11-25 15:12:54.239387249 +0000 UTC m=+632.681849869" watchObservedRunningTime="2025-11-25 15:12:54.242034364 +0000 UTC m=+632.684496984" Nov 25 15:12:54 crc kubenswrapper[4890]: I1125 15:12:54.264522 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-mrnt9" podStartSLOduration=1.8773786609999998 podStartE2EDuration="4.26450028s" podCreationTimestamp="2025-11-25 15:12:50 +0000 UTC" firstStartedPulling="2025-11-25 15:12:51.470257074 +0000 UTC m=+629.912719684" lastFinishedPulling="2025-11-25 15:12:53.857378693 +0000 UTC m=+632.299841303" observedRunningTime="2025-11-25 15:12:54.261307002 +0000 UTC m=+632.703769632" watchObservedRunningTime="2025-11-25 15:12:54.26450028 +0000 UTC m=+632.706962890" Nov 25 15:12:57 crc kubenswrapper[4890]: I1125 15:12:57.239859 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-8mxxw" event={"ID":"9d98268c-c10a-4157-9737-e2df6cbdcfe5","Type":"ContainerStarted","Data":"0c6f080ef1e5d974ba687cdbf98ab4131490a34648663620b0b60739ba5da0d7"} Nov 25 15:12:57 crc kubenswrapper[4890]: I1125 15:12:57.261107 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-8mxxw" podStartSLOduration=1.338395089 podStartE2EDuration="7.261087805s" podCreationTimestamp="2025-11-25 15:12:50 +0000 UTC" firstStartedPulling="2025-11-25 15:12:50.937707618 +0000 UTC m=+629.380170218" lastFinishedPulling="2025-11-25 15:12:56.860400324 +0000 UTC m=+635.302862934" observedRunningTime="2025-11-25 15:12:57.255588351 +0000 UTC m=+635.698051011" watchObservedRunningTime="2025-11-25 15:12:57.261087805 +0000 UTC m=+635.703550425" Nov 25 15:13:00 crc kubenswrapper[4890]: I1125 15:13:00.574758 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-sk8fz" Nov 25 15:13:00 crc kubenswrapper[4890]: I1125 15:13:00.920205 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-cf9bfd646-gbdr8" Nov 25 15:13:00 crc kubenswrapper[4890]: I1125 15:13:00.920636 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-cf9bfd646-gbdr8" Nov 25 15:13:00 crc kubenswrapper[4890]: I1125 15:13:00.927234 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-cf9bfd646-gbdr8" Nov 25 15:13:01 crc kubenswrapper[4890]: I1125 15:13:01.272416 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-cf9bfd646-gbdr8" Nov 25 15:13:01 crc kubenswrapper[4890]: I1125 15:13:01.390368 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-t29fd"] Nov 25 15:13:11 crc kubenswrapper[4890]: I1125 15:13:11.134720 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-nhh4b" Nov 25 15:13:24 crc kubenswrapper[4890]: I1125 15:13:24.149778 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mhgqc"] Nov 25 15:13:24 crc kubenswrapper[4890]: I1125 15:13:24.151601 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mhgqc" Nov 25 15:13:24 crc kubenswrapper[4890]: I1125 15:13:24.156406 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mhgqc"] Nov 25 15:13:24 crc kubenswrapper[4890]: I1125 15:13:24.156925 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 25 15:13:24 crc kubenswrapper[4890]: I1125 15:13:24.231236 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/76fcec1c-1c27-4e7f-9f3e-a2527b72581f-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mhgqc\" (UID: \"76fcec1c-1c27-4e7f-9f3e-a2527b72581f\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mhgqc" Nov 25 15:13:24 crc kubenswrapper[4890]: I1125 15:13:24.231615 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/76fcec1c-1c27-4e7f-9f3e-a2527b72581f-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mhgqc\" (UID: \"76fcec1c-1c27-4e7f-9f3e-a2527b72581f\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mhgqc" Nov 25 15:13:24 crc kubenswrapper[4890]: I1125 15:13:24.231794 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qqtx\" (UniqueName: \"kubernetes.io/projected/76fcec1c-1c27-4e7f-9f3e-a2527b72581f-kube-api-access-8qqtx\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mhgqc\" (UID: \"76fcec1c-1c27-4e7f-9f3e-a2527b72581f\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mhgqc" Nov 25 15:13:24 crc kubenswrapper[4890]: I1125 15:13:24.333135 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/76fcec1c-1c27-4e7f-9f3e-a2527b72581f-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mhgqc\" (UID: \"76fcec1c-1c27-4e7f-9f3e-a2527b72581f\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mhgqc" Nov 25 15:13:24 crc kubenswrapper[4890]: I1125 15:13:24.333271 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/76fcec1c-1c27-4e7f-9f3e-a2527b72581f-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mhgqc\" (UID: \"76fcec1c-1c27-4e7f-9f3e-a2527b72581f\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mhgqc" Nov 25 15:13:24 crc kubenswrapper[4890]: I1125 15:13:24.333349 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qqtx\" (UniqueName: \"kubernetes.io/projected/76fcec1c-1c27-4e7f-9f3e-a2527b72581f-kube-api-access-8qqtx\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mhgqc\" (UID: \"76fcec1c-1c27-4e7f-9f3e-a2527b72581f\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mhgqc" Nov 25 15:13:24 crc kubenswrapper[4890]: I1125 15:13:24.333676 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/76fcec1c-1c27-4e7f-9f3e-a2527b72581f-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mhgqc\" (UID: \"76fcec1c-1c27-4e7f-9f3e-a2527b72581f\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mhgqc" Nov 25 15:13:24 crc kubenswrapper[4890]: I1125 15:13:24.334009 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/76fcec1c-1c27-4e7f-9f3e-a2527b72581f-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mhgqc\" (UID: \"76fcec1c-1c27-4e7f-9f3e-a2527b72581f\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mhgqc" Nov 25 15:13:24 crc kubenswrapper[4890]: I1125 15:13:24.355094 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qqtx\" (UniqueName: \"kubernetes.io/projected/76fcec1c-1c27-4e7f-9f3e-a2527b72581f-kube-api-access-8qqtx\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mhgqc\" (UID: \"76fcec1c-1c27-4e7f-9f3e-a2527b72581f\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mhgqc" Nov 25 15:13:24 crc kubenswrapper[4890]: I1125 15:13:24.473802 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mhgqc" Nov 25 15:13:24 crc kubenswrapper[4890]: I1125 15:13:24.890558 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mhgqc"] Nov 25 15:13:24 crc kubenswrapper[4890]: W1125 15:13:24.902799 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod76fcec1c_1c27_4e7f_9f3e_a2527b72581f.slice/crio-40a923d8706b6928b8f4f9bdbfed9138861913e2aa16c99f5e762e3a455b8edc WatchSource:0}: Error finding container 40a923d8706b6928b8f4f9bdbfed9138861913e2aa16c99f5e762e3a455b8edc: Status 404 returned error can't find the container with id 40a923d8706b6928b8f4f9bdbfed9138861913e2aa16c99f5e762e3a455b8edc Nov 25 15:13:25 crc kubenswrapper[4890]: I1125 15:13:25.426642 4890 generic.go:334] "Generic (PLEG): container finished" podID="76fcec1c-1c27-4e7f-9f3e-a2527b72581f" containerID="35a24e5519bcccaedc463b2e54945a8febf75185aa4a53ae31181420fddb62b2" exitCode=0 Nov 25 15:13:25 crc kubenswrapper[4890]: I1125 15:13:25.426762 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mhgqc" event={"ID":"76fcec1c-1c27-4e7f-9f3e-a2527b72581f","Type":"ContainerDied","Data":"35a24e5519bcccaedc463b2e54945a8febf75185aa4a53ae31181420fddb62b2"} Nov 25 15:13:25 crc kubenswrapper[4890]: I1125 15:13:25.427327 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mhgqc" event={"ID":"76fcec1c-1c27-4e7f-9f3e-a2527b72581f","Type":"ContainerStarted","Data":"40a923d8706b6928b8f4f9bdbfed9138861913e2aa16c99f5e762e3a455b8edc"} Nov 25 15:13:26 crc kubenswrapper[4890]: I1125 15:13:26.436055 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-t29fd" podUID="f748296c-d668-46bd-bcb7-4eb21822a244" containerName="console" containerID="cri-o://a5a26bf2f348abf2477c73bf6de69eb2f00258d478a7795da8b21e776c8eeae1" gracePeriod=15 Nov 25 15:13:27 crc kubenswrapper[4890]: I1125 15:13:27.185077 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-t29fd_f748296c-d668-46bd-bcb7-4eb21822a244/console/0.log" Nov 25 15:13:27 crc kubenswrapper[4890]: I1125 15:13:27.185621 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-t29fd" Nov 25 15:13:27 crc kubenswrapper[4890]: I1125 15:13:27.283482 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f748296c-d668-46bd-bcb7-4eb21822a244-oauth-serving-cert\") pod \"f748296c-d668-46bd-bcb7-4eb21822a244\" (UID: \"f748296c-d668-46bd-bcb7-4eb21822a244\") " Nov 25 15:13:27 crc kubenswrapper[4890]: I1125 15:13:27.283590 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f748296c-d668-46bd-bcb7-4eb21822a244-console-oauth-config\") pod \"f748296c-d668-46bd-bcb7-4eb21822a244\" (UID: \"f748296c-d668-46bd-bcb7-4eb21822a244\") " Nov 25 15:13:27 crc kubenswrapper[4890]: I1125 15:13:27.283673 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f748296c-d668-46bd-bcb7-4eb21822a244-console-serving-cert\") pod \"f748296c-d668-46bd-bcb7-4eb21822a244\" (UID: \"f748296c-d668-46bd-bcb7-4eb21822a244\") " Nov 25 15:13:27 crc kubenswrapper[4890]: I1125 15:13:27.283791 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f748296c-d668-46bd-bcb7-4eb21822a244-service-ca\") pod \"f748296c-d668-46bd-bcb7-4eb21822a244\" (UID: \"f748296c-d668-46bd-bcb7-4eb21822a244\") " Nov 25 15:13:27 crc kubenswrapper[4890]: I1125 15:13:27.283864 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f748296c-d668-46bd-bcb7-4eb21822a244-console-config\") pod \"f748296c-d668-46bd-bcb7-4eb21822a244\" (UID: \"f748296c-d668-46bd-bcb7-4eb21822a244\") " Nov 25 15:13:27 crc kubenswrapper[4890]: I1125 15:13:27.283898 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8kw4\" (UniqueName: \"kubernetes.io/projected/f748296c-d668-46bd-bcb7-4eb21822a244-kube-api-access-c8kw4\") pod \"f748296c-d668-46bd-bcb7-4eb21822a244\" (UID: \"f748296c-d668-46bd-bcb7-4eb21822a244\") " Nov 25 15:13:27 crc kubenswrapper[4890]: I1125 15:13:27.283935 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f748296c-d668-46bd-bcb7-4eb21822a244-trusted-ca-bundle\") pod \"f748296c-d668-46bd-bcb7-4eb21822a244\" (UID: \"f748296c-d668-46bd-bcb7-4eb21822a244\") " Nov 25 15:13:27 crc kubenswrapper[4890]: I1125 15:13:27.284645 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f748296c-d668-46bd-bcb7-4eb21822a244-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "f748296c-d668-46bd-bcb7-4eb21822a244" (UID: "f748296c-d668-46bd-bcb7-4eb21822a244"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:13:27 crc kubenswrapper[4890]: I1125 15:13:27.284712 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f748296c-d668-46bd-bcb7-4eb21822a244-console-config" (OuterVolumeSpecName: "console-config") pod "f748296c-d668-46bd-bcb7-4eb21822a244" (UID: "f748296c-d668-46bd-bcb7-4eb21822a244"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:13:27 crc kubenswrapper[4890]: I1125 15:13:27.285287 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f748296c-d668-46bd-bcb7-4eb21822a244-service-ca" (OuterVolumeSpecName: "service-ca") pod "f748296c-d668-46bd-bcb7-4eb21822a244" (UID: "f748296c-d668-46bd-bcb7-4eb21822a244"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:13:27 crc kubenswrapper[4890]: I1125 15:13:27.285627 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f748296c-d668-46bd-bcb7-4eb21822a244-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "f748296c-d668-46bd-bcb7-4eb21822a244" (UID: "f748296c-d668-46bd-bcb7-4eb21822a244"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:13:27 crc kubenswrapper[4890]: I1125 15:13:27.288610 4890 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f748296c-d668-46bd-bcb7-4eb21822a244-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 15:13:27 crc kubenswrapper[4890]: I1125 15:13:27.288660 4890 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f748296c-d668-46bd-bcb7-4eb21822a244-console-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:13:27 crc kubenswrapper[4890]: I1125 15:13:27.288680 4890 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f748296c-d668-46bd-bcb7-4eb21822a244-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:13:27 crc kubenswrapper[4890]: I1125 15:13:27.288701 4890 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f748296c-d668-46bd-bcb7-4eb21822a244-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 15:13:27 crc kubenswrapper[4890]: I1125 15:13:27.290461 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f748296c-d668-46bd-bcb7-4eb21822a244-kube-api-access-c8kw4" (OuterVolumeSpecName: "kube-api-access-c8kw4") pod "f748296c-d668-46bd-bcb7-4eb21822a244" (UID: "f748296c-d668-46bd-bcb7-4eb21822a244"). InnerVolumeSpecName "kube-api-access-c8kw4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:13:27 crc kubenswrapper[4890]: I1125 15:13:27.290544 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f748296c-d668-46bd-bcb7-4eb21822a244-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "f748296c-d668-46bd-bcb7-4eb21822a244" (UID: "f748296c-d668-46bd-bcb7-4eb21822a244"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:13:27 crc kubenswrapper[4890]: I1125 15:13:27.291911 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f748296c-d668-46bd-bcb7-4eb21822a244-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "f748296c-d668-46bd-bcb7-4eb21822a244" (UID: "f748296c-d668-46bd-bcb7-4eb21822a244"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:13:27 crc kubenswrapper[4890]: I1125 15:13:27.389916 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c8kw4\" (UniqueName: \"kubernetes.io/projected/f748296c-d668-46bd-bcb7-4eb21822a244-kube-api-access-c8kw4\") on node \"crc\" DevicePath \"\"" Nov 25 15:13:27 crc kubenswrapper[4890]: I1125 15:13:27.389961 4890 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f748296c-d668-46bd-bcb7-4eb21822a244-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:13:27 crc kubenswrapper[4890]: I1125 15:13:27.389973 4890 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f748296c-d668-46bd-bcb7-4eb21822a244-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 15:13:27 crc kubenswrapper[4890]: I1125 15:13:27.440895 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-t29fd_f748296c-d668-46bd-bcb7-4eb21822a244/console/0.log" Nov 25 15:13:27 crc kubenswrapper[4890]: I1125 15:13:27.440948 4890 generic.go:334] "Generic (PLEG): container finished" podID="f748296c-d668-46bd-bcb7-4eb21822a244" containerID="a5a26bf2f348abf2477c73bf6de69eb2f00258d478a7795da8b21e776c8eeae1" exitCode=2 Nov 25 15:13:27 crc kubenswrapper[4890]: I1125 15:13:27.441004 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-t29fd" event={"ID":"f748296c-d668-46bd-bcb7-4eb21822a244","Type":"ContainerDied","Data":"a5a26bf2f348abf2477c73bf6de69eb2f00258d478a7795da8b21e776c8eeae1"} Nov 25 15:13:27 crc kubenswrapper[4890]: I1125 15:13:27.441035 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-t29fd" event={"ID":"f748296c-d668-46bd-bcb7-4eb21822a244","Type":"ContainerDied","Data":"9d0c93c82198fc751619075a936ae9aa328dc51d85d284cdc40d56221bbf7362"} Nov 25 15:13:27 crc kubenswrapper[4890]: I1125 15:13:27.441053 4890 scope.go:117] "RemoveContainer" containerID="a5a26bf2f348abf2477c73bf6de69eb2f00258d478a7795da8b21e776c8eeae1" Nov 25 15:13:27 crc kubenswrapper[4890]: I1125 15:13:27.441131 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-t29fd" Nov 25 15:13:27 crc kubenswrapper[4890]: I1125 15:13:27.443470 4890 generic.go:334] "Generic (PLEG): container finished" podID="76fcec1c-1c27-4e7f-9f3e-a2527b72581f" containerID="0b86f48bde9e4f879586f3a00792ed7361822d09afc66929d6db0f424aafa28d" exitCode=0 Nov 25 15:13:27 crc kubenswrapper[4890]: I1125 15:13:27.443492 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mhgqc" event={"ID":"76fcec1c-1c27-4e7f-9f3e-a2527b72581f","Type":"ContainerDied","Data":"0b86f48bde9e4f879586f3a00792ed7361822d09afc66929d6db0f424aafa28d"} Nov 25 15:13:27 crc kubenswrapper[4890]: I1125 15:13:27.467483 4890 scope.go:117] "RemoveContainer" containerID="a5a26bf2f348abf2477c73bf6de69eb2f00258d478a7795da8b21e776c8eeae1" Nov 25 15:13:27 crc kubenswrapper[4890]: E1125 15:13:27.469323 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5a26bf2f348abf2477c73bf6de69eb2f00258d478a7795da8b21e776c8eeae1\": container with ID starting with a5a26bf2f348abf2477c73bf6de69eb2f00258d478a7795da8b21e776c8eeae1 not found: ID does not exist" containerID="a5a26bf2f348abf2477c73bf6de69eb2f00258d478a7795da8b21e776c8eeae1" Nov 25 15:13:27 crc kubenswrapper[4890]: I1125 15:13:27.469403 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5a26bf2f348abf2477c73bf6de69eb2f00258d478a7795da8b21e776c8eeae1"} err="failed to get container status \"a5a26bf2f348abf2477c73bf6de69eb2f00258d478a7795da8b21e776c8eeae1\": rpc error: code = NotFound desc = could not find container \"a5a26bf2f348abf2477c73bf6de69eb2f00258d478a7795da8b21e776c8eeae1\": container with ID starting with a5a26bf2f348abf2477c73bf6de69eb2f00258d478a7795da8b21e776c8eeae1 not found: ID does not exist" Nov 25 15:13:27 crc kubenswrapper[4890]: I1125 15:13:27.484666 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-t29fd"] Nov 25 15:13:27 crc kubenswrapper[4890]: I1125 15:13:27.492271 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-t29fd"] Nov 25 15:13:28 crc kubenswrapper[4890]: I1125 15:13:28.186404 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f748296c-d668-46bd-bcb7-4eb21822a244" path="/var/lib/kubelet/pods/f748296c-d668-46bd-bcb7-4eb21822a244/volumes" Nov 25 15:13:28 crc kubenswrapper[4890]: I1125 15:13:28.453769 4890 generic.go:334] "Generic (PLEG): container finished" podID="76fcec1c-1c27-4e7f-9f3e-a2527b72581f" containerID="6f7478573116cfac1fde79533d1e77056000904aa841f0d7ed8bdc1cdc9430c5" exitCode=0 Nov 25 15:13:28 crc kubenswrapper[4890]: I1125 15:13:28.453807 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mhgqc" event={"ID":"76fcec1c-1c27-4e7f-9f3e-a2527b72581f","Type":"ContainerDied","Data":"6f7478573116cfac1fde79533d1e77056000904aa841f0d7ed8bdc1cdc9430c5"} Nov 25 15:13:29 crc kubenswrapper[4890]: I1125 15:13:29.747476 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mhgqc" Nov 25 15:13:29 crc kubenswrapper[4890]: I1125 15:13:29.823032 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8qqtx\" (UniqueName: \"kubernetes.io/projected/76fcec1c-1c27-4e7f-9f3e-a2527b72581f-kube-api-access-8qqtx\") pod \"76fcec1c-1c27-4e7f-9f3e-a2527b72581f\" (UID: \"76fcec1c-1c27-4e7f-9f3e-a2527b72581f\") " Nov 25 15:13:29 crc kubenswrapper[4890]: I1125 15:13:29.823126 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/76fcec1c-1c27-4e7f-9f3e-a2527b72581f-bundle\") pod \"76fcec1c-1c27-4e7f-9f3e-a2527b72581f\" (UID: \"76fcec1c-1c27-4e7f-9f3e-a2527b72581f\") " Nov 25 15:13:29 crc kubenswrapper[4890]: I1125 15:13:29.823191 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/76fcec1c-1c27-4e7f-9f3e-a2527b72581f-util\") pod \"76fcec1c-1c27-4e7f-9f3e-a2527b72581f\" (UID: \"76fcec1c-1c27-4e7f-9f3e-a2527b72581f\") " Nov 25 15:13:29 crc kubenswrapper[4890]: I1125 15:13:29.824437 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76fcec1c-1c27-4e7f-9f3e-a2527b72581f-bundle" (OuterVolumeSpecName: "bundle") pod "76fcec1c-1c27-4e7f-9f3e-a2527b72581f" (UID: "76fcec1c-1c27-4e7f-9f3e-a2527b72581f"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:13:29 crc kubenswrapper[4890]: I1125 15:13:29.832421 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76fcec1c-1c27-4e7f-9f3e-a2527b72581f-kube-api-access-8qqtx" (OuterVolumeSpecName: "kube-api-access-8qqtx") pod "76fcec1c-1c27-4e7f-9f3e-a2527b72581f" (UID: "76fcec1c-1c27-4e7f-9f3e-a2527b72581f"). InnerVolumeSpecName "kube-api-access-8qqtx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:13:29 crc kubenswrapper[4890]: I1125 15:13:29.838886 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76fcec1c-1c27-4e7f-9f3e-a2527b72581f-util" (OuterVolumeSpecName: "util") pod "76fcec1c-1c27-4e7f-9f3e-a2527b72581f" (UID: "76fcec1c-1c27-4e7f-9f3e-a2527b72581f"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:13:29 crc kubenswrapper[4890]: I1125 15:13:29.924609 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8qqtx\" (UniqueName: \"kubernetes.io/projected/76fcec1c-1c27-4e7f-9f3e-a2527b72581f-kube-api-access-8qqtx\") on node \"crc\" DevicePath \"\"" Nov 25 15:13:29 crc kubenswrapper[4890]: I1125 15:13:29.924637 4890 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/76fcec1c-1c27-4e7f-9f3e-a2527b72581f-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:13:29 crc kubenswrapper[4890]: I1125 15:13:29.924645 4890 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/76fcec1c-1c27-4e7f-9f3e-a2527b72581f-util\") on node \"crc\" DevicePath \"\"" Nov 25 15:13:30 crc kubenswrapper[4890]: I1125 15:13:30.471284 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mhgqc" event={"ID":"76fcec1c-1c27-4e7f-9f3e-a2527b72581f","Type":"ContainerDied","Data":"40a923d8706b6928b8f4f9bdbfed9138861913e2aa16c99f5e762e3a455b8edc"} Nov 25 15:13:30 crc kubenswrapper[4890]: I1125 15:13:30.471588 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="40a923d8706b6928b8f4f9bdbfed9138861913e2aa16c99f5e762e3a455b8edc" Nov 25 15:13:30 crc kubenswrapper[4890]: I1125 15:13:30.471431 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mhgqc" Nov 25 15:13:39 crc kubenswrapper[4890]: I1125 15:13:39.613552 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-66fc8c7fbf-lnjxm"] Nov 25 15:13:39 crc kubenswrapper[4890]: E1125 15:13:39.614738 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f748296c-d668-46bd-bcb7-4eb21822a244" containerName="console" Nov 25 15:13:39 crc kubenswrapper[4890]: I1125 15:13:39.614754 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="f748296c-d668-46bd-bcb7-4eb21822a244" containerName="console" Nov 25 15:13:39 crc kubenswrapper[4890]: E1125 15:13:39.614770 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76fcec1c-1c27-4e7f-9f3e-a2527b72581f" containerName="pull" Nov 25 15:13:39 crc kubenswrapper[4890]: I1125 15:13:39.614777 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="76fcec1c-1c27-4e7f-9f3e-a2527b72581f" containerName="pull" Nov 25 15:13:39 crc kubenswrapper[4890]: E1125 15:13:39.614793 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76fcec1c-1c27-4e7f-9f3e-a2527b72581f" containerName="util" Nov 25 15:13:39 crc kubenswrapper[4890]: I1125 15:13:39.614800 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="76fcec1c-1c27-4e7f-9f3e-a2527b72581f" containerName="util" Nov 25 15:13:39 crc kubenswrapper[4890]: E1125 15:13:39.614815 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76fcec1c-1c27-4e7f-9f3e-a2527b72581f" containerName="extract" Nov 25 15:13:39 crc kubenswrapper[4890]: I1125 15:13:39.614822 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="76fcec1c-1c27-4e7f-9f3e-a2527b72581f" containerName="extract" Nov 25 15:13:39 crc kubenswrapper[4890]: I1125 15:13:39.614948 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="76fcec1c-1c27-4e7f-9f3e-a2527b72581f" containerName="extract" Nov 25 15:13:39 crc kubenswrapper[4890]: I1125 15:13:39.614963 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="f748296c-d668-46bd-bcb7-4eb21822a244" containerName="console" Nov 25 15:13:39 crc kubenswrapper[4890]: I1125 15:13:39.615517 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-66fc8c7fbf-lnjxm" Nov 25 15:13:39 crc kubenswrapper[4890]: I1125 15:13:39.619836 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 25 15:13:39 crc kubenswrapper[4890]: I1125 15:13:39.621275 4890 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-4kv2m" Nov 25 15:13:39 crc kubenswrapper[4890]: I1125 15:13:39.621486 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 25 15:13:39 crc kubenswrapper[4890]: I1125 15:13:39.623557 4890 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 25 15:13:39 crc kubenswrapper[4890]: I1125 15:13:39.624081 4890 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Nov 25 15:13:39 crc kubenswrapper[4890]: I1125 15:13:39.637946 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-66fc8c7fbf-lnjxm"] Nov 25 15:13:39 crc kubenswrapper[4890]: I1125 15:13:39.771977 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkds7\" (UniqueName: \"kubernetes.io/projected/dad943fe-2209-40f9-a9bb-0be7ba80cca2-kube-api-access-kkds7\") pod \"metallb-operator-controller-manager-66fc8c7fbf-lnjxm\" (UID: \"dad943fe-2209-40f9-a9bb-0be7ba80cca2\") " pod="metallb-system/metallb-operator-controller-manager-66fc8c7fbf-lnjxm" Nov 25 15:13:39 crc kubenswrapper[4890]: I1125 15:13:39.772033 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/dad943fe-2209-40f9-a9bb-0be7ba80cca2-webhook-cert\") pod \"metallb-operator-controller-manager-66fc8c7fbf-lnjxm\" (UID: \"dad943fe-2209-40f9-a9bb-0be7ba80cca2\") " pod="metallb-system/metallb-operator-controller-manager-66fc8c7fbf-lnjxm" Nov 25 15:13:39 crc kubenswrapper[4890]: I1125 15:13:39.772068 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/dad943fe-2209-40f9-a9bb-0be7ba80cca2-apiservice-cert\") pod \"metallb-operator-controller-manager-66fc8c7fbf-lnjxm\" (UID: \"dad943fe-2209-40f9-a9bb-0be7ba80cca2\") " pod="metallb-system/metallb-operator-controller-manager-66fc8c7fbf-lnjxm" Nov 25 15:13:39 crc kubenswrapper[4890]: I1125 15:13:39.873396 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkds7\" (UniqueName: \"kubernetes.io/projected/dad943fe-2209-40f9-a9bb-0be7ba80cca2-kube-api-access-kkds7\") pod \"metallb-operator-controller-manager-66fc8c7fbf-lnjxm\" (UID: \"dad943fe-2209-40f9-a9bb-0be7ba80cca2\") " pod="metallb-system/metallb-operator-controller-manager-66fc8c7fbf-lnjxm" Nov 25 15:13:39 crc kubenswrapper[4890]: I1125 15:13:39.873478 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/dad943fe-2209-40f9-a9bb-0be7ba80cca2-webhook-cert\") pod \"metallb-operator-controller-manager-66fc8c7fbf-lnjxm\" (UID: \"dad943fe-2209-40f9-a9bb-0be7ba80cca2\") " pod="metallb-system/metallb-operator-controller-manager-66fc8c7fbf-lnjxm" Nov 25 15:13:39 crc kubenswrapper[4890]: I1125 15:13:39.873518 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/dad943fe-2209-40f9-a9bb-0be7ba80cca2-apiservice-cert\") pod \"metallb-operator-controller-manager-66fc8c7fbf-lnjxm\" (UID: \"dad943fe-2209-40f9-a9bb-0be7ba80cca2\") " pod="metallb-system/metallb-operator-controller-manager-66fc8c7fbf-lnjxm" Nov 25 15:13:39 crc kubenswrapper[4890]: I1125 15:13:39.881054 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/dad943fe-2209-40f9-a9bb-0be7ba80cca2-apiservice-cert\") pod \"metallb-operator-controller-manager-66fc8c7fbf-lnjxm\" (UID: \"dad943fe-2209-40f9-a9bb-0be7ba80cca2\") " pod="metallb-system/metallb-operator-controller-manager-66fc8c7fbf-lnjxm" Nov 25 15:13:39 crc kubenswrapper[4890]: I1125 15:13:39.882034 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/dad943fe-2209-40f9-a9bb-0be7ba80cca2-webhook-cert\") pod \"metallb-operator-controller-manager-66fc8c7fbf-lnjxm\" (UID: \"dad943fe-2209-40f9-a9bb-0be7ba80cca2\") " pod="metallb-system/metallb-operator-controller-manager-66fc8c7fbf-lnjxm" Nov 25 15:13:39 crc kubenswrapper[4890]: I1125 15:13:39.903684 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkds7\" (UniqueName: \"kubernetes.io/projected/dad943fe-2209-40f9-a9bb-0be7ba80cca2-kube-api-access-kkds7\") pod \"metallb-operator-controller-manager-66fc8c7fbf-lnjxm\" (UID: \"dad943fe-2209-40f9-a9bb-0be7ba80cca2\") " pod="metallb-system/metallb-operator-controller-manager-66fc8c7fbf-lnjxm" Nov 25 15:13:39 crc kubenswrapper[4890]: I1125 15:13:39.935367 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-66fc8c7fbf-lnjxm" Nov 25 15:13:39 crc kubenswrapper[4890]: I1125 15:13:39.952119 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-6b856c5bd9-slq4g"] Nov 25 15:13:39 crc kubenswrapper[4890]: I1125 15:13:39.959037 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6b856c5bd9-slq4g" Nov 25 15:13:39 crc kubenswrapper[4890]: I1125 15:13:39.964314 4890 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-v4zm5" Nov 25 15:13:39 crc kubenswrapper[4890]: I1125 15:13:39.964507 4890 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 25 15:13:39 crc kubenswrapper[4890]: I1125 15:13:39.964648 4890 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 25 15:13:39 crc kubenswrapper[4890]: I1125 15:13:39.996084 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6b856c5bd9-slq4g"] Nov 25 15:13:40 crc kubenswrapper[4890]: I1125 15:13:40.077738 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9k4bc\" (UniqueName: \"kubernetes.io/projected/fb5c6944-bf04-42b5-98f6-550b9b2811d5-kube-api-access-9k4bc\") pod \"metallb-operator-webhook-server-6b856c5bd9-slq4g\" (UID: \"fb5c6944-bf04-42b5-98f6-550b9b2811d5\") " pod="metallb-system/metallb-operator-webhook-server-6b856c5bd9-slq4g" Nov 25 15:13:40 crc kubenswrapper[4890]: I1125 15:13:40.077805 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fb5c6944-bf04-42b5-98f6-550b9b2811d5-apiservice-cert\") pod \"metallb-operator-webhook-server-6b856c5bd9-slq4g\" (UID: \"fb5c6944-bf04-42b5-98f6-550b9b2811d5\") " pod="metallb-system/metallb-operator-webhook-server-6b856c5bd9-slq4g" Nov 25 15:13:40 crc kubenswrapper[4890]: I1125 15:13:40.077838 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fb5c6944-bf04-42b5-98f6-550b9b2811d5-webhook-cert\") pod \"metallb-operator-webhook-server-6b856c5bd9-slq4g\" (UID: \"fb5c6944-bf04-42b5-98f6-550b9b2811d5\") " pod="metallb-system/metallb-operator-webhook-server-6b856c5bd9-slq4g" Nov 25 15:13:40 crc kubenswrapper[4890]: I1125 15:13:40.179419 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fb5c6944-bf04-42b5-98f6-550b9b2811d5-webhook-cert\") pod \"metallb-operator-webhook-server-6b856c5bd9-slq4g\" (UID: \"fb5c6944-bf04-42b5-98f6-550b9b2811d5\") " pod="metallb-system/metallb-operator-webhook-server-6b856c5bd9-slq4g" Nov 25 15:13:40 crc kubenswrapper[4890]: I1125 15:13:40.179577 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9k4bc\" (UniqueName: \"kubernetes.io/projected/fb5c6944-bf04-42b5-98f6-550b9b2811d5-kube-api-access-9k4bc\") pod \"metallb-operator-webhook-server-6b856c5bd9-slq4g\" (UID: \"fb5c6944-bf04-42b5-98f6-550b9b2811d5\") " pod="metallb-system/metallb-operator-webhook-server-6b856c5bd9-slq4g" Nov 25 15:13:40 crc kubenswrapper[4890]: I1125 15:13:40.179642 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fb5c6944-bf04-42b5-98f6-550b9b2811d5-apiservice-cert\") pod \"metallb-operator-webhook-server-6b856c5bd9-slq4g\" (UID: \"fb5c6944-bf04-42b5-98f6-550b9b2811d5\") " pod="metallb-system/metallb-operator-webhook-server-6b856c5bd9-slq4g" Nov 25 15:13:40 crc kubenswrapper[4890]: I1125 15:13:40.186665 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fb5c6944-bf04-42b5-98f6-550b9b2811d5-webhook-cert\") pod \"metallb-operator-webhook-server-6b856c5bd9-slq4g\" (UID: \"fb5c6944-bf04-42b5-98f6-550b9b2811d5\") " pod="metallb-system/metallb-operator-webhook-server-6b856c5bd9-slq4g" Nov 25 15:13:40 crc kubenswrapper[4890]: I1125 15:13:40.191499 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fb5c6944-bf04-42b5-98f6-550b9b2811d5-apiservice-cert\") pod \"metallb-operator-webhook-server-6b856c5bd9-slq4g\" (UID: \"fb5c6944-bf04-42b5-98f6-550b9b2811d5\") " pod="metallb-system/metallb-operator-webhook-server-6b856c5bd9-slq4g" Nov 25 15:13:40 crc kubenswrapper[4890]: I1125 15:13:40.203118 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9k4bc\" (UniqueName: \"kubernetes.io/projected/fb5c6944-bf04-42b5-98f6-550b9b2811d5-kube-api-access-9k4bc\") pod \"metallb-operator-webhook-server-6b856c5bd9-slq4g\" (UID: \"fb5c6944-bf04-42b5-98f6-550b9b2811d5\") " pod="metallb-system/metallb-operator-webhook-server-6b856c5bd9-slq4g" Nov 25 15:13:40 crc kubenswrapper[4890]: I1125 15:13:40.319356 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6b856c5bd9-slq4g" Nov 25 15:13:40 crc kubenswrapper[4890]: I1125 15:13:40.383546 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-66fc8c7fbf-lnjxm"] Nov 25 15:13:40 crc kubenswrapper[4890]: I1125 15:13:40.533347 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-66fc8c7fbf-lnjxm" event={"ID":"dad943fe-2209-40f9-a9bb-0be7ba80cca2","Type":"ContainerStarted","Data":"390eb14c7c98527437923e28b628bd865a643699dc58cbca9f9b733cd54bc720"} Nov 25 15:13:40 crc kubenswrapper[4890]: I1125 15:13:40.555216 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6b856c5bd9-slq4g"] Nov 25 15:13:40 crc kubenswrapper[4890]: W1125 15:13:40.564173 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfb5c6944_bf04_42b5_98f6_550b9b2811d5.slice/crio-0c6e554e27ec23c720b0dd16980b77a3566bc512dab4ec335a199fd22c72260d WatchSource:0}: Error finding container 0c6e554e27ec23c720b0dd16980b77a3566bc512dab4ec335a199fd22c72260d: Status 404 returned error can't find the container with id 0c6e554e27ec23c720b0dd16980b77a3566bc512dab4ec335a199fd22c72260d Nov 25 15:13:41 crc kubenswrapper[4890]: I1125 15:13:41.539681 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6b856c5bd9-slq4g" event={"ID":"fb5c6944-bf04-42b5-98f6-550b9b2811d5","Type":"ContainerStarted","Data":"0c6e554e27ec23c720b0dd16980b77a3566bc512dab4ec335a199fd22c72260d"} Nov 25 15:13:47 crc kubenswrapper[4890]: I1125 15:13:47.590412 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6b856c5bd9-slq4g" event={"ID":"fb5c6944-bf04-42b5-98f6-550b9b2811d5","Type":"ContainerStarted","Data":"9f8304a7ad902f77a80adb863215b95ba6bc04ea573dd4098f845528e307a602"} Nov 25 15:13:47 crc kubenswrapper[4890]: I1125 15:13:47.591357 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-6b856c5bd9-slq4g" Nov 25 15:13:47 crc kubenswrapper[4890]: I1125 15:13:47.592116 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-66fc8c7fbf-lnjxm" event={"ID":"dad943fe-2209-40f9-a9bb-0be7ba80cca2","Type":"ContainerStarted","Data":"0fc718ab7f1457fe75eff145c26dc4f53ab7a75dd29150bb92b277bdff7d608f"} Nov 25 15:13:47 crc kubenswrapper[4890]: I1125 15:13:47.592340 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-66fc8c7fbf-lnjxm" Nov 25 15:13:47 crc kubenswrapper[4890]: I1125 15:13:47.652780 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-6b856c5bd9-slq4g" podStartSLOduration=2.699742505 podStartE2EDuration="8.652760415s" podCreationTimestamp="2025-11-25 15:13:39 +0000 UTC" firstStartedPulling="2025-11-25 15:13:40.566948152 +0000 UTC m=+679.009410762" lastFinishedPulling="2025-11-25 15:13:46.519966062 +0000 UTC m=+684.962428672" observedRunningTime="2025-11-25 15:13:47.648052468 +0000 UTC m=+686.090515078" watchObservedRunningTime="2025-11-25 15:13:47.652760415 +0000 UTC m=+686.095223035" Nov 25 15:13:47 crc kubenswrapper[4890]: I1125 15:13:47.695961 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-66fc8c7fbf-lnjxm" podStartSLOduration=2.593048154 podStartE2EDuration="8.695944332s" podCreationTimestamp="2025-11-25 15:13:39 +0000 UTC" firstStartedPulling="2025-11-25 15:13:40.394431029 +0000 UTC m=+678.836893639" lastFinishedPulling="2025-11-25 15:13:46.497327187 +0000 UTC m=+684.939789817" observedRunningTime="2025-11-25 15:13:47.689372668 +0000 UTC m=+686.131835278" watchObservedRunningTime="2025-11-25 15:13:47.695944332 +0000 UTC m=+686.138406942" Nov 25 15:13:56 crc kubenswrapper[4890]: I1125 15:13:56.305982 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xldhl"] Nov 25 15:13:56 crc kubenswrapper[4890]: I1125 15:13:56.306947 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-xldhl" podUID="6e313ec3-70c2-4220-b08a-ef79b8a25847" containerName="controller-manager" containerID="cri-o://9fbede642703b66da34b00feaaa8a19598b017d4abe7de06ff6b12a64a93636f" gracePeriod=30 Nov 25 15:13:56 crc kubenswrapper[4890]: I1125 15:13:56.354730 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-mqvqx"] Nov 25 15:13:56 crc kubenswrapper[4890]: I1125 15:13:56.355432 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mqvqx" podUID="e09b8f75-2286-4a12-9571-5342b681690f" containerName="route-controller-manager" containerID="cri-o://54c69ecc23762a33f8d6eeca868dd59ac28361f509178948acf91cd5a6241e65" gracePeriod=30 Nov 25 15:13:56 crc kubenswrapper[4890]: I1125 15:13:56.447904 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 15:13:56 crc kubenswrapper[4890]: I1125 15:13:56.447981 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 15:13:56 crc kubenswrapper[4890]: I1125 15:13:56.660723 4890 generic.go:334] "Generic (PLEG): container finished" podID="e09b8f75-2286-4a12-9571-5342b681690f" containerID="54c69ecc23762a33f8d6eeca868dd59ac28361f509178948acf91cd5a6241e65" exitCode=0 Nov 25 15:13:56 crc kubenswrapper[4890]: I1125 15:13:56.660812 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mqvqx" event={"ID":"e09b8f75-2286-4a12-9571-5342b681690f","Type":"ContainerDied","Data":"54c69ecc23762a33f8d6eeca868dd59ac28361f509178948acf91cd5a6241e65"} Nov 25 15:13:56 crc kubenswrapper[4890]: I1125 15:13:56.662947 4890 generic.go:334] "Generic (PLEG): container finished" podID="6e313ec3-70c2-4220-b08a-ef79b8a25847" containerID="9fbede642703b66da34b00feaaa8a19598b017d4abe7de06ff6b12a64a93636f" exitCode=0 Nov 25 15:13:56 crc kubenswrapper[4890]: I1125 15:13:56.662981 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xldhl" event={"ID":"6e313ec3-70c2-4220-b08a-ef79b8a25847","Type":"ContainerDied","Data":"9fbede642703b66da34b00feaaa8a19598b017d4abe7de06ff6b12a64a93636f"} Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.496569 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xldhl" Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.548687 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7dbb7f8b57-qtlvh"] Nov 25 15:13:57 crc kubenswrapper[4890]: E1125 15:13:57.549001 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e313ec3-70c2-4220-b08a-ef79b8a25847" containerName="controller-manager" Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.549020 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e313ec3-70c2-4220-b08a-ef79b8a25847" containerName="controller-manager" Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.549231 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e313ec3-70c2-4220-b08a-ef79b8a25847-config\") pod \"6e313ec3-70c2-4220-b08a-ef79b8a25847\" (UID: \"6e313ec3-70c2-4220-b08a-ef79b8a25847\") " Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.549341 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e313ec3-70c2-4220-b08a-ef79b8a25847-serving-cert\") pod \"6e313ec3-70c2-4220-b08a-ef79b8a25847\" (UID: \"6e313ec3-70c2-4220-b08a-ef79b8a25847\") " Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.549419 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6e313ec3-70c2-4220-b08a-ef79b8a25847-proxy-ca-bundles\") pod \"6e313ec3-70c2-4220-b08a-ef79b8a25847\" (UID: \"6e313ec3-70c2-4220-b08a-ef79b8a25847\") " Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.549484 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qfl4f\" (UniqueName: \"kubernetes.io/projected/6e313ec3-70c2-4220-b08a-ef79b8a25847-kube-api-access-qfl4f\") pod \"6e313ec3-70c2-4220-b08a-ef79b8a25847\" (UID: \"6e313ec3-70c2-4220-b08a-ef79b8a25847\") " Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.549538 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6e313ec3-70c2-4220-b08a-ef79b8a25847-client-ca\") pod \"6e313ec3-70c2-4220-b08a-ef79b8a25847\" (UID: \"6e313ec3-70c2-4220-b08a-ef79b8a25847\") " Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.550634 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e313ec3-70c2-4220-b08a-ef79b8a25847-client-ca" (OuterVolumeSpecName: "client-ca") pod "6e313ec3-70c2-4220-b08a-ef79b8a25847" (UID: "6e313ec3-70c2-4220-b08a-ef79b8a25847"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.549261 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e313ec3-70c2-4220-b08a-ef79b8a25847" containerName="controller-manager" Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.551417 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7dbb7f8b57-qtlvh" Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.554266 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e313ec3-70c2-4220-b08a-ef79b8a25847-config" (OuterVolumeSpecName: "config") pod "6e313ec3-70c2-4220-b08a-ef79b8a25847" (UID: "6e313ec3-70c2-4220-b08a-ef79b8a25847"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.555382 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e313ec3-70c2-4220-b08a-ef79b8a25847-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "6e313ec3-70c2-4220-b08a-ef79b8a25847" (UID: "6e313ec3-70c2-4220-b08a-ef79b8a25847"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.564201 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e313ec3-70c2-4220-b08a-ef79b8a25847-kube-api-access-qfl4f" (OuterVolumeSpecName: "kube-api-access-qfl4f") pod "6e313ec3-70c2-4220-b08a-ef79b8a25847" (UID: "6e313ec3-70c2-4220-b08a-ef79b8a25847"). InnerVolumeSpecName "kube-api-access-qfl4f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.566126 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e313ec3-70c2-4220-b08a-ef79b8a25847-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6e313ec3-70c2-4220-b08a-ef79b8a25847" (UID: "6e313ec3-70c2-4220-b08a-ef79b8a25847"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.568743 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mqvqx" Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.574728 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7dbb7f8b57-qtlvh"] Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.651848 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e09b8f75-2286-4a12-9571-5342b681690f-config\") pod \"e09b8f75-2286-4a12-9571-5342b681690f\" (UID: \"e09b8f75-2286-4a12-9571-5342b681690f\") " Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.651932 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e09b8f75-2286-4a12-9571-5342b681690f-client-ca\") pod \"e09b8f75-2286-4a12-9571-5342b681690f\" (UID: \"e09b8f75-2286-4a12-9571-5342b681690f\") " Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.651994 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lhs55\" (UniqueName: \"kubernetes.io/projected/e09b8f75-2286-4a12-9571-5342b681690f-kube-api-access-lhs55\") pod \"e09b8f75-2286-4a12-9571-5342b681690f\" (UID: \"e09b8f75-2286-4a12-9571-5342b681690f\") " Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.652083 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e09b8f75-2286-4a12-9571-5342b681690f-serving-cert\") pod \"e09b8f75-2286-4a12-9571-5342b681690f\" (UID: \"e09b8f75-2286-4a12-9571-5342b681690f\") " Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.652368 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/464b19b9-c536-4601-adba-598a9777d43d-proxy-ca-bundles\") pod \"controller-manager-7dbb7f8b57-qtlvh\" (UID: \"464b19b9-c536-4601-adba-598a9777d43d\") " pod="openshift-controller-manager/controller-manager-7dbb7f8b57-qtlvh" Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.652397 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/464b19b9-c536-4601-adba-598a9777d43d-config\") pod \"controller-manager-7dbb7f8b57-qtlvh\" (UID: \"464b19b9-c536-4601-adba-598a9777d43d\") " pod="openshift-controller-manager/controller-manager-7dbb7f8b57-qtlvh" Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.652433 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ncbc\" (UniqueName: \"kubernetes.io/projected/464b19b9-c536-4601-adba-598a9777d43d-kube-api-access-9ncbc\") pod \"controller-manager-7dbb7f8b57-qtlvh\" (UID: \"464b19b9-c536-4601-adba-598a9777d43d\") " pod="openshift-controller-manager/controller-manager-7dbb7f8b57-qtlvh" Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.652462 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/464b19b9-c536-4601-adba-598a9777d43d-serving-cert\") pod \"controller-manager-7dbb7f8b57-qtlvh\" (UID: \"464b19b9-c536-4601-adba-598a9777d43d\") " pod="openshift-controller-manager/controller-manager-7dbb7f8b57-qtlvh" Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.652489 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/464b19b9-c536-4601-adba-598a9777d43d-client-ca\") pod \"controller-manager-7dbb7f8b57-qtlvh\" (UID: \"464b19b9-c536-4601-adba-598a9777d43d\") " pod="openshift-controller-manager/controller-manager-7dbb7f8b57-qtlvh" Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.652556 4890 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e313ec3-70c2-4220-b08a-ef79b8a25847-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.652569 4890 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e313ec3-70c2-4220-b08a-ef79b8a25847-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.652578 4890 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6e313ec3-70c2-4220-b08a-ef79b8a25847-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.652588 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qfl4f\" (UniqueName: \"kubernetes.io/projected/6e313ec3-70c2-4220-b08a-ef79b8a25847-kube-api-access-qfl4f\") on node \"crc\" DevicePath \"\"" Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.652598 4890 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6e313ec3-70c2-4220-b08a-ef79b8a25847-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.653774 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e09b8f75-2286-4a12-9571-5342b681690f-config" (OuterVolumeSpecName: "config") pod "e09b8f75-2286-4a12-9571-5342b681690f" (UID: "e09b8f75-2286-4a12-9571-5342b681690f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.654203 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e09b8f75-2286-4a12-9571-5342b681690f-client-ca" (OuterVolumeSpecName: "client-ca") pod "e09b8f75-2286-4a12-9571-5342b681690f" (UID: "e09b8f75-2286-4a12-9571-5342b681690f"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.658794 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e09b8f75-2286-4a12-9571-5342b681690f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e09b8f75-2286-4a12-9571-5342b681690f" (UID: "e09b8f75-2286-4a12-9571-5342b681690f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.660184 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e09b8f75-2286-4a12-9571-5342b681690f-kube-api-access-lhs55" (OuterVolumeSpecName: "kube-api-access-lhs55") pod "e09b8f75-2286-4a12-9571-5342b681690f" (UID: "e09b8f75-2286-4a12-9571-5342b681690f"). InnerVolumeSpecName "kube-api-access-lhs55". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.680592 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mqvqx" event={"ID":"e09b8f75-2286-4a12-9571-5342b681690f","Type":"ContainerDied","Data":"7503f9b1edca457c0b527a0174f9a5cc859a63fdb99e56b388c0337d94c57ab8"} Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.680670 4890 scope.go:117] "RemoveContainer" containerID="54c69ecc23762a33f8d6eeca868dd59ac28361f509178948acf91cd5a6241e65" Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.680797 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mqvqx" Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.685371 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xldhl" event={"ID":"6e313ec3-70c2-4220-b08a-ef79b8a25847","Type":"ContainerDied","Data":"1129dc184de9e2277418c632ff648e55150aaf4685472496dc6ab548f48fdf03"} Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.685456 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xldhl" Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.720584 4890 scope.go:117] "RemoveContainer" containerID="9fbede642703b66da34b00feaaa8a19598b017d4abe7de06ff6b12a64a93636f" Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.730853 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xldhl"] Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.739152 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xldhl"] Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.743066 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-mqvqx"] Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.745502 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-mqvqx"] Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.753513 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/464b19b9-c536-4601-adba-598a9777d43d-proxy-ca-bundles\") pod \"controller-manager-7dbb7f8b57-qtlvh\" (UID: \"464b19b9-c536-4601-adba-598a9777d43d\") " pod="openshift-controller-manager/controller-manager-7dbb7f8b57-qtlvh" Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.753555 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/464b19b9-c536-4601-adba-598a9777d43d-config\") pod \"controller-manager-7dbb7f8b57-qtlvh\" (UID: \"464b19b9-c536-4601-adba-598a9777d43d\") " pod="openshift-controller-manager/controller-manager-7dbb7f8b57-qtlvh" Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.753592 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ncbc\" (UniqueName: \"kubernetes.io/projected/464b19b9-c536-4601-adba-598a9777d43d-kube-api-access-9ncbc\") pod \"controller-manager-7dbb7f8b57-qtlvh\" (UID: \"464b19b9-c536-4601-adba-598a9777d43d\") " pod="openshift-controller-manager/controller-manager-7dbb7f8b57-qtlvh" Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.753622 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/464b19b9-c536-4601-adba-598a9777d43d-serving-cert\") pod \"controller-manager-7dbb7f8b57-qtlvh\" (UID: \"464b19b9-c536-4601-adba-598a9777d43d\") " pod="openshift-controller-manager/controller-manager-7dbb7f8b57-qtlvh" Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.753650 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/464b19b9-c536-4601-adba-598a9777d43d-client-ca\") pod \"controller-manager-7dbb7f8b57-qtlvh\" (UID: \"464b19b9-c536-4601-adba-598a9777d43d\") " pod="openshift-controller-manager/controller-manager-7dbb7f8b57-qtlvh" Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.753750 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lhs55\" (UniqueName: \"kubernetes.io/projected/e09b8f75-2286-4a12-9571-5342b681690f-kube-api-access-lhs55\") on node \"crc\" DevicePath \"\"" Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.753765 4890 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e09b8f75-2286-4a12-9571-5342b681690f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.753774 4890 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e09b8f75-2286-4a12-9571-5342b681690f-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.753784 4890 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e09b8f75-2286-4a12-9571-5342b681690f-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.754927 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/464b19b9-c536-4601-adba-598a9777d43d-client-ca\") pod \"controller-manager-7dbb7f8b57-qtlvh\" (UID: \"464b19b9-c536-4601-adba-598a9777d43d\") " pod="openshift-controller-manager/controller-manager-7dbb7f8b57-qtlvh" Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.755152 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/464b19b9-c536-4601-adba-598a9777d43d-proxy-ca-bundles\") pod \"controller-manager-7dbb7f8b57-qtlvh\" (UID: \"464b19b9-c536-4601-adba-598a9777d43d\") " pod="openshift-controller-manager/controller-manager-7dbb7f8b57-qtlvh" Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.755253 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/464b19b9-c536-4601-adba-598a9777d43d-config\") pod \"controller-manager-7dbb7f8b57-qtlvh\" (UID: \"464b19b9-c536-4601-adba-598a9777d43d\") " pod="openshift-controller-manager/controller-manager-7dbb7f8b57-qtlvh" Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.758231 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/464b19b9-c536-4601-adba-598a9777d43d-serving-cert\") pod \"controller-manager-7dbb7f8b57-qtlvh\" (UID: \"464b19b9-c536-4601-adba-598a9777d43d\") " pod="openshift-controller-manager/controller-manager-7dbb7f8b57-qtlvh" Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.774526 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ncbc\" (UniqueName: \"kubernetes.io/projected/464b19b9-c536-4601-adba-598a9777d43d-kube-api-access-9ncbc\") pod \"controller-manager-7dbb7f8b57-qtlvh\" (UID: \"464b19b9-c536-4601-adba-598a9777d43d\") " pod="openshift-controller-manager/controller-manager-7dbb7f8b57-qtlvh" Nov 25 15:13:57 crc kubenswrapper[4890]: I1125 15:13:57.897893 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7dbb7f8b57-qtlvh" Nov 25 15:13:58 crc kubenswrapper[4890]: I1125 15:13:58.171807 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7dbb7f8b57-qtlvh"] Nov 25 15:13:58 crc kubenswrapper[4890]: I1125 15:13:58.182318 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e313ec3-70c2-4220-b08a-ef79b8a25847" path="/var/lib/kubelet/pods/6e313ec3-70c2-4220-b08a-ef79b8a25847/volumes" Nov 25 15:13:58 crc kubenswrapper[4890]: I1125 15:13:58.183094 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e09b8f75-2286-4a12-9571-5342b681690f" path="/var/lib/kubelet/pods/e09b8f75-2286-4a12-9571-5342b681690f/volumes" Nov 25 15:13:58 crc kubenswrapper[4890]: I1125 15:13:58.693761 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7dbb7f8b57-qtlvh" event={"ID":"464b19b9-c536-4601-adba-598a9777d43d","Type":"ContainerStarted","Data":"4927fb8febd6548d174d767498249ccc1b6c08ef479da38c3575938277bc08b4"} Nov 25 15:13:58 crc kubenswrapper[4890]: I1125 15:13:58.693845 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7dbb7f8b57-qtlvh" event={"ID":"464b19b9-c536-4601-adba-598a9777d43d","Type":"ContainerStarted","Data":"bfdd3f8e7e8b9ddf0cc49787e8a084da37404e88d89c0333ee464ebccc880fb4"} Nov 25 15:13:58 crc kubenswrapper[4890]: I1125 15:13:58.693866 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7dbb7f8b57-qtlvh" Nov 25 15:13:58 crc kubenswrapper[4890]: I1125 15:13:58.699281 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7dbb7f8b57-qtlvh" Nov 25 15:13:58 crc kubenswrapper[4890]: I1125 15:13:58.720659 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7dbb7f8b57-qtlvh" podStartSLOduration=1.720639111 podStartE2EDuration="1.720639111s" podCreationTimestamp="2025-11-25 15:13:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:13:58.713478742 +0000 UTC m=+697.155941342" watchObservedRunningTime="2025-11-25 15:13:58.720639111 +0000 UTC m=+697.163101721" Nov 25 15:13:59 crc kubenswrapper[4890]: I1125 15:13:59.021500 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59f8fb67d5-r49dm"] Nov 25 15:13:59 crc kubenswrapper[4890]: E1125 15:13:59.021785 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e09b8f75-2286-4a12-9571-5342b681690f" containerName="route-controller-manager" Nov 25 15:13:59 crc kubenswrapper[4890]: I1125 15:13:59.021807 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="e09b8f75-2286-4a12-9571-5342b681690f" containerName="route-controller-manager" Nov 25 15:13:59 crc kubenswrapper[4890]: I1125 15:13:59.021936 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="e09b8f75-2286-4a12-9571-5342b681690f" containerName="route-controller-manager" Nov 25 15:13:59 crc kubenswrapper[4890]: I1125 15:13:59.022410 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59f8fb67d5-r49dm" Nov 25 15:13:59 crc kubenswrapper[4890]: I1125 15:13:59.024550 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 25 15:13:59 crc kubenswrapper[4890]: I1125 15:13:59.025701 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 25 15:13:59 crc kubenswrapper[4890]: I1125 15:13:59.034709 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 25 15:13:59 crc kubenswrapper[4890]: I1125 15:13:59.035423 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 25 15:13:59 crc kubenswrapper[4890]: I1125 15:13:59.036637 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 25 15:13:59 crc kubenswrapper[4890]: I1125 15:13:59.036876 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 25 15:13:59 crc kubenswrapper[4890]: I1125 15:13:59.041099 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59f8fb67d5-r49dm"] Nov 25 15:13:59 crc kubenswrapper[4890]: I1125 15:13:59.072956 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22sdw\" (UniqueName: \"kubernetes.io/projected/820cd23d-a547-4264-9715-b5afc66af87b-kube-api-access-22sdw\") pod \"route-controller-manager-59f8fb67d5-r49dm\" (UID: \"820cd23d-a547-4264-9715-b5afc66af87b\") " pod="openshift-route-controller-manager/route-controller-manager-59f8fb67d5-r49dm" Nov 25 15:13:59 crc kubenswrapper[4890]: I1125 15:13:59.073071 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/820cd23d-a547-4264-9715-b5afc66af87b-serving-cert\") pod \"route-controller-manager-59f8fb67d5-r49dm\" (UID: \"820cd23d-a547-4264-9715-b5afc66af87b\") " pod="openshift-route-controller-manager/route-controller-manager-59f8fb67d5-r49dm" Nov 25 15:13:59 crc kubenswrapper[4890]: I1125 15:13:59.073101 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/820cd23d-a547-4264-9715-b5afc66af87b-client-ca\") pod \"route-controller-manager-59f8fb67d5-r49dm\" (UID: \"820cd23d-a547-4264-9715-b5afc66af87b\") " pod="openshift-route-controller-manager/route-controller-manager-59f8fb67d5-r49dm" Nov 25 15:13:59 crc kubenswrapper[4890]: I1125 15:13:59.073225 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/820cd23d-a547-4264-9715-b5afc66af87b-config\") pod \"route-controller-manager-59f8fb67d5-r49dm\" (UID: \"820cd23d-a547-4264-9715-b5afc66af87b\") " pod="openshift-route-controller-manager/route-controller-manager-59f8fb67d5-r49dm" Nov 25 15:13:59 crc kubenswrapper[4890]: I1125 15:13:59.174083 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/820cd23d-a547-4264-9715-b5afc66af87b-config\") pod \"route-controller-manager-59f8fb67d5-r49dm\" (UID: \"820cd23d-a547-4264-9715-b5afc66af87b\") " pod="openshift-route-controller-manager/route-controller-manager-59f8fb67d5-r49dm" Nov 25 15:13:59 crc kubenswrapper[4890]: I1125 15:13:59.174143 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22sdw\" (UniqueName: \"kubernetes.io/projected/820cd23d-a547-4264-9715-b5afc66af87b-kube-api-access-22sdw\") pod \"route-controller-manager-59f8fb67d5-r49dm\" (UID: \"820cd23d-a547-4264-9715-b5afc66af87b\") " pod="openshift-route-controller-manager/route-controller-manager-59f8fb67d5-r49dm" Nov 25 15:13:59 crc kubenswrapper[4890]: I1125 15:13:59.174201 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/820cd23d-a547-4264-9715-b5afc66af87b-serving-cert\") pod \"route-controller-manager-59f8fb67d5-r49dm\" (UID: \"820cd23d-a547-4264-9715-b5afc66af87b\") " pod="openshift-route-controller-manager/route-controller-manager-59f8fb67d5-r49dm" Nov 25 15:13:59 crc kubenswrapper[4890]: I1125 15:13:59.174224 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/820cd23d-a547-4264-9715-b5afc66af87b-client-ca\") pod \"route-controller-manager-59f8fb67d5-r49dm\" (UID: \"820cd23d-a547-4264-9715-b5afc66af87b\") " pod="openshift-route-controller-manager/route-controller-manager-59f8fb67d5-r49dm" Nov 25 15:13:59 crc kubenswrapper[4890]: I1125 15:13:59.175374 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/820cd23d-a547-4264-9715-b5afc66af87b-config\") pod \"route-controller-manager-59f8fb67d5-r49dm\" (UID: \"820cd23d-a547-4264-9715-b5afc66af87b\") " pod="openshift-route-controller-manager/route-controller-manager-59f8fb67d5-r49dm" Nov 25 15:13:59 crc kubenswrapper[4890]: I1125 15:13:59.175371 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/820cd23d-a547-4264-9715-b5afc66af87b-client-ca\") pod \"route-controller-manager-59f8fb67d5-r49dm\" (UID: \"820cd23d-a547-4264-9715-b5afc66af87b\") " pod="openshift-route-controller-manager/route-controller-manager-59f8fb67d5-r49dm" Nov 25 15:13:59 crc kubenswrapper[4890]: I1125 15:13:59.183443 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/820cd23d-a547-4264-9715-b5afc66af87b-serving-cert\") pod \"route-controller-manager-59f8fb67d5-r49dm\" (UID: \"820cd23d-a547-4264-9715-b5afc66af87b\") " pod="openshift-route-controller-manager/route-controller-manager-59f8fb67d5-r49dm" Nov 25 15:13:59 crc kubenswrapper[4890]: I1125 15:13:59.195392 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22sdw\" (UniqueName: \"kubernetes.io/projected/820cd23d-a547-4264-9715-b5afc66af87b-kube-api-access-22sdw\") pod \"route-controller-manager-59f8fb67d5-r49dm\" (UID: \"820cd23d-a547-4264-9715-b5afc66af87b\") " pod="openshift-route-controller-manager/route-controller-manager-59f8fb67d5-r49dm" Nov 25 15:13:59 crc kubenswrapper[4890]: I1125 15:13:59.337153 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59f8fb67d5-r49dm" Nov 25 15:13:59 crc kubenswrapper[4890]: I1125 15:13:59.634193 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59f8fb67d5-r49dm"] Nov 25 15:13:59 crc kubenswrapper[4890]: I1125 15:13:59.706755 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-59f8fb67d5-r49dm" event={"ID":"820cd23d-a547-4264-9715-b5afc66af87b","Type":"ContainerStarted","Data":"62bf9a1a576b41a2f599ef973938d55dae64d59952c5ca956a87b722c79cc92e"} Nov 25 15:14:00 crc kubenswrapper[4890]: I1125 15:14:00.332600 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-6b856c5bd9-slq4g" Nov 25 15:14:00 crc kubenswrapper[4890]: I1125 15:14:00.716123 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-59f8fb67d5-r49dm" event={"ID":"820cd23d-a547-4264-9715-b5afc66af87b","Type":"ContainerStarted","Data":"61d3d48b0131d6badd881a7a35844a23f1052972f9a28ffb38a39efee0398618"} Nov 25 15:14:00 crc kubenswrapper[4890]: I1125 15:14:00.716383 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-59f8fb67d5-r49dm" Nov 25 15:14:00 crc kubenswrapper[4890]: I1125 15:14:00.723262 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-59f8fb67d5-r49dm" Nov 25 15:14:00 crc kubenswrapper[4890]: I1125 15:14:00.736825 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-59f8fb67d5-r49dm" podStartSLOduration=3.736807068 podStartE2EDuration="3.736807068s" podCreationTimestamp="2025-11-25 15:13:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:14:00.73488944 +0000 UTC m=+699.177352090" watchObservedRunningTime="2025-11-25 15:14:00.736807068 +0000 UTC m=+699.179269678" Nov 25 15:14:09 crc kubenswrapper[4890]: I1125 15:14:09.517471 4890 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 25 15:14:19 crc kubenswrapper[4890]: I1125 15:14:19.938215 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-66fc8c7fbf-lnjxm" Nov 25 15:14:20 crc kubenswrapper[4890]: I1125 15:14:20.769489 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-nhqzp"] Nov 25 15:14:20 crc kubenswrapper[4890]: I1125 15:14:20.772780 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-nhqzp" Nov 25 15:14:20 crc kubenswrapper[4890]: I1125 15:14:20.786344 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 25 15:14:20 crc kubenswrapper[4890]: I1125 15:14:20.789783 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-9ncz9"] Nov 25 15:14:20 crc kubenswrapper[4890]: I1125 15:14:20.790645 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-9ncz9" Nov 25 15:14:20 crc kubenswrapper[4890]: I1125 15:14:20.791047 4890 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-lv5hp" Nov 25 15:14:20 crc kubenswrapper[4890]: I1125 15:14:20.799735 4890 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 25 15:14:20 crc kubenswrapper[4890]: I1125 15:14:20.799773 4890 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 25 15:14:20 crc kubenswrapper[4890]: I1125 15:14:20.805466 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-9ncz9"] Nov 25 15:14:20 crc kubenswrapper[4890]: I1125 15:14:20.808711 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/742ce93d-d49e-4b12-a1cf-5413ec8475ac-cert\") pod \"frr-k8s-webhook-server-6998585d5-9ncz9\" (UID: \"742ce93d-d49e-4b12-a1cf-5413ec8475ac\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-9ncz9" Nov 25 15:14:20 crc kubenswrapper[4890]: I1125 15:14:20.808767 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxq7x\" (UniqueName: \"kubernetes.io/projected/60b160c7-c63e-4ffc-b031-9c9fa5c248e2-kube-api-access-nxq7x\") pod \"frr-k8s-nhqzp\" (UID: \"60b160c7-c63e-4ffc-b031-9c9fa5c248e2\") " pod="metallb-system/frr-k8s-nhqzp" Nov 25 15:14:20 crc kubenswrapper[4890]: I1125 15:14:20.808812 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/60b160c7-c63e-4ffc-b031-9c9fa5c248e2-metrics\") pod \"frr-k8s-nhqzp\" (UID: \"60b160c7-c63e-4ffc-b031-9c9fa5c248e2\") " pod="metallb-system/frr-k8s-nhqzp" Nov 25 15:14:20 crc kubenswrapper[4890]: I1125 15:14:20.808886 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/60b160c7-c63e-4ffc-b031-9c9fa5c248e2-frr-startup\") pod \"frr-k8s-nhqzp\" (UID: \"60b160c7-c63e-4ffc-b031-9c9fa5c248e2\") " pod="metallb-system/frr-k8s-nhqzp" Nov 25 15:14:20 crc kubenswrapper[4890]: I1125 15:14:20.808985 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/60b160c7-c63e-4ffc-b031-9c9fa5c248e2-reloader\") pod \"frr-k8s-nhqzp\" (UID: \"60b160c7-c63e-4ffc-b031-9c9fa5c248e2\") " pod="metallb-system/frr-k8s-nhqzp" Nov 25 15:14:20 crc kubenswrapper[4890]: I1125 15:14:20.809018 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/60b160c7-c63e-4ffc-b031-9c9fa5c248e2-frr-sockets\") pod \"frr-k8s-nhqzp\" (UID: \"60b160c7-c63e-4ffc-b031-9c9fa5c248e2\") " pod="metallb-system/frr-k8s-nhqzp" Nov 25 15:14:20 crc kubenswrapper[4890]: I1125 15:14:20.809033 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/60b160c7-c63e-4ffc-b031-9c9fa5c248e2-metrics-certs\") pod \"frr-k8s-nhqzp\" (UID: \"60b160c7-c63e-4ffc-b031-9c9fa5c248e2\") " pod="metallb-system/frr-k8s-nhqzp" Nov 25 15:14:20 crc kubenswrapper[4890]: I1125 15:14:20.809074 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/60b160c7-c63e-4ffc-b031-9c9fa5c248e2-frr-conf\") pod \"frr-k8s-nhqzp\" (UID: \"60b160c7-c63e-4ffc-b031-9c9fa5c248e2\") " pod="metallb-system/frr-k8s-nhqzp" Nov 25 15:14:20 crc kubenswrapper[4890]: I1125 15:14:20.809124 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hg762\" (UniqueName: \"kubernetes.io/projected/742ce93d-d49e-4b12-a1cf-5413ec8475ac-kube-api-access-hg762\") pod \"frr-k8s-webhook-server-6998585d5-9ncz9\" (UID: \"742ce93d-d49e-4b12-a1cf-5413ec8475ac\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-9ncz9" Nov 25 15:14:20 crc kubenswrapper[4890]: I1125 15:14:20.909837 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/60b160c7-c63e-4ffc-b031-9c9fa5c248e2-reloader\") pod \"frr-k8s-nhqzp\" (UID: \"60b160c7-c63e-4ffc-b031-9c9fa5c248e2\") " pod="metallb-system/frr-k8s-nhqzp" Nov 25 15:14:20 crc kubenswrapper[4890]: I1125 15:14:20.909895 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/60b160c7-c63e-4ffc-b031-9c9fa5c248e2-frr-sockets\") pod \"frr-k8s-nhqzp\" (UID: \"60b160c7-c63e-4ffc-b031-9c9fa5c248e2\") " pod="metallb-system/frr-k8s-nhqzp" Nov 25 15:14:20 crc kubenswrapper[4890]: I1125 15:14:20.909925 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/60b160c7-c63e-4ffc-b031-9c9fa5c248e2-metrics-certs\") pod \"frr-k8s-nhqzp\" (UID: \"60b160c7-c63e-4ffc-b031-9c9fa5c248e2\") " pod="metallb-system/frr-k8s-nhqzp" Nov 25 15:14:20 crc kubenswrapper[4890]: I1125 15:14:20.909946 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/60b160c7-c63e-4ffc-b031-9c9fa5c248e2-frr-conf\") pod \"frr-k8s-nhqzp\" (UID: \"60b160c7-c63e-4ffc-b031-9c9fa5c248e2\") " pod="metallb-system/frr-k8s-nhqzp" Nov 25 15:14:20 crc kubenswrapper[4890]: I1125 15:14:20.909975 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hg762\" (UniqueName: \"kubernetes.io/projected/742ce93d-d49e-4b12-a1cf-5413ec8475ac-kube-api-access-hg762\") pod \"frr-k8s-webhook-server-6998585d5-9ncz9\" (UID: \"742ce93d-d49e-4b12-a1cf-5413ec8475ac\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-9ncz9" Nov 25 15:14:20 crc kubenswrapper[4890]: I1125 15:14:20.909998 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/742ce93d-d49e-4b12-a1cf-5413ec8475ac-cert\") pod \"frr-k8s-webhook-server-6998585d5-9ncz9\" (UID: \"742ce93d-d49e-4b12-a1cf-5413ec8475ac\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-9ncz9" Nov 25 15:14:20 crc kubenswrapper[4890]: I1125 15:14:20.910021 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxq7x\" (UniqueName: \"kubernetes.io/projected/60b160c7-c63e-4ffc-b031-9c9fa5c248e2-kube-api-access-nxq7x\") pod \"frr-k8s-nhqzp\" (UID: \"60b160c7-c63e-4ffc-b031-9c9fa5c248e2\") " pod="metallb-system/frr-k8s-nhqzp" Nov 25 15:14:20 crc kubenswrapper[4890]: I1125 15:14:20.910040 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/60b160c7-c63e-4ffc-b031-9c9fa5c248e2-metrics\") pod \"frr-k8s-nhqzp\" (UID: \"60b160c7-c63e-4ffc-b031-9c9fa5c248e2\") " pod="metallb-system/frr-k8s-nhqzp" Nov 25 15:14:20 crc kubenswrapper[4890]: I1125 15:14:20.910065 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/60b160c7-c63e-4ffc-b031-9c9fa5c248e2-frr-startup\") pod \"frr-k8s-nhqzp\" (UID: \"60b160c7-c63e-4ffc-b031-9c9fa5c248e2\") " pod="metallb-system/frr-k8s-nhqzp" Nov 25 15:14:20 crc kubenswrapper[4890]: I1125 15:14:20.910456 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/60b160c7-c63e-4ffc-b031-9c9fa5c248e2-reloader\") pod \"frr-k8s-nhqzp\" (UID: \"60b160c7-c63e-4ffc-b031-9c9fa5c248e2\") " pod="metallb-system/frr-k8s-nhqzp" Nov 25 15:14:20 crc kubenswrapper[4890]: I1125 15:14:20.910982 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/60b160c7-c63e-4ffc-b031-9c9fa5c248e2-frr-startup\") pod \"frr-k8s-nhqzp\" (UID: \"60b160c7-c63e-4ffc-b031-9c9fa5c248e2\") " pod="metallb-system/frr-k8s-nhqzp" Nov 25 15:14:20 crc kubenswrapper[4890]: I1125 15:14:20.911016 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/60b160c7-c63e-4ffc-b031-9c9fa5c248e2-frr-sockets\") pod \"frr-k8s-nhqzp\" (UID: \"60b160c7-c63e-4ffc-b031-9c9fa5c248e2\") " pod="metallb-system/frr-k8s-nhqzp" Nov 25 15:14:20 crc kubenswrapper[4890]: E1125 15:14:20.911084 4890 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Nov 25 15:14:20 crc kubenswrapper[4890]: E1125 15:14:20.911103 4890 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Nov 25 15:14:20 crc kubenswrapper[4890]: E1125 15:14:20.911130 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/742ce93d-d49e-4b12-a1cf-5413ec8475ac-cert podName:742ce93d-d49e-4b12-a1cf-5413ec8475ac nodeName:}" failed. No retries permitted until 2025-11-25 15:14:21.411114252 +0000 UTC m=+719.853576862 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/742ce93d-d49e-4b12-a1cf-5413ec8475ac-cert") pod "frr-k8s-webhook-server-6998585d5-9ncz9" (UID: "742ce93d-d49e-4b12-a1cf-5413ec8475ac") : secret "frr-k8s-webhook-server-cert" not found Nov 25 15:14:20 crc kubenswrapper[4890]: E1125 15:14:20.911178 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/60b160c7-c63e-4ffc-b031-9c9fa5c248e2-metrics-certs podName:60b160c7-c63e-4ffc-b031-9c9fa5c248e2 nodeName:}" failed. No retries permitted until 2025-11-25 15:14:21.411140073 +0000 UTC m=+719.853602743 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/60b160c7-c63e-4ffc-b031-9c9fa5c248e2-metrics-certs") pod "frr-k8s-nhqzp" (UID: "60b160c7-c63e-4ffc-b031-9c9fa5c248e2") : secret "frr-k8s-certs-secret" not found Nov 25 15:14:20 crc kubenswrapper[4890]: I1125 15:14:20.911335 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/60b160c7-c63e-4ffc-b031-9c9fa5c248e2-metrics\") pod \"frr-k8s-nhqzp\" (UID: \"60b160c7-c63e-4ffc-b031-9c9fa5c248e2\") " pod="metallb-system/frr-k8s-nhqzp" Nov 25 15:14:20 crc kubenswrapper[4890]: I1125 15:14:20.913801 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-ndl5n"] Nov 25 15:14:20 crc kubenswrapper[4890]: I1125 15:14:20.914860 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-ndl5n" Nov 25 15:14:20 crc kubenswrapper[4890]: I1125 15:14:20.915471 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/60b160c7-c63e-4ffc-b031-9c9fa5c248e2-frr-conf\") pod \"frr-k8s-nhqzp\" (UID: \"60b160c7-c63e-4ffc-b031-9c9fa5c248e2\") " pod="metallb-system/frr-k8s-nhqzp" Nov 25 15:14:20 crc kubenswrapper[4890]: I1125 15:14:20.920515 4890 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 25 15:14:20 crc kubenswrapper[4890]: I1125 15:14:20.920597 4890 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-k7z2s" Nov 25 15:14:20 crc kubenswrapper[4890]: I1125 15:14:20.920748 4890 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 25 15:14:20 crc kubenswrapper[4890]: I1125 15:14:20.922725 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 25 15:14:20 crc kubenswrapper[4890]: I1125 15:14:20.932256 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6c7b4b5f48-hknzk"] Nov 25 15:14:20 crc kubenswrapper[4890]: I1125 15:14:20.933328 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-hknzk" Nov 25 15:14:20 crc kubenswrapper[4890]: I1125 15:14:20.939437 4890 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 25 15:14:20 crc kubenswrapper[4890]: I1125 15:14:20.958360 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-hknzk"] Nov 25 15:14:20 crc kubenswrapper[4890]: I1125 15:14:20.968239 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxq7x\" (UniqueName: \"kubernetes.io/projected/60b160c7-c63e-4ffc-b031-9c9fa5c248e2-kube-api-access-nxq7x\") pod \"frr-k8s-nhqzp\" (UID: \"60b160c7-c63e-4ffc-b031-9c9fa5c248e2\") " pod="metallb-system/frr-k8s-nhqzp" Nov 25 15:14:20 crc kubenswrapper[4890]: I1125 15:14:20.968299 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hg762\" (UniqueName: \"kubernetes.io/projected/742ce93d-d49e-4b12-a1cf-5413ec8475ac-kube-api-access-hg762\") pod \"frr-k8s-webhook-server-6998585d5-9ncz9\" (UID: \"742ce93d-d49e-4b12-a1cf-5413ec8475ac\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-9ncz9" Nov 25 15:14:21 crc kubenswrapper[4890]: I1125 15:14:21.010985 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d97f8477-4597-4fb0-b2fc-e3faf124ba4e-metrics-certs\") pod \"controller-6c7b4b5f48-hknzk\" (UID: \"d97f8477-4597-4fb0-b2fc-e3faf124ba4e\") " pod="metallb-system/controller-6c7b4b5f48-hknzk" Nov 25 15:14:21 crc kubenswrapper[4890]: I1125 15:14:21.011044 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/9c563abc-f1fc-4c15-9a03-901371600338-memberlist\") pod \"speaker-ndl5n\" (UID: \"9c563abc-f1fc-4c15-9a03-901371600338\") " pod="metallb-system/speaker-ndl5n" Nov 25 15:14:21 crc kubenswrapper[4890]: I1125 15:14:21.011108 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9c563abc-f1fc-4c15-9a03-901371600338-metrics-certs\") pod \"speaker-ndl5n\" (UID: \"9c563abc-f1fc-4c15-9a03-901371600338\") " pod="metallb-system/speaker-ndl5n" Nov 25 15:14:21 crc kubenswrapper[4890]: I1125 15:14:21.011198 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/9c563abc-f1fc-4c15-9a03-901371600338-metallb-excludel2\") pod \"speaker-ndl5n\" (UID: \"9c563abc-f1fc-4c15-9a03-901371600338\") " pod="metallb-system/speaker-ndl5n" Nov 25 15:14:21 crc kubenswrapper[4890]: I1125 15:14:21.011233 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7784k\" (UniqueName: \"kubernetes.io/projected/9c563abc-f1fc-4c15-9a03-901371600338-kube-api-access-7784k\") pod \"speaker-ndl5n\" (UID: \"9c563abc-f1fc-4c15-9a03-901371600338\") " pod="metallb-system/speaker-ndl5n" Nov 25 15:14:21 crc kubenswrapper[4890]: I1125 15:14:21.011267 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d97f8477-4597-4fb0-b2fc-e3faf124ba4e-cert\") pod \"controller-6c7b4b5f48-hknzk\" (UID: \"d97f8477-4597-4fb0-b2fc-e3faf124ba4e\") " pod="metallb-system/controller-6c7b4b5f48-hknzk" Nov 25 15:14:21 crc kubenswrapper[4890]: I1125 15:14:21.011282 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84vm2\" (UniqueName: \"kubernetes.io/projected/d97f8477-4597-4fb0-b2fc-e3faf124ba4e-kube-api-access-84vm2\") pod \"controller-6c7b4b5f48-hknzk\" (UID: \"d97f8477-4597-4fb0-b2fc-e3faf124ba4e\") " pod="metallb-system/controller-6c7b4b5f48-hknzk" Nov 25 15:14:21 crc kubenswrapper[4890]: I1125 15:14:21.112786 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9c563abc-f1fc-4c15-9a03-901371600338-metrics-certs\") pod \"speaker-ndl5n\" (UID: \"9c563abc-f1fc-4c15-9a03-901371600338\") " pod="metallb-system/speaker-ndl5n" Nov 25 15:14:21 crc kubenswrapper[4890]: I1125 15:14:21.112867 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/9c563abc-f1fc-4c15-9a03-901371600338-metallb-excludel2\") pod \"speaker-ndl5n\" (UID: \"9c563abc-f1fc-4c15-9a03-901371600338\") " pod="metallb-system/speaker-ndl5n" Nov 25 15:14:21 crc kubenswrapper[4890]: I1125 15:14:21.112945 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7784k\" (UniqueName: \"kubernetes.io/projected/9c563abc-f1fc-4c15-9a03-901371600338-kube-api-access-7784k\") pod \"speaker-ndl5n\" (UID: \"9c563abc-f1fc-4c15-9a03-901371600338\") " pod="metallb-system/speaker-ndl5n" Nov 25 15:14:21 crc kubenswrapper[4890]: I1125 15:14:21.112978 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d97f8477-4597-4fb0-b2fc-e3faf124ba4e-cert\") pod \"controller-6c7b4b5f48-hknzk\" (UID: \"d97f8477-4597-4fb0-b2fc-e3faf124ba4e\") " pod="metallb-system/controller-6c7b4b5f48-hknzk" Nov 25 15:14:21 crc kubenswrapper[4890]: I1125 15:14:21.113005 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84vm2\" (UniqueName: \"kubernetes.io/projected/d97f8477-4597-4fb0-b2fc-e3faf124ba4e-kube-api-access-84vm2\") pod \"controller-6c7b4b5f48-hknzk\" (UID: \"d97f8477-4597-4fb0-b2fc-e3faf124ba4e\") " pod="metallb-system/controller-6c7b4b5f48-hknzk" Nov 25 15:14:21 crc kubenswrapper[4890]: I1125 15:14:21.113103 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d97f8477-4597-4fb0-b2fc-e3faf124ba4e-metrics-certs\") pod \"controller-6c7b4b5f48-hknzk\" (UID: \"d97f8477-4597-4fb0-b2fc-e3faf124ba4e\") " pod="metallb-system/controller-6c7b4b5f48-hknzk" Nov 25 15:14:21 crc kubenswrapper[4890]: I1125 15:14:21.113137 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/9c563abc-f1fc-4c15-9a03-901371600338-memberlist\") pod \"speaker-ndl5n\" (UID: \"9c563abc-f1fc-4c15-9a03-901371600338\") " pod="metallb-system/speaker-ndl5n" Nov 25 15:14:21 crc kubenswrapper[4890]: E1125 15:14:21.113261 4890 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Nov 25 15:14:21 crc kubenswrapper[4890]: E1125 15:14:21.113314 4890 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 25 15:14:21 crc kubenswrapper[4890]: E1125 15:14:21.113347 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d97f8477-4597-4fb0-b2fc-e3faf124ba4e-metrics-certs podName:d97f8477-4597-4fb0-b2fc-e3faf124ba4e nodeName:}" failed. No retries permitted until 2025-11-25 15:14:21.613325936 +0000 UTC m=+720.055788546 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d97f8477-4597-4fb0-b2fc-e3faf124ba4e-metrics-certs") pod "controller-6c7b4b5f48-hknzk" (UID: "d97f8477-4597-4fb0-b2fc-e3faf124ba4e") : secret "controller-certs-secret" not found Nov 25 15:14:21 crc kubenswrapper[4890]: E1125 15:14:21.113379 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9c563abc-f1fc-4c15-9a03-901371600338-memberlist podName:9c563abc-f1fc-4c15-9a03-901371600338 nodeName:}" failed. No retries permitted until 2025-11-25 15:14:21.613363137 +0000 UTC m=+720.055825757 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/9c563abc-f1fc-4c15-9a03-901371600338-memberlist") pod "speaker-ndl5n" (UID: "9c563abc-f1fc-4c15-9a03-901371600338") : secret "metallb-memberlist" not found Nov 25 15:14:21 crc kubenswrapper[4890]: I1125 15:14:21.113850 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/9c563abc-f1fc-4c15-9a03-901371600338-metallb-excludel2\") pod \"speaker-ndl5n\" (UID: \"9c563abc-f1fc-4c15-9a03-901371600338\") " pod="metallb-system/speaker-ndl5n" Nov 25 15:14:21 crc kubenswrapper[4890]: I1125 15:14:21.118475 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d97f8477-4597-4fb0-b2fc-e3faf124ba4e-cert\") pod \"controller-6c7b4b5f48-hknzk\" (UID: \"d97f8477-4597-4fb0-b2fc-e3faf124ba4e\") " pod="metallb-system/controller-6c7b4b5f48-hknzk" Nov 25 15:14:21 crc kubenswrapper[4890]: I1125 15:14:21.118505 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9c563abc-f1fc-4c15-9a03-901371600338-metrics-certs\") pod \"speaker-ndl5n\" (UID: \"9c563abc-f1fc-4c15-9a03-901371600338\") " pod="metallb-system/speaker-ndl5n" Nov 25 15:14:21 crc kubenswrapper[4890]: I1125 15:14:21.133894 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84vm2\" (UniqueName: \"kubernetes.io/projected/d97f8477-4597-4fb0-b2fc-e3faf124ba4e-kube-api-access-84vm2\") pod \"controller-6c7b4b5f48-hknzk\" (UID: \"d97f8477-4597-4fb0-b2fc-e3faf124ba4e\") " pod="metallb-system/controller-6c7b4b5f48-hknzk" Nov 25 15:14:21 crc kubenswrapper[4890]: I1125 15:14:21.138184 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7784k\" (UniqueName: \"kubernetes.io/projected/9c563abc-f1fc-4c15-9a03-901371600338-kube-api-access-7784k\") pod \"speaker-ndl5n\" (UID: \"9c563abc-f1fc-4c15-9a03-901371600338\") " pod="metallb-system/speaker-ndl5n" Nov 25 15:14:21 crc kubenswrapper[4890]: I1125 15:14:21.415504 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/60b160c7-c63e-4ffc-b031-9c9fa5c248e2-metrics-certs\") pod \"frr-k8s-nhqzp\" (UID: \"60b160c7-c63e-4ffc-b031-9c9fa5c248e2\") " pod="metallb-system/frr-k8s-nhqzp" Nov 25 15:14:21 crc kubenswrapper[4890]: I1125 15:14:21.415816 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/742ce93d-d49e-4b12-a1cf-5413ec8475ac-cert\") pod \"frr-k8s-webhook-server-6998585d5-9ncz9\" (UID: \"742ce93d-d49e-4b12-a1cf-5413ec8475ac\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-9ncz9" Nov 25 15:14:21 crc kubenswrapper[4890]: I1125 15:14:21.419080 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/60b160c7-c63e-4ffc-b031-9c9fa5c248e2-metrics-certs\") pod \"frr-k8s-nhqzp\" (UID: \"60b160c7-c63e-4ffc-b031-9c9fa5c248e2\") " pod="metallb-system/frr-k8s-nhqzp" Nov 25 15:14:21 crc kubenswrapper[4890]: I1125 15:14:21.419912 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/742ce93d-d49e-4b12-a1cf-5413ec8475ac-cert\") pod \"frr-k8s-webhook-server-6998585d5-9ncz9\" (UID: \"742ce93d-d49e-4b12-a1cf-5413ec8475ac\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-9ncz9" Nov 25 15:14:21 crc kubenswrapper[4890]: I1125 15:14:21.618691 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d97f8477-4597-4fb0-b2fc-e3faf124ba4e-metrics-certs\") pod \"controller-6c7b4b5f48-hknzk\" (UID: \"d97f8477-4597-4fb0-b2fc-e3faf124ba4e\") " pod="metallb-system/controller-6c7b4b5f48-hknzk" Nov 25 15:14:21 crc kubenswrapper[4890]: I1125 15:14:21.618787 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/9c563abc-f1fc-4c15-9a03-901371600338-memberlist\") pod \"speaker-ndl5n\" (UID: \"9c563abc-f1fc-4c15-9a03-901371600338\") " pod="metallb-system/speaker-ndl5n" Nov 25 15:14:21 crc kubenswrapper[4890]: E1125 15:14:21.618983 4890 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 25 15:14:21 crc kubenswrapper[4890]: E1125 15:14:21.619067 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9c563abc-f1fc-4c15-9a03-901371600338-memberlist podName:9c563abc-f1fc-4c15-9a03-901371600338 nodeName:}" failed. No retries permitted until 2025-11-25 15:14:22.61904779 +0000 UTC m=+721.061510390 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/9c563abc-f1fc-4c15-9a03-901371600338-memberlist") pod "speaker-ndl5n" (UID: "9c563abc-f1fc-4c15-9a03-901371600338") : secret "metallb-memberlist" not found Nov 25 15:14:21 crc kubenswrapper[4890]: I1125 15:14:21.622890 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d97f8477-4597-4fb0-b2fc-e3faf124ba4e-metrics-certs\") pod \"controller-6c7b4b5f48-hknzk\" (UID: \"d97f8477-4597-4fb0-b2fc-e3faf124ba4e\") " pod="metallb-system/controller-6c7b4b5f48-hknzk" Nov 25 15:14:21 crc kubenswrapper[4890]: I1125 15:14:21.695594 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-nhqzp" Nov 25 15:14:21 crc kubenswrapper[4890]: I1125 15:14:21.710149 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-9ncz9" Nov 25 15:14:21 crc kubenswrapper[4890]: I1125 15:14:21.870646 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-hknzk" Nov 25 15:14:21 crc kubenswrapper[4890]: I1125 15:14:21.998813 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-nhqzp" event={"ID":"60b160c7-c63e-4ffc-b031-9c9fa5c248e2","Type":"ContainerStarted","Data":"4e3a795e135a12142f229495d7148eeb62273400c61a252a5ec527ff772c9ed2"} Nov 25 15:14:22 crc kubenswrapper[4890]: I1125 15:14:22.134609 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-9ncz9"] Nov 25 15:14:22 crc kubenswrapper[4890]: W1125 15:14:22.148912 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod742ce93d_d49e_4b12_a1cf_5413ec8475ac.slice/crio-3f91990660ddd93d1bf5d9f561beca04f3af262c66255aa66c5c1c09f039a460 WatchSource:0}: Error finding container 3f91990660ddd93d1bf5d9f561beca04f3af262c66255aa66c5c1c09f039a460: Status 404 returned error can't find the container with id 3f91990660ddd93d1bf5d9f561beca04f3af262c66255aa66c5c1c09f039a460 Nov 25 15:14:22 crc kubenswrapper[4890]: I1125 15:14:22.297972 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-hknzk"] Nov 25 15:14:22 crc kubenswrapper[4890]: W1125 15:14:22.301555 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd97f8477_4597_4fb0_b2fc_e3faf124ba4e.slice/crio-d906a79451303d6164413fb2cef306223415c7d2669814ddbdb88f83e2c38c75 WatchSource:0}: Error finding container d906a79451303d6164413fb2cef306223415c7d2669814ddbdb88f83e2c38c75: Status 404 returned error can't find the container with id d906a79451303d6164413fb2cef306223415c7d2669814ddbdb88f83e2c38c75 Nov 25 15:14:22 crc kubenswrapper[4890]: I1125 15:14:22.632577 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/9c563abc-f1fc-4c15-9a03-901371600338-memberlist\") pod \"speaker-ndl5n\" (UID: \"9c563abc-f1fc-4c15-9a03-901371600338\") " pod="metallb-system/speaker-ndl5n" Nov 25 15:14:22 crc kubenswrapper[4890]: I1125 15:14:22.637991 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/9c563abc-f1fc-4c15-9a03-901371600338-memberlist\") pod \"speaker-ndl5n\" (UID: \"9c563abc-f1fc-4c15-9a03-901371600338\") " pod="metallb-system/speaker-ndl5n" Nov 25 15:14:22 crc kubenswrapper[4890]: I1125 15:14:22.737787 4890 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-k7z2s" Nov 25 15:14:22 crc kubenswrapper[4890]: I1125 15:14:22.747349 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-ndl5n" Nov 25 15:14:23 crc kubenswrapper[4890]: I1125 15:14:23.005634 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-hknzk" event={"ID":"d97f8477-4597-4fb0-b2fc-e3faf124ba4e","Type":"ContainerStarted","Data":"e9ab8b556e31c266995d29b44677bd494002ff88d24c88355af4f0a5bc83f986"} Nov 25 15:14:23 crc kubenswrapper[4890]: I1125 15:14:23.005679 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-hknzk" event={"ID":"d97f8477-4597-4fb0-b2fc-e3faf124ba4e","Type":"ContainerStarted","Data":"6cb850f62e5ae60a4aef7b5de816067c2dc4cbdf7c4d9635cf9bd2c30640ca0a"} Nov 25 15:14:23 crc kubenswrapper[4890]: I1125 15:14:23.005690 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-hknzk" event={"ID":"d97f8477-4597-4fb0-b2fc-e3faf124ba4e","Type":"ContainerStarted","Data":"d906a79451303d6164413fb2cef306223415c7d2669814ddbdb88f83e2c38c75"} Nov 25 15:14:23 crc kubenswrapper[4890]: I1125 15:14:23.005732 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6c7b4b5f48-hknzk" Nov 25 15:14:23 crc kubenswrapper[4890]: I1125 15:14:23.007929 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-9ncz9" event={"ID":"742ce93d-d49e-4b12-a1cf-5413ec8475ac","Type":"ContainerStarted","Data":"3f91990660ddd93d1bf5d9f561beca04f3af262c66255aa66c5c1c09f039a460"} Nov 25 15:14:23 crc kubenswrapper[4890]: I1125 15:14:23.009335 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-ndl5n" event={"ID":"9c563abc-f1fc-4c15-9a03-901371600338","Type":"ContainerStarted","Data":"46067af5001d4360c94304143005db8925d00af5246f128db6cdbef99bbf82f2"} Nov 25 15:14:23 crc kubenswrapper[4890]: I1125 15:14:23.031102 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6c7b4b5f48-hknzk" podStartSLOduration=3.031082688 podStartE2EDuration="3.031082688s" podCreationTimestamp="2025-11-25 15:14:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:14:23.030609116 +0000 UTC m=+721.473071726" watchObservedRunningTime="2025-11-25 15:14:23.031082688 +0000 UTC m=+721.473545288" Nov 25 15:14:24 crc kubenswrapper[4890]: I1125 15:14:24.032813 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-ndl5n" event={"ID":"9c563abc-f1fc-4c15-9a03-901371600338","Type":"ContainerStarted","Data":"e50be2e4187247bef7059941ef64a9d0b0d2e36a9e3d1d311bfb51f29f81eb00"} Nov 25 15:14:24 crc kubenswrapper[4890]: I1125 15:14:24.033222 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-ndl5n" event={"ID":"9c563abc-f1fc-4c15-9a03-901371600338","Type":"ContainerStarted","Data":"f199eefebc80967d98c8fc1e0072f29686521c2847178b2c47a2201631e2a07c"} Nov 25 15:14:24 crc kubenswrapper[4890]: I1125 15:14:24.065355 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-ndl5n" podStartSLOduration=4.065331014 podStartE2EDuration="4.065331014s" podCreationTimestamp="2025-11-25 15:14:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:14:24.059447138 +0000 UTC m=+722.501909748" watchObservedRunningTime="2025-11-25 15:14:24.065331014 +0000 UTC m=+722.507793624" Nov 25 15:14:25 crc kubenswrapper[4890]: I1125 15:14:25.039012 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-ndl5n" Nov 25 15:14:26 crc kubenswrapper[4890]: I1125 15:14:26.447496 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 15:14:26 crc kubenswrapper[4890]: I1125 15:14:26.447573 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 15:14:30 crc kubenswrapper[4890]: I1125 15:14:30.071541 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-9ncz9" event={"ID":"742ce93d-d49e-4b12-a1cf-5413ec8475ac","Type":"ContainerStarted","Data":"7c395dec228d7533aa5863fb8b23fe1d77f562eec1d0c14628a6e51dcc45e909"} Nov 25 15:14:30 crc kubenswrapper[4890]: I1125 15:14:30.072381 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-6998585d5-9ncz9" Nov 25 15:14:30 crc kubenswrapper[4890]: I1125 15:14:30.073630 4890 generic.go:334] "Generic (PLEG): container finished" podID="60b160c7-c63e-4ffc-b031-9c9fa5c248e2" containerID="ddce860ce92e87cc98ec0c75c7b58e3f4109248ce88cdfa28f2243671a897281" exitCode=0 Nov 25 15:14:30 crc kubenswrapper[4890]: I1125 15:14:30.073691 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-nhqzp" event={"ID":"60b160c7-c63e-4ffc-b031-9c9fa5c248e2","Type":"ContainerDied","Data":"ddce860ce92e87cc98ec0c75c7b58e3f4109248ce88cdfa28f2243671a897281"} Nov 25 15:14:30 crc kubenswrapper[4890]: I1125 15:14:30.092594 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-6998585d5-9ncz9" podStartSLOduration=3.141316858 podStartE2EDuration="10.092577085s" podCreationTimestamp="2025-11-25 15:14:20 +0000 UTC" firstStartedPulling="2025-11-25 15:14:22.150965427 +0000 UTC m=+720.593428037" lastFinishedPulling="2025-11-25 15:14:29.102225654 +0000 UTC m=+727.544688264" observedRunningTime="2025-11-25 15:14:30.090284367 +0000 UTC m=+728.532746997" watchObservedRunningTime="2025-11-25 15:14:30.092577085 +0000 UTC m=+728.535039695" Nov 25 15:14:31 crc kubenswrapper[4890]: I1125 15:14:31.086464 4890 generic.go:334] "Generic (PLEG): container finished" podID="60b160c7-c63e-4ffc-b031-9c9fa5c248e2" containerID="291bc70ed3446fe2a0fec8abf2fecb094fb9b320f42a1e96a22f7d119e0d3d79" exitCode=0 Nov 25 15:14:31 crc kubenswrapper[4890]: I1125 15:14:31.086598 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-nhqzp" event={"ID":"60b160c7-c63e-4ffc-b031-9c9fa5c248e2","Type":"ContainerDied","Data":"291bc70ed3446fe2a0fec8abf2fecb094fb9b320f42a1e96a22f7d119e0d3d79"} Nov 25 15:14:32 crc kubenswrapper[4890]: I1125 15:14:32.094637 4890 generic.go:334] "Generic (PLEG): container finished" podID="60b160c7-c63e-4ffc-b031-9c9fa5c248e2" containerID="20c9be998edc674081cb4d6b6b181461d569b2a289054a136aafd48f0ab3e789" exitCode=0 Nov 25 15:14:32 crc kubenswrapper[4890]: I1125 15:14:32.094703 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-nhqzp" event={"ID":"60b160c7-c63e-4ffc-b031-9c9fa5c248e2","Type":"ContainerDied","Data":"20c9be998edc674081cb4d6b6b181461d569b2a289054a136aafd48f0ab3e789"} Nov 25 15:14:32 crc kubenswrapper[4890]: I1125 15:14:32.866981 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-t65nr"] Nov 25 15:14:32 crc kubenswrapper[4890]: I1125 15:14:32.868383 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t65nr" Nov 25 15:14:32 crc kubenswrapper[4890]: I1125 15:14:32.888799 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t65nr"] Nov 25 15:14:32 crc kubenswrapper[4890]: I1125 15:14:32.923939 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03edf3d6-c374-4786-aaf6-7b0dc784ae54-catalog-content\") pod \"redhat-operators-t65nr\" (UID: \"03edf3d6-c374-4786-aaf6-7b0dc784ae54\") " pod="openshift-marketplace/redhat-operators-t65nr" Nov 25 15:14:32 crc kubenswrapper[4890]: I1125 15:14:32.924004 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnvrn\" (UniqueName: \"kubernetes.io/projected/03edf3d6-c374-4786-aaf6-7b0dc784ae54-kube-api-access-nnvrn\") pod \"redhat-operators-t65nr\" (UID: \"03edf3d6-c374-4786-aaf6-7b0dc784ae54\") " pod="openshift-marketplace/redhat-operators-t65nr" Nov 25 15:14:32 crc kubenswrapper[4890]: I1125 15:14:32.924021 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03edf3d6-c374-4786-aaf6-7b0dc784ae54-utilities\") pod \"redhat-operators-t65nr\" (UID: \"03edf3d6-c374-4786-aaf6-7b0dc784ae54\") " pod="openshift-marketplace/redhat-operators-t65nr" Nov 25 15:14:33 crc kubenswrapper[4890]: I1125 15:14:33.024596 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03edf3d6-c374-4786-aaf6-7b0dc784ae54-catalog-content\") pod \"redhat-operators-t65nr\" (UID: \"03edf3d6-c374-4786-aaf6-7b0dc784ae54\") " pod="openshift-marketplace/redhat-operators-t65nr" Nov 25 15:14:33 crc kubenswrapper[4890]: I1125 15:14:33.024706 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnvrn\" (UniqueName: \"kubernetes.io/projected/03edf3d6-c374-4786-aaf6-7b0dc784ae54-kube-api-access-nnvrn\") pod \"redhat-operators-t65nr\" (UID: \"03edf3d6-c374-4786-aaf6-7b0dc784ae54\") " pod="openshift-marketplace/redhat-operators-t65nr" Nov 25 15:14:33 crc kubenswrapper[4890]: I1125 15:14:33.024735 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03edf3d6-c374-4786-aaf6-7b0dc784ae54-utilities\") pod \"redhat-operators-t65nr\" (UID: \"03edf3d6-c374-4786-aaf6-7b0dc784ae54\") " pod="openshift-marketplace/redhat-operators-t65nr" Nov 25 15:14:33 crc kubenswrapper[4890]: I1125 15:14:33.025226 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03edf3d6-c374-4786-aaf6-7b0dc784ae54-catalog-content\") pod \"redhat-operators-t65nr\" (UID: \"03edf3d6-c374-4786-aaf6-7b0dc784ae54\") " pod="openshift-marketplace/redhat-operators-t65nr" Nov 25 15:14:33 crc kubenswrapper[4890]: I1125 15:14:33.025275 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03edf3d6-c374-4786-aaf6-7b0dc784ae54-utilities\") pod \"redhat-operators-t65nr\" (UID: \"03edf3d6-c374-4786-aaf6-7b0dc784ae54\") " pod="openshift-marketplace/redhat-operators-t65nr" Nov 25 15:14:33 crc kubenswrapper[4890]: I1125 15:14:33.046798 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnvrn\" (UniqueName: \"kubernetes.io/projected/03edf3d6-c374-4786-aaf6-7b0dc784ae54-kube-api-access-nnvrn\") pod \"redhat-operators-t65nr\" (UID: \"03edf3d6-c374-4786-aaf6-7b0dc784ae54\") " pod="openshift-marketplace/redhat-operators-t65nr" Nov 25 15:14:33 crc kubenswrapper[4890]: I1125 15:14:33.108446 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-nhqzp" event={"ID":"60b160c7-c63e-4ffc-b031-9c9fa5c248e2","Type":"ContainerStarted","Data":"6bbd15e62c4e71a811e1f25f03ba2a21b58551231b40e5e0ce03488737136e9a"} Nov 25 15:14:33 crc kubenswrapper[4890]: I1125 15:14:33.108486 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-nhqzp" event={"ID":"60b160c7-c63e-4ffc-b031-9c9fa5c248e2","Type":"ContainerStarted","Data":"3fdb5c688bf994649b3c3a7a4f9bfa4a4a8b964efab35bf6517b7fb91fe59104"} Nov 25 15:14:33 crc kubenswrapper[4890]: I1125 15:14:33.108495 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-nhqzp" event={"ID":"60b160c7-c63e-4ffc-b031-9c9fa5c248e2","Type":"ContainerStarted","Data":"ac9cb31c3879aca4b2cc42705dddc6cbabdbe4bbe33aefa212e8e23e84299fcd"} Nov 25 15:14:33 crc kubenswrapper[4890]: I1125 15:14:33.108505 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-nhqzp" event={"ID":"60b160c7-c63e-4ffc-b031-9c9fa5c248e2","Type":"ContainerStarted","Data":"c5e1b0b74f6a0be89f2f39a93821cae219874af6f1b70e7d15ff26979a3cf262"} Nov 25 15:14:33 crc kubenswrapper[4890]: I1125 15:14:33.108513 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-nhqzp" event={"ID":"60b160c7-c63e-4ffc-b031-9c9fa5c248e2","Type":"ContainerStarted","Data":"2b20f18cd8f72afb83e5cc31a2351b9ca433e66fa72b0494b4a9cd0121aa122c"} Nov 25 15:14:33 crc kubenswrapper[4890]: I1125 15:14:33.187843 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t65nr" Nov 25 15:14:33 crc kubenswrapper[4890]: I1125 15:14:33.645013 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t65nr"] Nov 25 15:14:33 crc kubenswrapper[4890]: W1125 15:14:33.653980 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod03edf3d6_c374_4786_aaf6_7b0dc784ae54.slice/crio-28c782fb84ce223dd8596f146cef6ba55a3adb45c74e78fcea835e244524ea68 WatchSource:0}: Error finding container 28c782fb84ce223dd8596f146cef6ba55a3adb45c74e78fcea835e244524ea68: Status 404 returned error can't find the container with id 28c782fb84ce223dd8596f146cef6ba55a3adb45c74e78fcea835e244524ea68 Nov 25 15:14:34 crc kubenswrapper[4890]: I1125 15:14:34.119610 4890 generic.go:334] "Generic (PLEG): container finished" podID="03edf3d6-c374-4786-aaf6-7b0dc784ae54" containerID="4a3aa834ec87abb4b22a47a3dfcddf47fd8ad6e997cdf705d70da73e18e1e441" exitCode=0 Nov 25 15:14:34 crc kubenswrapper[4890]: I1125 15:14:34.119699 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t65nr" event={"ID":"03edf3d6-c374-4786-aaf6-7b0dc784ae54","Type":"ContainerDied","Data":"4a3aa834ec87abb4b22a47a3dfcddf47fd8ad6e997cdf705d70da73e18e1e441"} Nov 25 15:14:34 crc kubenswrapper[4890]: I1125 15:14:34.119777 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t65nr" event={"ID":"03edf3d6-c374-4786-aaf6-7b0dc784ae54","Type":"ContainerStarted","Data":"28c782fb84ce223dd8596f146cef6ba55a3adb45c74e78fcea835e244524ea68"} Nov 25 15:14:34 crc kubenswrapper[4890]: I1125 15:14:34.124829 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-nhqzp" event={"ID":"60b160c7-c63e-4ffc-b031-9c9fa5c248e2","Type":"ContainerStarted","Data":"0fe505569c56a9e996cfe6b8f085ab1605fedab0596eafc8e12141f6f11ee60b"} Nov 25 15:14:34 crc kubenswrapper[4890]: I1125 15:14:34.125061 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-nhqzp" Nov 25 15:14:34 crc kubenswrapper[4890]: I1125 15:14:34.175815 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-nhqzp" podStartSLOduration=6.926881077 podStartE2EDuration="14.175784048s" podCreationTimestamp="2025-11-25 15:14:20 +0000 UTC" firstStartedPulling="2025-11-25 15:14:21.834093484 +0000 UTC m=+720.276556094" lastFinishedPulling="2025-11-25 15:14:29.082996455 +0000 UTC m=+727.525459065" observedRunningTime="2025-11-25 15:14:34.175305156 +0000 UTC m=+732.617767766" watchObservedRunningTime="2025-11-25 15:14:34.175784048 +0000 UTC m=+732.618246658" Nov 25 15:14:35 crc kubenswrapper[4890]: I1125 15:14:35.133256 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t65nr" event={"ID":"03edf3d6-c374-4786-aaf6-7b0dc784ae54","Type":"ContainerStarted","Data":"ef7003b2e2354f20c7c53ab49307125dacf3ab7c1dad6343274901045094ff2f"} Nov 25 15:14:36 crc kubenswrapper[4890]: I1125 15:14:36.141911 4890 generic.go:334] "Generic (PLEG): container finished" podID="03edf3d6-c374-4786-aaf6-7b0dc784ae54" containerID="ef7003b2e2354f20c7c53ab49307125dacf3ab7c1dad6343274901045094ff2f" exitCode=0 Nov 25 15:14:36 crc kubenswrapper[4890]: I1125 15:14:36.141972 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t65nr" event={"ID":"03edf3d6-c374-4786-aaf6-7b0dc784ae54","Type":"ContainerDied","Data":"ef7003b2e2354f20c7c53ab49307125dacf3ab7c1dad6343274901045094ff2f"} Nov 25 15:14:36 crc kubenswrapper[4890]: I1125 15:14:36.695841 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-nhqzp" Nov 25 15:14:36 crc kubenswrapper[4890]: I1125 15:14:36.746467 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-nhqzp" Nov 25 15:14:37 crc kubenswrapper[4890]: I1125 15:14:37.161963 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t65nr" event={"ID":"03edf3d6-c374-4786-aaf6-7b0dc784ae54","Type":"ContainerStarted","Data":"e24781d0dab8145a235b05cb6a33ac3c958daac1299bfdedb7c2c620b766e4f9"} Nov 25 15:14:37 crc kubenswrapper[4890]: I1125 15:14:37.185561 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-t65nr" podStartSLOduration=2.5886943650000003 podStartE2EDuration="5.185538276s" podCreationTimestamp="2025-11-25 15:14:32 +0000 UTC" firstStartedPulling="2025-11-25 15:14:34.121267268 +0000 UTC m=+732.563729878" lastFinishedPulling="2025-11-25 15:14:36.718111179 +0000 UTC m=+735.160573789" observedRunningTime="2025-11-25 15:14:37.181402283 +0000 UTC m=+735.623864903" watchObservedRunningTime="2025-11-25 15:14:37.185538276 +0000 UTC m=+735.628000886" Nov 25 15:14:41 crc kubenswrapper[4890]: I1125 15:14:41.716666 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-6998585d5-9ncz9" Nov 25 15:14:41 crc kubenswrapper[4890]: I1125 15:14:41.877585 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6c7b4b5f48-hknzk" Nov 25 15:14:42 crc kubenswrapper[4890]: I1125 15:14:42.755682 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-ndl5n" Nov 25 15:14:43 crc kubenswrapper[4890]: I1125 15:14:43.189024 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-t65nr" Nov 25 15:14:43 crc kubenswrapper[4890]: I1125 15:14:43.189444 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-t65nr" Nov 25 15:14:43 crc kubenswrapper[4890]: I1125 15:14:43.243992 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-t65nr" Nov 25 15:14:44 crc kubenswrapper[4890]: I1125 15:14:44.244376 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-t65nr" Nov 25 15:14:44 crc kubenswrapper[4890]: I1125 15:14:44.298831 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t65nr"] Nov 25 15:14:45 crc kubenswrapper[4890]: I1125 15:14:45.689685 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-vdc6n"] Nov 25 15:14:45 crc kubenswrapper[4890]: I1125 15:14:45.690644 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-vdc6n" Nov 25 15:14:45 crc kubenswrapper[4890]: I1125 15:14:45.693928 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-94xds" Nov 25 15:14:45 crc kubenswrapper[4890]: I1125 15:14:45.694379 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 25 15:14:45 crc kubenswrapper[4890]: I1125 15:14:45.694686 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 25 15:14:45 crc kubenswrapper[4890]: I1125 15:14:45.705451 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-vdc6n"] Nov 25 15:14:45 crc kubenswrapper[4890]: I1125 15:14:45.808740 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hj7g\" (UniqueName: \"kubernetes.io/projected/7ada5ebc-0ba5-482a-b0af-267a544a844c-kube-api-access-2hj7g\") pod \"openstack-operator-index-vdc6n\" (UID: \"7ada5ebc-0ba5-482a-b0af-267a544a844c\") " pod="openstack-operators/openstack-operator-index-vdc6n" Nov 25 15:14:45 crc kubenswrapper[4890]: I1125 15:14:45.910600 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hj7g\" (UniqueName: \"kubernetes.io/projected/7ada5ebc-0ba5-482a-b0af-267a544a844c-kube-api-access-2hj7g\") pod \"openstack-operator-index-vdc6n\" (UID: \"7ada5ebc-0ba5-482a-b0af-267a544a844c\") " pod="openstack-operators/openstack-operator-index-vdc6n" Nov 25 15:14:45 crc kubenswrapper[4890]: I1125 15:14:45.930982 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hj7g\" (UniqueName: \"kubernetes.io/projected/7ada5ebc-0ba5-482a-b0af-267a544a844c-kube-api-access-2hj7g\") pod \"openstack-operator-index-vdc6n\" (UID: \"7ada5ebc-0ba5-482a-b0af-267a544a844c\") " pod="openstack-operators/openstack-operator-index-vdc6n" Nov 25 15:14:46 crc kubenswrapper[4890]: I1125 15:14:46.020911 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-vdc6n" Nov 25 15:14:46 crc kubenswrapper[4890]: I1125 15:14:46.219352 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-t65nr" podUID="03edf3d6-c374-4786-aaf6-7b0dc784ae54" containerName="registry-server" containerID="cri-o://e24781d0dab8145a235b05cb6a33ac3c958daac1299bfdedb7c2c620b766e4f9" gracePeriod=2 Nov 25 15:14:46 crc kubenswrapper[4890]: I1125 15:14:46.440377 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-vdc6n"] Nov 25 15:14:46 crc kubenswrapper[4890]: W1125 15:14:46.445128 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7ada5ebc_0ba5_482a_b0af_267a544a844c.slice/crio-440d6a525de541665c57d28c90ecf4fc64f0ab844f38513f944f8fe450731910 WatchSource:0}: Error finding container 440d6a525de541665c57d28c90ecf4fc64f0ab844f38513f944f8fe450731910: Status 404 returned error can't find the container with id 440d6a525de541665c57d28c90ecf4fc64f0ab844f38513f944f8fe450731910 Nov 25 15:14:47 crc kubenswrapper[4890]: I1125 15:14:47.160816 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t65nr" Nov 25 15:14:47 crc kubenswrapper[4890]: I1125 15:14:47.227016 4890 generic.go:334] "Generic (PLEG): container finished" podID="03edf3d6-c374-4786-aaf6-7b0dc784ae54" containerID="e24781d0dab8145a235b05cb6a33ac3c958daac1299bfdedb7c2c620b766e4f9" exitCode=0 Nov 25 15:14:47 crc kubenswrapper[4890]: I1125 15:14:47.227080 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t65nr" Nov 25 15:14:47 crc kubenswrapper[4890]: I1125 15:14:47.227093 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t65nr" event={"ID":"03edf3d6-c374-4786-aaf6-7b0dc784ae54","Type":"ContainerDied","Data":"e24781d0dab8145a235b05cb6a33ac3c958daac1299bfdedb7c2c620b766e4f9"} Nov 25 15:14:47 crc kubenswrapper[4890]: I1125 15:14:47.227224 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t65nr" event={"ID":"03edf3d6-c374-4786-aaf6-7b0dc784ae54","Type":"ContainerDied","Data":"28c782fb84ce223dd8596f146cef6ba55a3adb45c74e78fcea835e244524ea68"} Nov 25 15:14:47 crc kubenswrapper[4890]: I1125 15:14:47.227253 4890 scope.go:117] "RemoveContainer" containerID="e24781d0dab8145a235b05cb6a33ac3c958daac1299bfdedb7c2c620b766e4f9" Nov 25 15:14:47 crc kubenswrapper[4890]: I1125 15:14:47.229866 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-vdc6n" event={"ID":"7ada5ebc-0ba5-482a-b0af-267a544a844c","Type":"ContainerStarted","Data":"440d6a525de541665c57d28c90ecf4fc64f0ab844f38513f944f8fe450731910"} Nov 25 15:14:47 crc kubenswrapper[4890]: I1125 15:14:47.335287 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03edf3d6-c374-4786-aaf6-7b0dc784ae54-catalog-content\") pod \"03edf3d6-c374-4786-aaf6-7b0dc784ae54\" (UID: \"03edf3d6-c374-4786-aaf6-7b0dc784ae54\") " Nov 25 15:14:47 crc kubenswrapper[4890]: I1125 15:14:47.335407 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nnvrn\" (UniqueName: \"kubernetes.io/projected/03edf3d6-c374-4786-aaf6-7b0dc784ae54-kube-api-access-nnvrn\") pod \"03edf3d6-c374-4786-aaf6-7b0dc784ae54\" (UID: \"03edf3d6-c374-4786-aaf6-7b0dc784ae54\") " Nov 25 15:14:47 crc kubenswrapper[4890]: I1125 15:14:47.335486 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03edf3d6-c374-4786-aaf6-7b0dc784ae54-utilities\") pod \"03edf3d6-c374-4786-aaf6-7b0dc784ae54\" (UID: \"03edf3d6-c374-4786-aaf6-7b0dc784ae54\") " Nov 25 15:14:47 crc kubenswrapper[4890]: I1125 15:14:47.336992 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03edf3d6-c374-4786-aaf6-7b0dc784ae54-utilities" (OuterVolumeSpecName: "utilities") pod "03edf3d6-c374-4786-aaf6-7b0dc784ae54" (UID: "03edf3d6-c374-4786-aaf6-7b0dc784ae54"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:14:47 crc kubenswrapper[4890]: I1125 15:14:47.347036 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03edf3d6-c374-4786-aaf6-7b0dc784ae54-kube-api-access-nnvrn" (OuterVolumeSpecName: "kube-api-access-nnvrn") pod "03edf3d6-c374-4786-aaf6-7b0dc784ae54" (UID: "03edf3d6-c374-4786-aaf6-7b0dc784ae54"). InnerVolumeSpecName "kube-api-access-nnvrn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:14:47 crc kubenswrapper[4890]: I1125 15:14:47.429623 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03edf3d6-c374-4786-aaf6-7b0dc784ae54-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "03edf3d6-c374-4786-aaf6-7b0dc784ae54" (UID: "03edf3d6-c374-4786-aaf6-7b0dc784ae54"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:14:47 crc kubenswrapper[4890]: I1125 15:14:47.437218 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nnvrn\" (UniqueName: \"kubernetes.io/projected/03edf3d6-c374-4786-aaf6-7b0dc784ae54-kube-api-access-nnvrn\") on node \"crc\" DevicePath \"\"" Nov 25 15:14:47 crc kubenswrapper[4890]: I1125 15:14:47.437259 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03edf3d6-c374-4786-aaf6-7b0dc784ae54-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 15:14:47 crc kubenswrapper[4890]: I1125 15:14:47.437271 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03edf3d6-c374-4786-aaf6-7b0dc784ae54-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 15:14:47 crc kubenswrapper[4890]: I1125 15:14:47.561956 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t65nr"] Nov 25 15:14:47 crc kubenswrapper[4890]: I1125 15:14:47.565653 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-t65nr"] Nov 25 15:14:47 crc kubenswrapper[4890]: I1125 15:14:47.595687 4890 scope.go:117] "RemoveContainer" containerID="ef7003b2e2354f20c7c53ab49307125dacf3ab7c1dad6343274901045094ff2f" Nov 25 15:14:47 crc kubenswrapper[4890]: I1125 15:14:47.789836 4890 scope.go:117] "RemoveContainer" containerID="4a3aa834ec87abb4b22a47a3dfcddf47fd8ad6e997cdf705d70da73e18e1e441" Nov 25 15:14:48 crc kubenswrapper[4890]: I1125 15:14:48.043610 4890 scope.go:117] "RemoveContainer" containerID="e24781d0dab8145a235b05cb6a33ac3c958daac1299bfdedb7c2c620b766e4f9" Nov 25 15:14:48 crc kubenswrapper[4890]: E1125 15:14:48.044336 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e24781d0dab8145a235b05cb6a33ac3c958daac1299bfdedb7c2c620b766e4f9\": container with ID starting with e24781d0dab8145a235b05cb6a33ac3c958daac1299bfdedb7c2c620b766e4f9 not found: ID does not exist" containerID="e24781d0dab8145a235b05cb6a33ac3c958daac1299bfdedb7c2c620b766e4f9" Nov 25 15:14:48 crc kubenswrapper[4890]: I1125 15:14:48.044413 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e24781d0dab8145a235b05cb6a33ac3c958daac1299bfdedb7c2c620b766e4f9"} err="failed to get container status \"e24781d0dab8145a235b05cb6a33ac3c958daac1299bfdedb7c2c620b766e4f9\": rpc error: code = NotFound desc = could not find container \"e24781d0dab8145a235b05cb6a33ac3c958daac1299bfdedb7c2c620b766e4f9\": container with ID starting with e24781d0dab8145a235b05cb6a33ac3c958daac1299bfdedb7c2c620b766e4f9 not found: ID does not exist" Nov 25 15:14:48 crc kubenswrapper[4890]: I1125 15:14:48.044444 4890 scope.go:117] "RemoveContainer" containerID="ef7003b2e2354f20c7c53ab49307125dacf3ab7c1dad6343274901045094ff2f" Nov 25 15:14:48 crc kubenswrapper[4890]: E1125 15:14:48.045057 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef7003b2e2354f20c7c53ab49307125dacf3ab7c1dad6343274901045094ff2f\": container with ID starting with ef7003b2e2354f20c7c53ab49307125dacf3ab7c1dad6343274901045094ff2f not found: ID does not exist" containerID="ef7003b2e2354f20c7c53ab49307125dacf3ab7c1dad6343274901045094ff2f" Nov 25 15:14:48 crc kubenswrapper[4890]: I1125 15:14:48.045104 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef7003b2e2354f20c7c53ab49307125dacf3ab7c1dad6343274901045094ff2f"} err="failed to get container status \"ef7003b2e2354f20c7c53ab49307125dacf3ab7c1dad6343274901045094ff2f\": rpc error: code = NotFound desc = could not find container \"ef7003b2e2354f20c7c53ab49307125dacf3ab7c1dad6343274901045094ff2f\": container with ID starting with ef7003b2e2354f20c7c53ab49307125dacf3ab7c1dad6343274901045094ff2f not found: ID does not exist" Nov 25 15:14:48 crc kubenswrapper[4890]: I1125 15:14:48.045132 4890 scope.go:117] "RemoveContainer" containerID="4a3aa834ec87abb4b22a47a3dfcddf47fd8ad6e997cdf705d70da73e18e1e441" Nov 25 15:14:48 crc kubenswrapper[4890]: E1125 15:14:48.045587 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a3aa834ec87abb4b22a47a3dfcddf47fd8ad6e997cdf705d70da73e18e1e441\": container with ID starting with 4a3aa834ec87abb4b22a47a3dfcddf47fd8ad6e997cdf705d70da73e18e1e441 not found: ID does not exist" containerID="4a3aa834ec87abb4b22a47a3dfcddf47fd8ad6e997cdf705d70da73e18e1e441" Nov 25 15:14:48 crc kubenswrapper[4890]: I1125 15:14:48.045664 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a3aa834ec87abb4b22a47a3dfcddf47fd8ad6e997cdf705d70da73e18e1e441"} err="failed to get container status \"4a3aa834ec87abb4b22a47a3dfcddf47fd8ad6e997cdf705d70da73e18e1e441\": rpc error: code = NotFound desc = could not find container \"4a3aa834ec87abb4b22a47a3dfcddf47fd8ad6e997cdf705d70da73e18e1e441\": container with ID starting with 4a3aa834ec87abb4b22a47a3dfcddf47fd8ad6e997cdf705d70da73e18e1e441 not found: ID does not exist" Nov 25 15:14:48 crc kubenswrapper[4890]: I1125 15:14:48.183343 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03edf3d6-c374-4786-aaf6-7b0dc784ae54" path="/var/lib/kubelet/pods/03edf3d6-c374-4786-aaf6-7b0dc784ae54/volumes" Nov 25 15:14:49 crc kubenswrapper[4890]: I1125 15:14:49.246487 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-vdc6n" event={"ID":"7ada5ebc-0ba5-482a-b0af-267a544a844c","Type":"ContainerStarted","Data":"f498eac41483d3893aabdf1e09ceab569282d0ebe416129ac923351d12f99cdb"} Nov 25 15:14:50 crc kubenswrapper[4890]: I1125 15:14:50.878372 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-vdc6n" podStartSLOduration=3.321768628 podStartE2EDuration="5.878338323s" podCreationTimestamp="2025-11-25 15:14:45 +0000 UTC" firstStartedPulling="2025-11-25 15:14:46.44751344 +0000 UTC m=+744.889976050" lastFinishedPulling="2025-11-25 15:14:49.004083135 +0000 UTC m=+747.446545745" observedRunningTime="2025-11-25 15:14:49.271756622 +0000 UTC m=+747.714219292" watchObservedRunningTime="2025-11-25 15:14:50.878338323 +0000 UTC m=+749.320800963" Nov 25 15:14:50 crc kubenswrapper[4890]: I1125 15:14:50.891274 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-vdc6n"] Nov 25 15:14:51 crc kubenswrapper[4890]: I1125 15:14:51.260927 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-vdc6n" podUID="7ada5ebc-0ba5-482a-b0af-267a544a844c" containerName="registry-server" containerID="cri-o://f498eac41483d3893aabdf1e09ceab569282d0ebe416129ac923351d12f99cdb" gracePeriod=2 Nov 25 15:14:51 crc kubenswrapper[4890]: E1125 15:14:51.389536 4890 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7ada5ebc_0ba5_482a_b0af_267a544a844c.slice/crio-f498eac41483d3893aabdf1e09ceab569282d0ebe416129ac923351d12f99cdb.scope\": RecentStats: unable to find data in memory cache]" Nov 25 15:14:51 crc kubenswrapper[4890]: I1125 15:14:51.685678 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-vtg6p"] Nov 25 15:14:51 crc kubenswrapper[4890]: E1125 15:14:51.686393 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03edf3d6-c374-4786-aaf6-7b0dc784ae54" containerName="extract-content" Nov 25 15:14:51 crc kubenswrapper[4890]: I1125 15:14:51.686413 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="03edf3d6-c374-4786-aaf6-7b0dc784ae54" containerName="extract-content" Nov 25 15:14:51 crc kubenswrapper[4890]: E1125 15:14:51.686438 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03edf3d6-c374-4786-aaf6-7b0dc784ae54" containerName="extract-utilities" Nov 25 15:14:51 crc kubenswrapper[4890]: I1125 15:14:51.686448 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="03edf3d6-c374-4786-aaf6-7b0dc784ae54" containerName="extract-utilities" Nov 25 15:14:51 crc kubenswrapper[4890]: E1125 15:14:51.686465 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03edf3d6-c374-4786-aaf6-7b0dc784ae54" containerName="registry-server" Nov 25 15:14:51 crc kubenswrapper[4890]: I1125 15:14:51.686474 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="03edf3d6-c374-4786-aaf6-7b0dc784ae54" containerName="registry-server" Nov 25 15:14:51 crc kubenswrapper[4890]: I1125 15:14:51.686639 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="03edf3d6-c374-4786-aaf6-7b0dc784ae54" containerName="registry-server" Nov 25 15:14:51 crc kubenswrapper[4890]: I1125 15:14:51.687294 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-vtg6p" Nov 25 15:14:51 crc kubenswrapper[4890]: I1125 15:14:51.699410 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-vtg6p"] Nov 25 15:14:51 crc kubenswrapper[4890]: I1125 15:14:51.699940 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-nhqzp" Nov 25 15:14:51 crc kubenswrapper[4890]: I1125 15:14:51.730604 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-vdc6n" Nov 25 15:14:51 crc kubenswrapper[4890]: I1125 15:14:51.799664 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpqdd\" (UniqueName: \"kubernetes.io/projected/bd47d0df-89c3-465a-8f11-8f0e7e1117eb-kube-api-access-qpqdd\") pod \"openstack-operator-index-vtg6p\" (UID: \"bd47d0df-89c3-465a-8f11-8f0e7e1117eb\") " pod="openstack-operators/openstack-operator-index-vtg6p" Nov 25 15:14:51 crc kubenswrapper[4890]: I1125 15:14:51.901531 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2hj7g\" (UniqueName: \"kubernetes.io/projected/7ada5ebc-0ba5-482a-b0af-267a544a844c-kube-api-access-2hj7g\") pod \"7ada5ebc-0ba5-482a-b0af-267a544a844c\" (UID: \"7ada5ebc-0ba5-482a-b0af-267a544a844c\") " Nov 25 15:14:51 crc kubenswrapper[4890]: I1125 15:14:51.901863 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpqdd\" (UniqueName: \"kubernetes.io/projected/bd47d0df-89c3-465a-8f11-8f0e7e1117eb-kube-api-access-qpqdd\") pod \"openstack-operator-index-vtg6p\" (UID: \"bd47d0df-89c3-465a-8f11-8f0e7e1117eb\") " pod="openstack-operators/openstack-operator-index-vtg6p" Nov 25 15:14:51 crc kubenswrapper[4890]: I1125 15:14:51.907416 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ada5ebc-0ba5-482a-b0af-267a544a844c-kube-api-access-2hj7g" (OuterVolumeSpecName: "kube-api-access-2hj7g") pod "7ada5ebc-0ba5-482a-b0af-267a544a844c" (UID: "7ada5ebc-0ba5-482a-b0af-267a544a844c"). InnerVolumeSpecName "kube-api-access-2hj7g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:14:51 crc kubenswrapper[4890]: I1125 15:14:51.918061 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpqdd\" (UniqueName: \"kubernetes.io/projected/bd47d0df-89c3-465a-8f11-8f0e7e1117eb-kube-api-access-qpqdd\") pod \"openstack-operator-index-vtg6p\" (UID: \"bd47d0df-89c3-465a-8f11-8f0e7e1117eb\") " pod="openstack-operators/openstack-operator-index-vtg6p" Nov 25 15:14:52 crc kubenswrapper[4890]: I1125 15:14:52.003139 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2hj7g\" (UniqueName: \"kubernetes.io/projected/7ada5ebc-0ba5-482a-b0af-267a544a844c-kube-api-access-2hj7g\") on node \"crc\" DevicePath \"\"" Nov 25 15:14:52 crc kubenswrapper[4890]: I1125 15:14:52.048763 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-vtg6p" Nov 25 15:14:52 crc kubenswrapper[4890]: I1125 15:14:52.269930 4890 generic.go:334] "Generic (PLEG): container finished" podID="7ada5ebc-0ba5-482a-b0af-267a544a844c" containerID="f498eac41483d3893aabdf1e09ceab569282d0ebe416129ac923351d12f99cdb" exitCode=0 Nov 25 15:14:52 crc kubenswrapper[4890]: I1125 15:14:52.269989 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-vdc6n" event={"ID":"7ada5ebc-0ba5-482a-b0af-267a544a844c","Type":"ContainerDied","Data":"f498eac41483d3893aabdf1e09ceab569282d0ebe416129ac923351d12f99cdb"} Nov 25 15:14:52 crc kubenswrapper[4890]: I1125 15:14:52.270051 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-vdc6n" event={"ID":"7ada5ebc-0ba5-482a-b0af-267a544a844c","Type":"ContainerDied","Data":"440d6a525de541665c57d28c90ecf4fc64f0ab844f38513f944f8fe450731910"} Nov 25 15:14:52 crc kubenswrapper[4890]: I1125 15:14:52.270081 4890 scope.go:117] "RemoveContainer" containerID="f498eac41483d3893aabdf1e09ceab569282d0ebe416129ac923351d12f99cdb" Nov 25 15:14:52 crc kubenswrapper[4890]: I1125 15:14:52.269997 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-vdc6n" Nov 25 15:14:52 crc kubenswrapper[4890]: I1125 15:14:52.291877 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-vdc6n"] Nov 25 15:14:52 crc kubenswrapper[4890]: I1125 15:14:52.296736 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-vdc6n"] Nov 25 15:14:52 crc kubenswrapper[4890]: I1125 15:14:52.299413 4890 scope.go:117] "RemoveContainer" containerID="f498eac41483d3893aabdf1e09ceab569282d0ebe416129ac923351d12f99cdb" Nov 25 15:14:52 crc kubenswrapper[4890]: E1125 15:14:52.300432 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f498eac41483d3893aabdf1e09ceab569282d0ebe416129ac923351d12f99cdb\": container with ID starting with f498eac41483d3893aabdf1e09ceab569282d0ebe416129ac923351d12f99cdb not found: ID does not exist" containerID="f498eac41483d3893aabdf1e09ceab569282d0ebe416129ac923351d12f99cdb" Nov 25 15:14:52 crc kubenswrapper[4890]: I1125 15:14:52.300490 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f498eac41483d3893aabdf1e09ceab569282d0ebe416129ac923351d12f99cdb"} err="failed to get container status \"f498eac41483d3893aabdf1e09ceab569282d0ebe416129ac923351d12f99cdb\": rpc error: code = NotFound desc = could not find container \"f498eac41483d3893aabdf1e09ceab569282d0ebe416129ac923351d12f99cdb\": container with ID starting with f498eac41483d3893aabdf1e09ceab569282d0ebe416129ac923351d12f99cdb not found: ID does not exist" Nov 25 15:14:52 crc kubenswrapper[4890]: I1125 15:14:52.449684 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-vtg6p"] Nov 25 15:14:52 crc kubenswrapper[4890]: W1125 15:14:52.458445 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbd47d0df_89c3_465a_8f11_8f0e7e1117eb.slice/crio-a46c15e96f012c37c814f9e58a3cb9675ac4d207ea94bf4cec7b09c05e68fe98 WatchSource:0}: Error finding container a46c15e96f012c37c814f9e58a3cb9675ac4d207ea94bf4cec7b09c05e68fe98: Status 404 returned error can't find the container with id a46c15e96f012c37c814f9e58a3cb9675ac4d207ea94bf4cec7b09c05e68fe98 Nov 25 15:14:53 crc kubenswrapper[4890]: I1125 15:14:53.279856 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-vtg6p" event={"ID":"bd47d0df-89c3-465a-8f11-8f0e7e1117eb","Type":"ContainerStarted","Data":"ac3abe6247453ce1123994ccc1ae89cc1dc0a83e44e68626431278b99fd81832"} Nov 25 15:14:53 crc kubenswrapper[4890]: I1125 15:14:53.279905 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-vtg6p" event={"ID":"bd47d0df-89c3-465a-8f11-8f0e7e1117eb","Type":"ContainerStarted","Data":"a46c15e96f012c37c814f9e58a3cb9675ac4d207ea94bf4cec7b09c05e68fe98"} Nov 25 15:14:53 crc kubenswrapper[4890]: I1125 15:14:53.302520 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-vtg6p" podStartSLOduration=2.252017097 podStartE2EDuration="2.302493316s" podCreationTimestamp="2025-11-25 15:14:51 +0000 UTC" firstStartedPulling="2025-11-25 15:14:52.461795237 +0000 UTC m=+750.904257847" lastFinishedPulling="2025-11-25 15:14:52.512271416 +0000 UTC m=+750.954734066" observedRunningTime="2025-11-25 15:14:53.300985248 +0000 UTC m=+751.743447868" watchObservedRunningTime="2025-11-25 15:14:53.302493316 +0000 UTC m=+751.744955936" Nov 25 15:14:54 crc kubenswrapper[4890]: I1125 15:14:54.186449 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ada5ebc-0ba5-482a-b0af-267a544a844c" path="/var/lib/kubelet/pods/7ada5ebc-0ba5-482a-b0af-267a544a844c/volumes" Nov 25 15:14:56 crc kubenswrapper[4890]: I1125 15:14:56.447683 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 15:14:56 crc kubenswrapper[4890]: I1125 15:14:56.448097 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 15:14:56 crc kubenswrapper[4890]: I1125 15:14:56.448153 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" Nov 25 15:14:56 crc kubenswrapper[4890]: I1125 15:14:56.448859 4890 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"361ef05d7f3bcf13e88a1632cd64dd7b6044add1c3fc4ac44bf729fc819bec71"} pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 15:14:56 crc kubenswrapper[4890]: I1125 15:14:56.448980 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" containerID="cri-o://361ef05d7f3bcf13e88a1632cd64dd7b6044add1c3fc4ac44bf729fc819bec71" gracePeriod=600 Nov 25 15:14:57 crc kubenswrapper[4890]: I1125 15:14:57.309178 4890 generic.go:334] "Generic (PLEG): container finished" podID="4e4f849d-f239-4727-a73e-18327856929a" containerID="361ef05d7f3bcf13e88a1632cd64dd7b6044add1c3fc4ac44bf729fc819bec71" exitCode=0 Nov 25 15:14:57 crc kubenswrapper[4890]: I1125 15:14:57.309201 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" event={"ID":"4e4f849d-f239-4727-a73e-18327856929a","Type":"ContainerDied","Data":"361ef05d7f3bcf13e88a1632cd64dd7b6044add1c3fc4ac44bf729fc819bec71"} Nov 25 15:14:57 crc kubenswrapper[4890]: I1125 15:14:57.309595 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" event={"ID":"4e4f849d-f239-4727-a73e-18327856929a","Type":"ContainerStarted","Data":"b51c3e1b3270ec2e9b5e884e9f00a4274f242cf4d6c3240274a36d149cb2249e"} Nov 25 15:14:57 crc kubenswrapper[4890]: I1125 15:14:57.309693 4890 scope.go:117] "RemoveContainer" containerID="1a14507b222e3efb8d4e5a5a36210ae167e2bc170367e62ff6f845c2d500391c" Nov 25 15:15:00 crc kubenswrapper[4890]: I1125 15:15:00.137036 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401395-hslqf"] Nov 25 15:15:00 crc kubenswrapper[4890]: E1125 15:15:00.137863 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ada5ebc-0ba5-482a-b0af-267a544a844c" containerName="registry-server" Nov 25 15:15:00 crc kubenswrapper[4890]: I1125 15:15:00.137879 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ada5ebc-0ba5-482a-b0af-267a544a844c" containerName="registry-server" Nov 25 15:15:00 crc kubenswrapper[4890]: I1125 15:15:00.138023 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ada5ebc-0ba5-482a-b0af-267a544a844c" containerName="registry-server" Nov 25 15:15:00 crc kubenswrapper[4890]: I1125 15:15:00.138565 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401395-hslqf" Nov 25 15:15:00 crc kubenswrapper[4890]: I1125 15:15:00.142694 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 15:15:00 crc kubenswrapper[4890]: I1125 15:15:00.143584 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 15:15:00 crc kubenswrapper[4890]: I1125 15:15:00.155206 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401395-hslqf"] Nov 25 15:15:00 crc kubenswrapper[4890]: I1125 15:15:00.222881 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6pm2\" (UniqueName: \"kubernetes.io/projected/ed35cb6c-f2b1-4d8a-9b43-782ca62594c7-kube-api-access-s6pm2\") pod \"collect-profiles-29401395-hslqf\" (UID: \"ed35cb6c-f2b1-4d8a-9b43-782ca62594c7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401395-hslqf" Nov 25 15:15:00 crc kubenswrapper[4890]: I1125 15:15:00.223353 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ed35cb6c-f2b1-4d8a-9b43-782ca62594c7-secret-volume\") pod \"collect-profiles-29401395-hslqf\" (UID: \"ed35cb6c-f2b1-4d8a-9b43-782ca62594c7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401395-hslqf" Nov 25 15:15:00 crc kubenswrapper[4890]: I1125 15:15:00.223550 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ed35cb6c-f2b1-4d8a-9b43-782ca62594c7-config-volume\") pod \"collect-profiles-29401395-hslqf\" (UID: \"ed35cb6c-f2b1-4d8a-9b43-782ca62594c7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401395-hslqf" Nov 25 15:15:00 crc kubenswrapper[4890]: I1125 15:15:00.324857 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ed35cb6c-f2b1-4d8a-9b43-782ca62594c7-config-volume\") pod \"collect-profiles-29401395-hslqf\" (UID: \"ed35cb6c-f2b1-4d8a-9b43-782ca62594c7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401395-hslqf" Nov 25 15:15:00 crc kubenswrapper[4890]: I1125 15:15:00.324995 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6pm2\" (UniqueName: \"kubernetes.io/projected/ed35cb6c-f2b1-4d8a-9b43-782ca62594c7-kube-api-access-s6pm2\") pod \"collect-profiles-29401395-hslqf\" (UID: \"ed35cb6c-f2b1-4d8a-9b43-782ca62594c7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401395-hslqf" Nov 25 15:15:00 crc kubenswrapper[4890]: I1125 15:15:00.325069 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ed35cb6c-f2b1-4d8a-9b43-782ca62594c7-secret-volume\") pod \"collect-profiles-29401395-hslqf\" (UID: \"ed35cb6c-f2b1-4d8a-9b43-782ca62594c7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401395-hslqf" Nov 25 15:15:00 crc kubenswrapper[4890]: I1125 15:15:00.326424 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ed35cb6c-f2b1-4d8a-9b43-782ca62594c7-config-volume\") pod \"collect-profiles-29401395-hslqf\" (UID: \"ed35cb6c-f2b1-4d8a-9b43-782ca62594c7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401395-hslqf" Nov 25 15:15:00 crc kubenswrapper[4890]: I1125 15:15:00.333028 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ed35cb6c-f2b1-4d8a-9b43-782ca62594c7-secret-volume\") pod \"collect-profiles-29401395-hslqf\" (UID: \"ed35cb6c-f2b1-4d8a-9b43-782ca62594c7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401395-hslqf" Nov 25 15:15:00 crc kubenswrapper[4890]: I1125 15:15:00.347452 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6pm2\" (UniqueName: \"kubernetes.io/projected/ed35cb6c-f2b1-4d8a-9b43-782ca62594c7-kube-api-access-s6pm2\") pod \"collect-profiles-29401395-hslqf\" (UID: \"ed35cb6c-f2b1-4d8a-9b43-782ca62594c7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401395-hslqf" Nov 25 15:15:00 crc kubenswrapper[4890]: I1125 15:15:00.463779 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401395-hslqf" Nov 25 15:15:00 crc kubenswrapper[4890]: I1125 15:15:00.937063 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401395-hslqf"] Nov 25 15:15:01 crc kubenswrapper[4890]: I1125 15:15:01.343328 4890 generic.go:334] "Generic (PLEG): container finished" podID="ed35cb6c-f2b1-4d8a-9b43-782ca62594c7" containerID="2229f386779c716886cbedd7df7fc219b411ba788eae258f76d6836ac9dc6d8f" exitCode=0 Nov 25 15:15:01 crc kubenswrapper[4890]: I1125 15:15:01.343366 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401395-hslqf" event={"ID":"ed35cb6c-f2b1-4d8a-9b43-782ca62594c7","Type":"ContainerDied","Data":"2229f386779c716886cbedd7df7fc219b411ba788eae258f76d6836ac9dc6d8f"} Nov 25 15:15:01 crc kubenswrapper[4890]: I1125 15:15:01.343424 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401395-hslqf" event={"ID":"ed35cb6c-f2b1-4d8a-9b43-782ca62594c7","Type":"ContainerStarted","Data":"e713d65433ef5ec1836024459406c19e47a50a92bfc94d5d9d4d547b5f662522"} Nov 25 15:15:02 crc kubenswrapper[4890]: I1125 15:15:02.049565 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-vtg6p" Nov 25 15:15:02 crc kubenswrapper[4890]: I1125 15:15:02.050478 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-vtg6p" Nov 25 15:15:02 crc kubenswrapper[4890]: I1125 15:15:02.078188 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-vtg6p" Nov 25 15:15:02 crc kubenswrapper[4890]: I1125 15:15:02.390260 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-vtg6p" Nov 25 15:15:02 crc kubenswrapper[4890]: I1125 15:15:02.665653 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401395-hslqf" Nov 25 15:15:02 crc kubenswrapper[4890]: I1125 15:15:02.760528 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ed35cb6c-f2b1-4d8a-9b43-782ca62594c7-config-volume\") pod \"ed35cb6c-f2b1-4d8a-9b43-782ca62594c7\" (UID: \"ed35cb6c-f2b1-4d8a-9b43-782ca62594c7\") " Nov 25 15:15:02 crc kubenswrapper[4890]: I1125 15:15:02.760576 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ed35cb6c-f2b1-4d8a-9b43-782ca62594c7-secret-volume\") pod \"ed35cb6c-f2b1-4d8a-9b43-782ca62594c7\" (UID: \"ed35cb6c-f2b1-4d8a-9b43-782ca62594c7\") " Nov 25 15:15:02 crc kubenswrapper[4890]: I1125 15:15:02.760645 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s6pm2\" (UniqueName: \"kubernetes.io/projected/ed35cb6c-f2b1-4d8a-9b43-782ca62594c7-kube-api-access-s6pm2\") pod \"ed35cb6c-f2b1-4d8a-9b43-782ca62594c7\" (UID: \"ed35cb6c-f2b1-4d8a-9b43-782ca62594c7\") " Nov 25 15:15:02 crc kubenswrapper[4890]: I1125 15:15:02.762037 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed35cb6c-f2b1-4d8a-9b43-782ca62594c7-config-volume" (OuterVolumeSpecName: "config-volume") pod "ed35cb6c-f2b1-4d8a-9b43-782ca62594c7" (UID: "ed35cb6c-f2b1-4d8a-9b43-782ca62594c7"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:15:02 crc kubenswrapper[4890]: I1125 15:15:02.767291 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed35cb6c-f2b1-4d8a-9b43-782ca62594c7-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "ed35cb6c-f2b1-4d8a-9b43-782ca62594c7" (UID: "ed35cb6c-f2b1-4d8a-9b43-782ca62594c7"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:15:02 crc kubenswrapper[4890]: I1125 15:15:02.768294 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed35cb6c-f2b1-4d8a-9b43-782ca62594c7-kube-api-access-s6pm2" (OuterVolumeSpecName: "kube-api-access-s6pm2") pod "ed35cb6c-f2b1-4d8a-9b43-782ca62594c7" (UID: "ed35cb6c-f2b1-4d8a-9b43-782ca62594c7"). InnerVolumeSpecName "kube-api-access-s6pm2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:15:02 crc kubenswrapper[4890]: I1125 15:15:02.862755 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s6pm2\" (UniqueName: \"kubernetes.io/projected/ed35cb6c-f2b1-4d8a-9b43-782ca62594c7-kube-api-access-s6pm2\") on node \"crc\" DevicePath \"\"" Nov 25 15:15:02 crc kubenswrapper[4890]: I1125 15:15:02.862798 4890 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ed35cb6c-f2b1-4d8a-9b43-782ca62594c7-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 15:15:02 crc kubenswrapper[4890]: I1125 15:15:02.862810 4890 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ed35cb6c-f2b1-4d8a-9b43-782ca62594c7-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 15:15:03 crc kubenswrapper[4890]: I1125 15:15:03.357295 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401395-hslqf" Nov 25 15:15:03 crc kubenswrapper[4890]: I1125 15:15:03.357387 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401395-hslqf" event={"ID":"ed35cb6c-f2b1-4d8a-9b43-782ca62594c7","Type":"ContainerDied","Data":"e713d65433ef5ec1836024459406c19e47a50a92bfc94d5d9d4d547b5f662522"} Nov 25 15:15:03 crc kubenswrapper[4890]: I1125 15:15:03.357452 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e713d65433ef5ec1836024459406c19e47a50a92bfc94d5d9d4d547b5f662522" Nov 25 15:15:06 crc kubenswrapper[4890]: I1125 15:15:06.093748 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ctt6j"] Nov 25 15:15:06 crc kubenswrapper[4890]: E1125 15:15:06.094794 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed35cb6c-f2b1-4d8a-9b43-782ca62594c7" containerName="collect-profiles" Nov 25 15:15:06 crc kubenswrapper[4890]: I1125 15:15:06.094817 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed35cb6c-f2b1-4d8a-9b43-782ca62594c7" containerName="collect-profiles" Nov 25 15:15:06 crc kubenswrapper[4890]: I1125 15:15:06.095006 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed35cb6c-f2b1-4d8a-9b43-782ca62594c7" containerName="collect-profiles" Nov 25 15:15:06 crc kubenswrapper[4890]: I1125 15:15:06.096033 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ctt6j" Nov 25 15:15:06 crc kubenswrapper[4890]: I1125 15:15:06.108997 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ctt6j"] Nov 25 15:15:06 crc kubenswrapper[4890]: I1125 15:15:06.209646 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ba31633-f93b-4389-8809-9499cab796f3-catalog-content\") pod \"certified-operators-ctt6j\" (UID: \"9ba31633-f93b-4389-8809-9499cab796f3\") " pod="openshift-marketplace/certified-operators-ctt6j" Nov 25 15:15:06 crc kubenswrapper[4890]: I1125 15:15:06.209703 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnsnn\" (UniqueName: \"kubernetes.io/projected/9ba31633-f93b-4389-8809-9499cab796f3-kube-api-access-nnsnn\") pod \"certified-operators-ctt6j\" (UID: \"9ba31633-f93b-4389-8809-9499cab796f3\") " pod="openshift-marketplace/certified-operators-ctt6j" Nov 25 15:15:06 crc kubenswrapper[4890]: I1125 15:15:06.209738 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ba31633-f93b-4389-8809-9499cab796f3-utilities\") pod \"certified-operators-ctt6j\" (UID: \"9ba31633-f93b-4389-8809-9499cab796f3\") " pod="openshift-marketplace/certified-operators-ctt6j" Nov 25 15:15:06 crc kubenswrapper[4890]: I1125 15:15:06.312003 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ba31633-f93b-4389-8809-9499cab796f3-catalog-content\") pod \"certified-operators-ctt6j\" (UID: \"9ba31633-f93b-4389-8809-9499cab796f3\") " pod="openshift-marketplace/certified-operators-ctt6j" Nov 25 15:15:06 crc kubenswrapper[4890]: I1125 15:15:06.312091 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnsnn\" (UniqueName: \"kubernetes.io/projected/9ba31633-f93b-4389-8809-9499cab796f3-kube-api-access-nnsnn\") pod \"certified-operators-ctt6j\" (UID: \"9ba31633-f93b-4389-8809-9499cab796f3\") " pod="openshift-marketplace/certified-operators-ctt6j" Nov 25 15:15:06 crc kubenswrapper[4890]: I1125 15:15:06.312189 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ba31633-f93b-4389-8809-9499cab796f3-utilities\") pod \"certified-operators-ctt6j\" (UID: \"9ba31633-f93b-4389-8809-9499cab796f3\") " pod="openshift-marketplace/certified-operators-ctt6j" Nov 25 15:15:06 crc kubenswrapper[4890]: I1125 15:15:06.313497 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ba31633-f93b-4389-8809-9499cab796f3-catalog-content\") pod \"certified-operators-ctt6j\" (UID: \"9ba31633-f93b-4389-8809-9499cab796f3\") " pod="openshift-marketplace/certified-operators-ctt6j" Nov 25 15:15:06 crc kubenswrapper[4890]: I1125 15:15:06.313848 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ba31633-f93b-4389-8809-9499cab796f3-utilities\") pod \"certified-operators-ctt6j\" (UID: \"9ba31633-f93b-4389-8809-9499cab796f3\") " pod="openshift-marketplace/certified-operators-ctt6j" Nov 25 15:15:06 crc kubenswrapper[4890]: I1125 15:15:06.340747 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnsnn\" (UniqueName: \"kubernetes.io/projected/9ba31633-f93b-4389-8809-9499cab796f3-kube-api-access-nnsnn\") pod \"certified-operators-ctt6j\" (UID: \"9ba31633-f93b-4389-8809-9499cab796f3\") " pod="openshift-marketplace/certified-operators-ctt6j" Nov 25 15:15:06 crc kubenswrapper[4890]: I1125 15:15:06.417264 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ctt6j" Nov 25 15:15:06 crc kubenswrapper[4890]: I1125 15:15:06.643192 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ctt6j"] Nov 25 15:15:07 crc kubenswrapper[4890]: I1125 15:15:07.382148 4890 generic.go:334] "Generic (PLEG): container finished" podID="9ba31633-f93b-4389-8809-9499cab796f3" containerID="616cf1a3678bb884d11e03bef505989cd6d989592e6fba7bb8fde81767be3f76" exitCode=0 Nov 25 15:15:07 crc kubenswrapper[4890]: I1125 15:15:07.382433 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ctt6j" event={"ID":"9ba31633-f93b-4389-8809-9499cab796f3","Type":"ContainerDied","Data":"616cf1a3678bb884d11e03bef505989cd6d989592e6fba7bb8fde81767be3f76"} Nov 25 15:15:07 crc kubenswrapper[4890]: I1125 15:15:07.382539 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ctt6j" event={"ID":"9ba31633-f93b-4389-8809-9499cab796f3","Type":"ContainerStarted","Data":"eeae85a886f99782bcc2d0f40ec1dca175b1d314ce15a3a52c2d48994085e95e"} Nov 25 15:15:08 crc kubenswrapper[4890]: I1125 15:15:08.393660 4890 generic.go:334] "Generic (PLEG): container finished" podID="9ba31633-f93b-4389-8809-9499cab796f3" containerID="6cfb20ac631984779180046718d98987050c7231d5d019819e37511bc1b293bc" exitCode=0 Nov 25 15:15:08 crc kubenswrapper[4890]: I1125 15:15:08.393725 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ctt6j" event={"ID":"9ba31633-f93b-4389-8809-9499cab796f3","Type":"ContainerDied","Data":"6cfb20ac631984779180046718d98987050c7231d5d019819e37511bc1b293bc"} Nov 25 15:15:09 crc kubenswrapper[4890]: I1125 15:15:09.401965 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ctt6j" event={"ID":"9ba31633-f93b-4389-8809-9499cab796f3","Type":"ContainerStarted","Data":"d52f79661ec696363c90ec67333ac60ef817109f6472075099adac841811d047"} Nov 25 15:15:09 crc kubenswrapper[4890]: I1125 15:15:09.430588 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ctt6j" podStartSLOduration=1.968664467 podStartE2EDuration="3.43056825s" podCreationTimestamp="2025-11-25 15:15:06 +0000 UTC" firstStartedPulling="2025-11-25 15:15:07.38498513 +0000 UTC m=+765.827447740" lastFinishedPulling="2025-11-25 15:15:08.846888903 +0000 UTC m=+767.289351523" observedRunningTime="2025-11-25 15:15:09.428622582 +0000 UTC m=+767.871085262" watchObservedRunningTime="2025-11-25 15:15:09.43056825 +0000 UTC m=+767.873030860" Nov 25 15:15:09 crc kubenswrapper[4890]: I1125 15:15:09.926246 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/9b547be2eac4c3e8f15e2800f340fb44715065f09b75b108844ab9dee67zk4x"] Nov 25 15:15:09 crc kubenswrapper[4890]: I1125 15:15:09.927724 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9b547be2eac4c3e8f15e2800f340fb44715065f09b75b108844ab9dee67zk4x" Nov 25 15:15:09 crc kubenswrapper[4890]: I1125 15:15:09.929836 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-rwwxf" Nov 25 15:15:09 crc kubenswrapper[4890]: I1125 15:15:09.938452 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9b547be2eac4c3e8f15e2800f340fb44715065f09b75b108844ab9dee67zk4x"] Nov 25 15:15:10 crc kubenswrapper[4890]: I1125 15:15:10.064885 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4050c336-b097-49f8-bc65-17bb8d191fa4-bundle\") pod \"9b547be2eac4c3e8f15e2800f340fb44715065f09b75b108844ab9dee67zk4x\" (UID: \"4050c336-b097-49f8-bc65-17bb8d191fa4\") " pod="openstack-operators/9b547be2eac4c3e8f15e2800f340fb44715065f09b75b108844ab9dee67zk4x" Nov 25 15:15:10 crc kubenswrapper[4890]: I1125 15:15:10.064951 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4050c336-b097-49f8-bc65-17bb8d191fa4-util\") pod \"9b547be2eac4c3e8f15e2800f340fb44715065f09b75b108844ab9dee67zk4x\" (UID: \"4050c336-b097-49f8-bc65-17bb8d191fa4\") " pod="openstack-operators/9b547be2eac4c3e8f15e2800f340fb44715065f09b75b108844ab9dee67zk4x" Nov 25 15:15:10 crc kubenswrapper[4890]: I1125 15:15:10.065016 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52tdz\" (UniqueName: \"kubernetes.io/projected/4050c336-b097-49f8-bc65-17bb8d191fa4-kube-api-access-52tdz\") pod \"9b547be2eac4c3e8f15e2800f340fb44715065f09b75b108844ab9dee67zk4x\" (UID: \"4050c336-b097-49f8-bc65-17bb8d191fa4\") " pod="openstack-operators/9b547be2eac4c3e8f15e2800f340fb44715065f09b75b108844ab9dee67zk4x" Nov 25 15:15:10 crc kubenswrapper[4890]: I1125 15:15:10.166747 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4050c336-b097-49f8-bc65-17bb8d191fa4-bundle\") pod \"9b547be2eac4c3e8f15e2800f340fb44715065f09b75b108844ab9dee67zk4x\" (UID: \"4050c336-b097-49f8-bc65-17bb8d191fa4\") " pod="openstack-operators/9b547be2eac4c3e8f15e2800f340fb44715065f09b75b108844ab9dee67zk4x" Nov 25 15:15:10 crc kubenswrapper[4890]: I1125 15:15:10.166819 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4050c336-b097-49f8-bc65-17bb8d191fa4-util\") pod \"9b547be2eac4c3e8f15e2800f340fb44715065f09b75b108844ab9dee67zk4x\" (UID: \"4050c336-b097-49f8-bc65-17bb8d191fa4\") " pod="openstack-operators/9b547be2eac4c3e8f15e2800f340fb44715065f09b75b108844ab9dee67zk4x" Nov 25 15:15:10 crc kubenswrapper[4890]: I1125 15:15:10.166850 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52tdz\" (UniqueName: \"kubernetes.io/projected/4050c336-b097-49f8-bc65-17bb8d191fa4-kube-api-access-52tdz\") pod \"9b547be2eac4c3e8f15e2800f340fb44715065f09b75b108844ab9dee67zk4x\" (UID: \"4050c336-b097-49f8-bc65-17bb8d191fa4\") " pod="openstack-operators/9b547be2eac4c3e8f15e2800f340fb44715065f09b75b108844ab9dee67zk4x" Nov 25 15:15:10 crc kubenswrapper[4890]: I1125 15:15:10.167517 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4050c336-b097-49f8-bc65-17bb8d191fa4-util\") pod \"9b547be2eac4c3e8f15e2800f340fb44715065f09b75b108844ab9dee67zk4x\" (UID: \"4050c336-b097-49f8-bc65-17bb8d191fa4\") " pod="openstack-operators/9b547be2eac4c3e8f15e2800f340fb44715065f09b75b108844ab9dee67zk4x" Nov 25 15:15:10 crc kubenswrapper[4890]: I1125 15:15:10.167546 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4050c336-b097-49f8-bc65-17bb8d191fa4-bundle\") pod \"9b547be2eac4c3e8f15e2800f340fb44715065f09b75b108844ab9dee67zk4x\" (UID: \"4050c336-b097-49f8-bc65-17bb8d191fa4\") " pod="openstack-operators/9b547be2eac4c3e8f15e2800f340fb44715065f09b75b108844ab9dee67zk4x" Nov 25 15:15:10 crc kubenswrapper[4890]: I1125 15:15:10.200198 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52tdz\" (UniqueName: \"kubernetes.io/projected/4050c336-b097-49f8-bc65-17bb8d191fa4-kube-api-access-52tdz\") pod \"9b547be2eac4c3e8f15e2800f340fb44715065f09b75b108844ab9dee67zk4x\" (UID: \"4050c336-b097-49f8-bc65-17bb8d191fa4\") " pod="openstack-operators/9b547be2eac4c3e8f15e2800f340fb44715065f09b75b108844ab9dee67zk4x" Nov 25 15:15:10 crc kubenswrapper[4890]: I1125 15:15:10.248650 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9b547be2eac4c3e8f15e2800f340fb44715065f09b75b108844ab9dee67zk4x" Nov 25 15:15:10 crc kubenswrapper[4890]: I1125 15:15:10.660874 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9b547be2eac4c3e8f15e2800f340fb44715065f09b75b108844ab9dee67zk4x"] Nov 25 15:15:10 crc kubenswrapper[4890]: W1125 15:15:10.666715 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4050c336_b097_49f8_bc65_17bb8d191fa4.slice/crio-0a58e9e9d740910347aab6ee5aa63c2f89f6bdc2966461fc4e5c473a19b94587 WatchSource:0}: Error finding container 0a58e9e9d740910347aab6ee5aa63c2f89f6bdc2966461fc4e5c473a19b94587: Status 404 returned error can't find the container with id 0a58e9e9d740910347aab6ee5aa63c2f89f6bdc2966461fc4e5c473a19b94587 Nov 25 15:15:11 crc kubenswrapper[4890]: I1125 15:15:11.422195 4890 generic.go:334] "Generic (PLEG): container finished" podID="4050c336-b097-49f8-bc65-17bb8d191fa4" containerID="78ebc8cd66f04d7069ac741700419ef9b135887978cf586f50cc0e44e87e08d3" exitCode=0 Nov 25 15:15:11 crc kubenswrapper[4890]: I1125 15:15:11.422289 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9b547be2eac4c3e8f15e2800f340fb44715065f09b75b108844ab9dee67zk4x" event={"ID":"4050c336-b097-49f8-bc65-17bb8d191fa4","Type":"ContainerDied","Data":"78ebc8cd66f04d7069ac741700419ef9b135887978cf586f50cc0e44e87e08d3"} Nov 25 15:15:11 crc kubenswrapper[4890]: I1125 15:15:11.422756 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9b547be2eac4c3e8f15e2800f340fb44715065f09b75b108844ab9dee67zk4x" event={"ID":"4050c336-b097-49f8-bc65-17bb8d191fa4","Type":"ContainerStarted","Data":"0a58e9e9d740910347aab6ee5aa63c2f89f6bdc2966461fc4e5c473a19b94587"} Nov 25 15:15:12 crc kubenswrapper[4890]: I1125 15:15:12.433220 4890 generic.go:334] "Generic (PLEG): container finished" podID="4050c336-b097-49f8-bc65-17bb8d191fa4" containerID="bd8d3f56809319187e638a5fbbdee928dd8dbe4dac9781bb58d71487e489925b" exitCode=0 Nov 25 15:15:12 crc kubenswrapper[4890]: I1125 15:15:12.433366 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9b547be2eac4c3e8f15e2800f340fb44715065f09b75b108844ab9dee67zk4x" event={"ID":"4050c336-b097-49f8-bc65-17bb8d191fa4","Type":"ContainerDied","Data":"bd8d3f56809319187e638a5fbbdee928dd8dbe4dac9781bb58d71487e489925b"} Nov 25 15:15:13 crc kubenswrapper[4890]: I1125 15:15:13.443251 4890 generic.go:334] "Generic (PLEG): container finished" podID="4050c336-b097-49f8-bc65-17bb8d191fa4" containerID="d13f42849a5b9986c34a4700cc02ad99fd0242965b8a36a0cf11c9f3de92d4b0" exitCode=0 Nov 25 15:15:13 crc kubenswrapper[4890]: I1125 15:15:13.443297 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9b547be2eac4c3e8f15e2800f340fb44715065f09b75b108844ab9dee67zk4x" event={"ID":"4050c336-b097-49f8-bc65-17bb8d191fa4","Type":"ContainerDied","Data":"d13f42849a5b9986c34a4700cc02ad99fd0242965b8a36a0cf11c9f3de92d4b0"} Nov 25 15:15:14 crc kubenswrapper[4890]: I1125 15:15:14.115218 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5p4bn"] Nov 25 15:15:14 crc kubenswrapper[4890]: I1125 15:15:14.117441 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5p4bn" Nov 25 15:15:14 crc kubenswrapper[4890]: I1125 15:15:14.132615 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5p4bn"] Nov 25 15:15:14 crc kubenswrapper[4890]: I1125 15:15:14.229855 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f02bea2-bd7d-4153-9b42-3a898dddea30-catalog-content\") pod \"redhat-marketplace-5p4bn\" (UID: \"2f02bea2-bd7d-4153-9b42-3a898dddea30\") " pod="openshift-marketplace/redhat-marketplace-5p4bn" Nov 25 15:15:14 crc kubenswrapper[4890]: I1125 15:15:14.230234 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f02bea2-bd7d-4153-9b42-3a898dddea30-utilities\") pod \"redhat-marketplace-5p4bn\" (UID: \"2f02bea2-bd7d-4153-9b42-3a898dddea30\") " pod="openshift-marketplace/redhat-marketplace-5p4bn" Nov 25 15:15:14 crc kubenswrapper[4890]: I1125 15:15:14.230373 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcz6h\" (UniqueName: \"kubernetes.io/projected/2f02bea2-bd7d-4153-9b42-3a898dddea30-kube-api-access-fcz6h\") pod \"redhat-marketplace-5p4bn\" (UID: \"2f02bea2-bd7d-4153-9b42-3a898dddea30\") " pod="openshift-marketplace/redhat-marketplace-5p4bn" Nov 25 15:15:14 crc kubenswrapper[4890]: I1125 15:15:14.332118 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fcz6h\" (UniqueName: \"kubernetes.io/projected/2f02bea2-bd7d-4153-9b42-3a898dddea30-kube-api-access-fcz6h\") pod \"redhat-marketplace-5p4bn\" (UID: \"2f02bea2-bd7d-4153-9b42-3a898dddea30\") " pod="openshift-marketplace/redhat-marketplace-5p4bn" Nov 25 15:15:14 crc kubenswrapper[4890]: I1125 15:15:14.332502 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f02bea2-bd7d-4153-9b42-3a898dddea30-catalog-content\") pod \"redhat-marketplace-5p4bn\" (UID: \"2f02bea2-bd7d-4153-9b42-3a898dddea30\") " pod="openshift-marketplace/redhat-marketplace-5p4bn" Nov 25 15:15:14 crc kubenswrapper[4890]: I1125 15:15:14.333730 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f02bea2-bd7d-4153-9b42-3a898dddea30-catalog-content\") pod \"redhat-marketplace-5p4bn\" (UID: \"2f02bea2-bd7d-4153-9b42-3a898dddea30\") " pod="openshift-marketplace/redhat-marketplace-5p4bn" Nov 25 15:15:14 crc kubenswrapper[4890]: I1125 15:15:14.334833 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f02bea2-bd7d-4153-9b42-3a898dddea30-utilities\") pod \"redhat-marketplace-5p4bn\" (UID: \"2f02bea2-bd7d-4153-9b42-3a898dddea30\") " pod="openshift-marketplace/redhat-marketplace-5p4bn" Nov 25 15:15:14 crc kubenswrapper[4890]: I1125 15:15:14.334124 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f02bea2-bd7d-4153-9b42-3a898dddea30-utilities\") pod \"redhat-marketplace-5p4bn\" (UID: \"2f02bea2-bd7d-4153-9b42-3a898dddea30\") " pod="openshift-marketplace/redhat-marketplace-5p4bn" Nov 25 15:15:14 crc kubenswrapper[4890]: I1125 15:15:14.352069 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fcz6h\" (UniqueName: \"kubernetes.io/projected/2f02bea2-bd7d-4153-9b42-3a898dddea30-kube-api-access-fcz6h\") pod \"redhat-marketplace-5p4bn\" (UID: \"2f02bea2-bd7d-4153-9b42-3a898dddea30\") " pod="openshift-marketplace/redhat-marketplace-5p4bn" Nov 25 15:15:14 crc kubenswrapper[4890]: I1125 15:15:14.452012 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5p4bn" Nov 25 15:15:14 crc kubenswrapper[4890]: I1125 15:15:14.747060 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9b547be2eac4c3e8f15e2800f340fb44715065f09b75b108844ab9dee67zk4x" Nov 25 15:15:14 crc kubenswrapper[4890]: I1125 15:15:14.842050 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-52tdz\" (UniqueName: \"kubernetes.io/projected/4050c336-b097-49f8-bc65-17bb8d191fa4-kube-api-access-52tdz\") pod \"4050c336-b097-49f8-bc65-17bb8d191fa4\" (UID: \"4050c336-b097-49f8-bc65-17bb8d191fa4\") " Nov 25 15:15:14 crc kubenswrapper[4890]: I1125 15:15:14.842261 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4050c336-b097-49f8-bc65-17bb8d191fa4-util\") pod \"4050c336-b097-49f8-bc65-17bb8d191fa4\" (UID: \"4050c336-b097-49f8-bc65-17bb8d191fa4\") " Nov 25 15:15:14 crc kubenswrapper[4890]: I1125 15:15:14.842301 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4050c336-b097-49f8-bc65-17bb8d191fa4-bundle\") pod \"4050c336-b097-49f8-bc65-17bb8d191fa4\" (UID: \"4050c336-b097-49f8-bc65-17bb8d191fa4\") " Nov 25 15:15:14 crc kubenswrapper[4890]: I1125 15:15:14.843585 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4050c336-b097-49f8-bc65-17bb8d191fa4-bundle" (OuterVolumeSpecName: "bundle") pod "4050c336-b097-49f8-bc65-17bb8d191fa4" (UID: "4050c336-b097-49f8-bc65-17bb8d191fa4"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:15:14 crc kubenswrapper[4890]: I1125 15:15:14.847920 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4050c336-b097-49f8-bc65-17bb8d191fa4-kube-api-access-52tdz" (OuterVolumeSpecName: "kube-api-access-52tdz") pod "4050c336-b097-49f8-bc65-17bb8d191fa4" (UID: "4050c336-b097-49f8-bc65-17bb8d191fa4"). InnerVolumeSpecName "kube-api-access-52tdz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:15:14 crc kubenswrapper[4890]: I1125 15:15:14.857380 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4050c336-b097-49f8-bc65-17bb8d191fa4-util" (OuterVolumeSpecName: "util") pod "4050c336-b097-49f8-bc65-17bb8d191fa4" (UID: "4050c336-b097-49f8-bc65-17bb8d191fa4"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:15:14 crc kubenswrapper[4890]: I1125 15:15:14.942669 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5p4bn"] Nov 25 15:15:14 crc kubenswrapper[4890]: I1125 15:15:14.943211 4890 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4050c336-b097-49f8-bc65-17bb8d191fa4-util\") on node \"crc\" DevicePath \"\"" Nov 25 15:15:14 crc kubenswrapper[4890]: I1125 15:15:14.943241 4890 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4050c336-b097-49f8-bc65-17bb8d191fa4-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:15:14 crc kubenswrapper[4890]: I1125 15:15:14.943250 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-52tdz\" (UniqueName: \"kubernetes.io/projected/4050c336-b097-49f8-bc65-17bb8d191fa4-kube-api-access-52tdz\") on node \"crc\" DevicePath \"\"" Nov 25 15:15:14 crc kubenswrapper[4890]: W1125 15:15:14.952337 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2f02bea2_bd7d_4153_9b42_3a898dddea30.slice/crio-32365b04975e703f859307891517355af55df8d80a26d78f4d61be9ab3d57805 WatchSource:0}: Error finding container 32365b04975e703f859307891517355af55df8d80a26d78f4d61be9ab3d57805: Status 404 returned error can't find the container with id 32365b04975e703f859307891517355af55df8d80a26d78f4d61be9ab3d57805 Nov 25 15:15:15 crc kubenswrapper[4890]: I1125 15:15:15.460124 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9b547be2eac4c3e8f15e2800f340fb44715065f09b75b108844ab9dee67zk4x" event={"ID":"4050c336-b097-49f8-bc65-17bb8d191fa4","Type":"ContainerDied","Data":"0a58e9e9d740910347aab6ee5aa63c2f89f6bdc2966461fc4e5c473a19b94587"} Nov 25 15:15:15 crc kubenswrapper[4890]: I1125 15:15:15.460181 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9b547be2eac4c3e8f15e2800f340fb44715065f09b75b108844ab9dee67zk4x" Nov 25 15:15:15 crc kubenswrapper[4890]: I1125 15:15:15.460184 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a58e9e9d740910347aab6ee5aa63c2f89f6bdc2966461fc4e5c473a19b94587" Nov 25 15:15:15 crc kubenswrapper[4890]: I1125 15:15:15.462229 4890 generic.go:334] "Generic (PLEG): container finished" podID="2f02bea2-bd7d-4153-9b42-3a898dddea30" containerID="d91b877ae1823ad0dffcba52d2099b23884c6897a141a7d85be3bdaf16ef9031" exitCode=0 Nov 25 15:15:15 crc kubenswrapper[4890]: I1125 15:15:15.462284 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5p4bn" event={"ID":"2f02bea2-bd7d-4153-9b42-3a898dddea30","Type":"ContainerDied","Data":"d91b877ae1823ad0dffcba52d2099b23884c6897a141a7d85be3bdaf16ef9031"} Nov 25 15:15:15 crc kubenswrapper[4890]: I1125 15:15:15.462336 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5p4bn" event={"ID":"2f02bea2-bd7d-4153-9b42-3a898dddea30","Type":"ContainerStarted","Data":"32365b04975e703f859307891517355af55df8d80a26d78f4d61be9ab3d57805"} Nov 25 15:15:16 crc kubenswrapper[4890]: I1125 15:15:16.417609 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ctt6j" Nov 25 15:15:16 crc kubenswrapper[4890]: I1125 15:15:16.418506 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ctt6j" Nov 25 15:15:16 crc kubenswrapper[4890]: I1125 15:15:16.457710 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ctt6j" Nov 25 15:15:16 crc kubenswrapper[4890]: I1125 15:15:16.474110 4890 generic.go:334] "Generic (PLEG): container finished" podID="2f02bea2-bd7d-4153-9b42-3a898dddea30" containerID="577ae3619c447d338e7ae675a53e32ff49ee3dded20e352a29232dab3617afe2" exitCode=0 Nov 25 15:15:16 crc kubenswrapper[4890]: I1125 15:15:16.474948 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5p4bn" event={"ID":"2f02bea2-bd7d-4153-9b42-3a898dddea30","Type":"ContainerDied","Data":"577ae3619c447d338e7ae675a53e32ff49ee3dded20e352a29232dab3617afe2"} Nov 25 15:15:16 crc kubenswrapper[4890]: I1125 15:15:16.524526 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ctt6j" Nov 25 15:15:17 crc kubenswrapper[4890]: I1125 15:15:17.482496 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5p4bn" event={"ID":"2f02bea2-bd7d-4153-9b42-3a898dddea30","Type":"ContainerStarted","Data":"55077ed2f4c41ddb9de994521d9468122ec92321709725330d94e7d84fe4e15d"} Nov 25 15:15:17 crc kubenswrapper[4890]: I1125 15:15:17.508609 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5p4bn" podStartSLOduration=2.102208864 podStartE2EDuration="3.508592342s" podCreationTimestamp="2025-11-25 15:15:14 +0000 UTC" firstStartedPulling="2025-11-25 15:15:15.464456527 +0000 UTC m=+773.906919147" lastFinishedPulling="2025-11-25 15:15:16.870839995 +0000 UTC m=+775.313302625" observedRunningTime="2025-11-25 15:15:17.506646263 +0000 UTC m=+775.949108883" watchObservedRunningTime="2025-11-25 15:15:17.508592342 +0000 UTC m=+775.951054942" Nov 25 15:15:17 crc kubenswrapper[4890]: I1125 15:15:17.677584 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6d88ccc4fc-9smnn"] Nov 25 15:15:17 crc kubenswrapper[4890]: E1125 15:15:17.678173 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4050c336-b097-49f8-bc65-17bb8d191fa4" containerName="util" Nov 25 15:15:17 crc kubenswrapper[4890]: I1125 15:15:17.678262 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="4050c336-b097-49f8-bc65-17bb8d191fa4" containerName="util" Nov 25 15:15:17 crc kubenswrapper[4890]: E1125 15:15:17.678337 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4050c336-b097-49f8-bc65-17bb8d191fa4" containerName="extract" Nov 25 15:15:17 crc kubenswrapper[4890]: I1125 15:15:17.678405 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="4050c336-b097-49f8-bc65-17bb8d191fa4" containerName="extract" Nov 25 15:15:17 crc kubenswrapper[4890]: E1125 15:15:17.678477 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4050c336-b097-49f8-bc65-17bb8d191fa4" containerName="pull" Nov 25 15:15:17 crc kubenswrapper[4890]: I1125 15:15:17.678537 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="4050c336-b097-49f8-bc65-17bb8d191fa4" containerName="pull" Nov 25 15:15:17 crc kubenswrapper[4890]: I1125 15:15:17.678771 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="4050c336-b097-49f8-bc65-17bb8d191fa4" containerName="extract" Nov 25 15:15:17 crc kubenswrapper[4890]: I1125 15:15:17.679359 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6d88ccc4fc-9smnn" Nov 25 15:15:17 crc kubenswrapper[4890]: I1125 15:15:17.686936 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-527z8" Nov 25 15:15:17 crc kubenswrapper[4890]: I1125 15:15:17.708029 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6d88ccc4fc-9smnn"] Nov 25 15:15:17 crc kubenswrapper[4890]: I1125 15:15:17.778925 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kx7s8\" (UniqueName: \"kubernetes.io/projected/bd60efda-e354-44d0-af56-c110651bdbaf-kube-api-access-kx7s8\") pod \"openstack-operator-controller-operator-6d88ccc4fc-9smnn\" (UID: \"bd60efda-e354-44d0-af56-c110651bdbaf\") " pod="openstack-operators/openstack-operator-controller-operator-6d88ccc4fc-9smnn" Nov 25 15:15:17 crc kubenswrapper[4890]: I1125 15:15:17.880490 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kx7s8\" (UniqueName: \"kubernetes.io/projected/bd60efda-e354-44d0-af56-c110651bdbaf-kube-api-access-kx7s8\") pod \"openstack-operator-controller-operator-6d88ccc4fc-9smnn\" (UID: \"bd60efda-e354-44d0-af56-c110651bdbaf\") " pod="openstack-operators/openstack-operator-controller-operator-6d88ccc4fc-9smnn" Nov 25 15:15:17 crc kubenswrapper[4890]: I1125 15:15:17.903772 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kx7s8\" (UniqueName: \"kubernetes.io/projected/bd60efda-e354-44d0-af56-c110651bdbaf-kube-api-access-kx7s8\") pod \"openstack-operator-controller-operator-6d88ccc4fc-9smnn\" (UID: \"bd60efda-e354-44d0-af56-c110651bdbaf\") " pod="openstack-operators/openstack-operator-controller-operator-6d88ccc4fc-9smnn" Nov 25 15:15:18 crc kubenswrapper[4890]: I1125 15:15:18.002091 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6d88ccc4fc-9smnn" Nov 25 15:15:18 crc kubenswrapper[4890]: I1125 15:15:18.449994 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6d88ccc4fc-9smnn"] Nov 25 15:15:18 crc kubenswrapper[4890]: W1125 15:15:18.452589 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbd60efda_e354_44d0_af56_c110651bdbaf.slice/crio-e595136579910bc8146b2beeaa972c07fe07a292e02a5a6ba8f3a32bbb173d70 WatchSource:0}: Error finding container e595136579910bc8146b2beeaa972c07fe07a292e02a5a6ba8f3a32bbb173d70: Status 404 returned error can't find the container with id e595136579910bc8146b2beeaa972c07fe07a292e02a5a6ba8f3a32bbb173d70 Nov 25 15:15:18 crc kubenswrapper[4890]: I1125 15:15:18.492704 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6d88ccc4fc-9smnn" event={"ID":"bd60efda-e354-44d0-af56-c110651bdbaf","Type":"ContainerStarted","Data":"e595136579910bc8146b2beeaa972c07fe07a292e02a5a6ba8f3a32bbb173d70"} Nov 25 15:15:19 crc kubenswrapper[4890]: I1125 15:15:19.878139 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ctt6j"] Nov 25 15:15:19 crc kubenswrapper[4890]: I1125 15:15:19.878833 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-ctt6j" podUID="9ba31633-f93b-4389-8809-9499cab796f3" containerName="registry-server" containerID="cri-o://d52f79661ec696363c90ec67333ac60ef817109f6472075099adac841811d047" gracePeriod=2 Nov 25 15:15:20 crc kubenswrapper[4890]: I1125 15:15:20.512247 4890 generic.go:334] "Generic (PLEG): container finished" podID="9ba31633-f93b-4389-8809-9499cab796f3" containerID="d52f79661ec696363c90ec67333ac60ef817109f6472075099adac841811d047" exitCode=0 Nov 25 15:15:20 crc kubenswrapper[4890]: I1125 15:15:20.512290 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ctt6j" event={"ID":"9ba31633-f93b-4389-8809-9499cab796f3","Type":"ContainerDied","Data":"d52f79661ec696363c90ec67333ac60ef817109f6472075099adac841811d047"} Nov 25 15:15:24 crc kubenswrapper[4890]: I1125 15:15:24.452929 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5p4bn" Nov 25 15:15:24 crc kubenswrapper[4890]: I1125 15:15:24.453442 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5p4bn" Nov 25 15:15:24 crc kubenswrapper[4890]: I1125 15:15:24.502266 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5p4bn" Nov 25 15:15:24 crc kubenswrapper[4890]: I1125 15:15:24.586641 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5p4bn" Nov 25 15:15:24 crc kubenswrapper[4890]: I1125 15:15:24.949651 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ctt6j" Nov 25 15:15:25 crc kubenswrapper[4890]: I1125 15:15:25.074082 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nnsnn\" (UniqueName: \"kubernetes.io/projected/9ba31633-f93b-4389-8809-9499cab796f3-kube-api-access-nnsnn\") pod \"9ba31633-f93b-4389-8809-9499cab796f3\" (UID: \"9ba31633-f93b-4389-8809-9499cab796f3\") " Nov 25 15:15:25 crc kubenswrapper[4890]: I1125 15:15:25.074184 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ba31633-f93b-4389-8809-9499cab796f3-utilities\") pod \"9ba31633-f93b-4389-8809-9499cab796f3\" (UID: \"9ba31633-f93b-4389-8809-9499cab796f3\") " Nov 25 15:15:25 crc kubenswrapper[4890]: I1125 15:15:25.074267 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ba31633-f93b-4389-8809-9499cab796f3-catalog-content\") pod \"9ba31633-f93b-4389-8809-9499cab796f3\" (UID: \"9ba31633-f93b-4389-8809-9499cab796f3\") " Nov 25 15:15:25 crc kubenswrapper[4890]: I1125 15:15:25.074858 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ba31633-f93b-4389-8809-9499cab796f3-utilities" (OuterVolumeSpecName: "utilities") pod "9ba31633-f93b-4389-8809-9499cab796f3" (UID: "9ba31633-f93b-4389-8809-9499cab796f3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:15:25 crc kubenswrapper[4890]: I1125 15:15:25.080569 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ba31633-f93b-4389-8809-9499cab796f3-kube-api-access-nnsnn" (OuterVolumeSpecName: "kube-api-access-nnsnn") pod "9ba31633-f93b-4389-8809-9499cab796f3" (UID: "9ba31633-f93b-4389-8809-9499cab796f3"). InnerVolumeSpecName "kube-api-access-nnsnn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:15:25 crc kubenswrapper[4890]: I1125 15:15:25.122805 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ba31633-f93b-4389-8809-9499cab796f3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9ba31633-f93b-4389-8809-9499cab796f3" (UID: "9ba31633-f93b-4389-8809-9499cab796f3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:15:25 crc kubenswrapper[4890]: I1125 15:15:25.176285 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nnsnn\" (UniqueName: \"kubernetes.io/projected/9ba31633-f93b-4389-8809-9499cab796f3-kube-api-access-nnsnn\") on node \"crc\" DevicePath \"\"" Nov 25 15:15:25 crc kubenswrapper[4890]: I1125 15:15:25.176317 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ba31633-f93b-4389-8809-9499cab796f3-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 15:15:25 crc kubenswrapper[4890]: I1125 15:15:25.176327 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ba31633-f93b-4389-8809-9499cab796f3-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 15:15:25 crc kubenswrapper[4890]: I1125 15:15:25.557953 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ctt6j" event={"ID":"9ba31633-f93b-4389-8809-9499cab796f3","Type":"ContainerDied","Data":"eeae85a886f99782bcc2d0f40ec1dca175b1d314ce15a3a52c2d48994085e95e"} Nov 25 15:15:25 crc kubenswrapper[4890]: I1125 15:15:25.558008 4890 scope.go:117] "RemoveContainer" containerID="d52f79661ec696363c90ec67333ac60ef817109f6472075099adac841811d047" Nov 25 15:15:25 crc kubenswrapper[4890]: I1125 15:15:25.558144 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ctt6j" Nov 25 15:15:25 crc kubenswrapper[4890]: I1125 15:15:25.568936 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6d88ccc4fc-9smnn" event={"ID":"bd60efda-e354-44d0-af56-c110651bdbaf","Type":"ContainerStarted","Data":"f6e34e9f7f991c3d5b3dc786e507a602a630e025bed28294ab4e01d68a0244e4"} Nov 25 15:15:25 crc kubenswrapper[4890]: I1125 15:15:25.569000 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-6d88ccc4fc-9smnn" Nov 25 15:15:25 crc kubenswrapper[4890]: I1125 15:15:25.578217 4890 scope.go:117] "RemoveContainer" containerID="6cfb20ac631984779180046718d98987050c7231d5d019819e37511bc1b293bc" Nov 25 15:15:25 crc kubenswrapper[4890]: I1125 15:15:25.599238 4890 scope.go:117] "RemoveContainer" containerID="616cf1a3678bb884d11e03bef505989cd6d989592e6fba7bb8fde81767be3f76" Nov 25 15:15:25 crc kubenswrapper[4890]: I1125 15:15:25.609132 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-6d88ccc4fc-9smnn" podStartSLOduration=1.7205371409999999 podStartE2EDuration="8.609105415s" podCreationTimestamp="2025-11-25 15:15:17 +0000 UTC" firstStartedPulling="2025-11-25 15:15:18.455416507 +0000 UTC m=+776.897879117" lastFinishedPulling="2025-11-25 15:15:25.343984781 +0000 UTC m=+783.786447391" observedRunningTime="2025-11-25 15:15:25.602607283 +0000 UTC m=+784.045069903" watchObservedRunningTime="2025-11-25 15:15:25.609105415 +0000 UTC m=+784.051568025" Nov 25 15:15:25 crc kubenswrapper[4890]: I1125 15:15:25.620374 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ctt6j"] Nov 25 15:15:25 crc kubenswrapper[4890]: I1125 15:15:25.625298 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-ctt6j"] Nov 25 15:15:26 crc kubenswrapper[4890]: I1125 15:15:26.089743 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vhkk5"] Nov 25 15:15:26 crc kubenswrapper[4890]: E1125 15:15:26.090073 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ba31633-f93b-4389-8809-9499cab796f3" containerName="registry-server" Nov 25 15:15:26 crc kubenswrapper[4890]: I1125 15:15:26.090089 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ba31633-f93b-4389-8809-9499cab796f3" containerName="registry-server" Nov 25 15:15:26 crc kubenswrapper[4890]: E1125 15:15:26.090101 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ba31633-f93b-4389-8809-9499cab796f3" containerName="extract-utilities" Nov 25 15:15:26 crc kubenswrapper[4890]: I1125 15:15:26.090109 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ba31633-f93b-4389-8809-9499cab796f3" containerName="extract-utilities" Nov 25 15:15:26 crc kubenswrapper[4890]: E1125 15:15:26.090141 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ba31633-f93b-4389-8809-9499cab796f3" containerName="extract-content" Nov 25 15:15:26 crc kubenswrapper[4890]: I1125 15:15:26.090153 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ba31633-f93b-4389-8809-9499cab796f3" containerName="extract-content" Nov 25 15:15:26 crc kubenswrapper[4890]: I1125 15:15:26.090327 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ba31633-f93b-4389-8809-9499cab796f3" containerName="registry-server" Nov 25 15:15:26 crc kubenswrapper[4890]: I1125 15:15:26.091389 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vhkk5" Nov 25 15:15:26 crc kubenswrapper[4890]: I1125 15:15:26.146684 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vhkk5"] Nov 25 15:15:26 crc kubenswrapper[4890]: I1125 15:15:26.182342 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ba31633-f93b-4389-8809-9499cab796f3" path="/var/lib/kubelet/pods/9ba31633-f93b-4389-8809-9499cab796f3/volumes" Nov 25 15:15:26 crc kubenswrapper[4890]: I1125 15:15:26.190962 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e8357ad-84de-4f9c-9902-7109dabb2dc3-catalog-content\") pod \"community-operators-vhkk5\" (UID: \"8e8357ad-84de-4f9c-9902-7109dabb2dc3\") " pod="openshift-marketplace/community-operators-vhkk5" Nov 25 15:15:26 crc kubenswrapper[4890]: I1125 15:15:26.191006 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e8357ad-84de-4f9c-9902-7109dabb2dc3-utilities\") pod \"community-operators-vhkk5\" (UID: \"8e8357ad-84de-4f9c-9902-7109dabb2dc3\") " pod="openshift-marketplace/community-operators-vhkk5" Nov 25 15:15:26 crc kubenswrapper[4890]: I1125 15:15:26.191031 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxfdc\" (UniqueName: \"kubernetes.io/projected/8e8357ad-84de-4f9c-9902-7109dabb2dc3-kube-api-access-zxfdc\") pod \"community-operators-vhkk5\" (UID: \"8e8357ad-84de-4f9c-9902-7109dabb2dc3\") " pod="openshift-marketplace/community-operators-vhkk5" Nov 25 15:15:26 crc kubenswrapper[4890]: I1125 15:15:26.293062 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e8357ad-84de-4f9c-9902-7109dabb2dc3-catalog-content\") pod \"community-operators-vhkk5\" (UID: \"8e8357ad-84de-4f9c-9902-7109dabb2dc3\") " pod="openshift-marketplace/community-operators-vhkk5" Nov 25 15:15:26 crc kubenswrapper[4890]: I1125 15:15:26.293113 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e8357ad-84de-4f9c-9902-7109dabb2dc3-utilities\") pod \"community-operators-vhkk5\" (UID: \"8e8357ad-84de-4f9c-9902-7109dabb2dc3\") " pod="openshift-marketplace/community-operators-vhkk5" Nov 25 15:15:26 crc kubenswrapper[4890]: I1125 15:15:26.293129 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxfdc\" (UniqueName: \"kubernetes.io/projected/8e8357ad-84de-4f9c-9902-7109dabb2dc3-kube-api-access-zxfdc\") pod \"community-operators-vhkk5\" (UID: \"8e8357ad-84de-4f9c-9902-7109dabb2dc3\") " pod="openshift-marketplace/community-operators-vhkk5" Nov 25 15:15:26 crc kubenswrapper[4890]: I1125 15:15:26.294228 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e8357ad-84de-4f9c-9902-7109dabb2dc3-catalog-content\") pod \"community-operators-vhkk5\" (UID: \"8e8357ad-84de-4f9c-9902-7109dabb2dc3\") " pod="openshift-marketplace/community-operators-vhkk5" Nov 25 15:15:26 crc kubenswrapper[4890]: I1125 15:15:26.294503 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e8357ad-84de-4f9c-9902-7109dabb2dc3-utilities\") pod \"community-operators-vhkk5\" (UID: \"8e8357ad-84de-4f9c-9902-7109dabb2dc3\") " pod="openshift-marketplace/community-operators-vhkk5" Nov 25 15:15:26 crc kubenswrapper[4890]: I1125 15:15:26.312263 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxfdc\" (UniqueName: \"kubernetes.io/projected/8e8357ad-84de-4f9c-9902-7109dabb2dc3-kube-api-access-zxfdc\") pod \"community-operators-vhkk5\" (UID: \"8e8357ad-84de-4f9c-9902-7109dabb2dc3\") " pod="openshift-marketplace/community-operators-vhkk5" Nov 25 15:15:26 crc kubenswrapper[4890]: I1125 15:15:26.429010 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vhkk5" Nov 25 15:15:26 crc kubenswrapper[4890]: I1125 15:15:26.736360 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vhkk5"] Nov 25 15:15:27 crc kubenswrapper[4890]: I1125 15:15:27.600196 4890 generic.go:334] "Generic (PLEG): container finished" podID="8e8357ad-84de-4f9c-9902-7109dabb2dc3" containerID="67afd6a182b3cd90bc50c60f35aaf92e4fbbd85de76d773b2c756a91870529b0" exitCode=0 Nov 25 15:15:27 crc kubenswrapper[4890]: I1125 15:15:27.600270 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vhkk5" event={"ID":"8e8357ad-84de-4f9c-9902-7109dabb2dc3","Type":"ContainerDied","Data":"67afd6a182b3cd90bc50c60f35aaf92e4fbbd85de76d773b2c756a91870529b0"} Nov 25 15:15:27 crc kubenswrapper[4890]: I1125 15:15:27.600663 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vhkk5" event={"ID":"8e8357ad-84de-4f9c-9902-7109dabb2dc3","Type":"ContainerStarted","Data":"8b6180d7ddb272cca4c2bba33651fa2a371158bcf32c299d49f28397226b1817"} Nov 25 15:15:28 crc kubenswrapper[4890]: I1125 15:15:28.876139 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5p4bn"] Nov 25 15:15:28 crc kubenswrapper[4890]: I1125 15:15:28.876428 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5p4bn" podUID="2f02bea2-bd7d-4153-9b42-3a898dddea30" containerName="registry-server" containerID="cri-o://55077ed2f4c41ddb9de994521d9468122ec92321709725330d94e7d84fe4e15d" gracePeriod=2 Nov 25 15:15:29 crc kubenswrapper[4890]: I1125 15:15:29.266297 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5p4bn" Nov 25 15:15:29 crc kubenswrapper[4890]: I1125 15:15:29.341190 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f02bea2-bd7d-4153-9b42-3a898dddea30-catalog-content\") pod \"2f02bea2-bd7d-4153-9b42-3a898dddea30\" (UID: \"2f02bea2-bd7d-4153-9b42-3a898dddea30\") " Nov 25 15:15:29 crc kubenswrapper[4890]: I1125 15:15:29.341307 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcz6h\" (UniqueName: \"kubernetes.io/projected/2f02bea2-bd7d-4153-9b42-3a898dddea30-kube-api-access-fcz6h\") pod \"2f02bea2-bd7d-4153-9b42-3a898dddea30\" (UID: \"2f02bea2-bd7d-4153-9b42-3a898dddea30\") " Nov 25 15:15:29 crc kubenswrapper[4890]: I1125 15:15:29.341383 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f02bea2-bd7d-4153-9b42-3a898dddea30-utilities\") pod \"2f02bea2-bd7d-4153-9b42-3a898dddea30\" (UID: \"2f02bea2-bd7d-4153-9b42-3a898dddea30\") " Nov 25 15:15:29 crc kubenswrapper[4890]: I1125 15:15:29.342457 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f02bea2-bd7d-4153-9b42-3a898dddea30-utilities" (OuterVolumeSpecName: "utilities") pod "2f02bea2-bd7d-4153-9b42-3a898dddea30" (UID: "2f02bea2-bd7d-4153-9b42-3a898dddea30"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:15:29 crc kubenswrapper[4890]: I1125 15:15:29.346136 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f02bea2-bd7d-4153-9b42-3a898dddea30-kube-api-access-fcz6h" (OuterVolumeSpecName: "kube-api-access-fcz6h") pod "2f02bea2-bd7d-4153-9b42-3a898dddea30" (UID: "2f02bea2-bd7d-4153-9b42-3a898dddea30"). InnerVolumeSpecName "kube-api-access-fcz6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:15:29 crc kubenswrapper[4890]: I1125 15:15:29.362644 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f02bea2-bd7d-4153-9b42-3a898dddea30-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2f02bea2-bd7d-4153-9b42-3a898dddea30" (UID: "2f02bea2-bd7d-4153-9b42-3a898dddea30"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:15:29 crc kubenswrapper[4890]: I1125 15:15:29.442660 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f02bea2-bd7d-4153-9b42-3a898dddea30-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 15:15:29 crc kubenswrapper[4890]: I1125 15:15:29.442704 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcz6h\" (UniqueName: \"kubernetes.io/projected/2f02bea2-bd7d-4153-9b42-3a898dddea30-kube-api-access-fcz6h\") on node \"crc\" DevicePath \"\"" Nov 25 15:15:29 crc kubenswrapper[4890]: I1125 15:15:29.442718 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f02bea2-bd7d-4153-9b42-3a898dddea30-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 15:15:29 crc kubenswrapper[4890]: I1125 15:15:29.619077 4890 generic.go:334] "Generic (PLEG): container finished" podID="8e8357ad-84de-4f9c-9902-7109dabb2dc3" containerID="43f56df7e9635d8afda56f4558cbfc1334f403e9c5098c2472d94f0b72394fd6" exitCode=0 Nov 25 15:15:29 crc kubenswrapper[4890]: I1125 15:15:29.619174 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vhkk5" event={"ID":"8e8357ad-84de-4f9c-9902-7109dabb2dc3","Type":"ContainerDied","Data":"43f56df7e9635d8afda56f4558cbfc1334f403e9c5098c2472d94f0b72394fd6"} Nov 25 15:15:29 crc kubenswrapper[4890]: I1125 15:15:29.622535 4890 generic.go:334] "Generic (PLEG): container finished" podID="2f02bea2-bd7d-4153-9b42-3a898dddea30" containerID="55077ed2f4c41ddb9de994521d9468122ec92321709725330d94e7d84fe4e15d" exitCode=0 Nov 25 15:15:29 crc kubenswrapper[4890]: I1125 15:15:29.622605 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5p4bn" Nov 25 15:15:29 crc kubenswrapper[4890]: I1125 15:15:29.622617 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5p4bn" event={"ID":"2f02bea2-bd7d-4153-9b42-3a898dddea30","Type":"ContainerDied","Data":"55077ed2f4c41ddb9de994521d9468122ec92321709725330d94e7d84fe4e15d"} Nov 25 15:15:29 crc kubenswrapper[4890]: I1125 15:15:29.622671 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5p4bn" event={"ID":"2f02bea2-bd7d-4153-9b42-3a898dddea30","Type":"ContainerDied","Data":"32365b04975e703f859307891517355af55df8d80a26d78f4d61be9ab3d57805"} Nov 25 15:15:29 crc kubenswrapper[4890]: I1125 15:15:29.622698 4890 scope.go:117] "RemoveContainer" containerID="55077ed2f4c41ddb9de994521d9468122ec92321709725330d94e7d84fe4e15d" Nov 25 15:15:29 crc kubenswrapper[4890]: I1125 15:15:29.652956 4890 scope.go:117] "RemoveContainer" containerID="577ae3619c447d338e7ae675a53e32ff49ee3dded20e352a29232dab3617afe2" Nov 25 15:15:29 crc kubenswrapper[4890]: I1125 15:15:29.679068 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5p4bn"] Nov 25 15:15:29 crc kubenswrapper[4890]: I1125 15:15:29.686092 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5p4bn"] Nov 25 15:15:29 crc kubenswrapper[4890]: I1125 15:15:29.689782 4890 scope.go:117] "RemoveContainer" containerID="d91b877ae1823ad0dffcba52d2099b23884c6897a141a7d85be3bdaf16ef9031" Nov 25 15:15:29 crc kubenswrapper[4890]: I1125 15:15:29.704581 4890 scope.go:117] "RemoveContainer" containerID="55077ed2f4c41ddb9de994521d9468122ec92321709725330d94e7d84fe4e15d" Nov 25 15:15:29 crc kubenswrapper[4890]: E1125 15:15:29.705058 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55077ed2f4c41ddb9de994521d9468122ec92321709725330d94e7d84fe4e15d\": container with ID starting with 55077ed2f4c41ddb9de994521d9468122ec92321709725330d94e7d84fe4e15d not found: ID does not exist" containerID="55077ed2f4c41ddb9de994521d9468122ec92321709725330d94e7d84fe4e15d" Nov 25 15:15:29 crc kubenswrapper[4890]: I1125 15:15:29.705097 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55077ed2f4c41ddb9de994521d9468122ec92321709725330d94e7d84fe4e15d"} err="failed to get container status \"55077ed2f4c41ddb9de994521d9468122ec92321709725330d94e7d84fe4e15d\": rpc error: code = NotFound desc = could not find container \"55077ed2f4c41ddb9de994521d9468122ec92321709725330d94e7d84fe4e15d\": container with ID starting with 55077ed2f4c41ddb9de994521d9468122ec92321709725330d94e7d84fe4e15d not found: ID does not exist" Nov 25 15:15:29 crc kubenswrapper[4890]: I1125 15:15:29.705130 4890 scope.go:117] "RemoveContainer" containerID="577ae3619c447d338e7ae675a53e32ff49ee3dded20e352a29232dab3617afe2" Nov 25 15:15:29 crc kubenswrapper[4890]: E1125 15:15:29.705549 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"577ae3619c447d338e7ae675a53e32ff49ee3dded20e352a29232dab3617afe2\": container with ID starting with 577ae3619c447d338e7ae675a53e32ff49ee3dded20e352a29232dab3617afe2 not found: ID does not exist" containerID="577ae3619c447d338e7ae675a53e32ff49ee3dded20e352a29232dab3617afe2" Nov 25 15:15:29 crc kubenswrapper[4890]: I1125 15:15:29.705575 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"577ae3619c447d338e7ae675a53e32ff49ee3dded20e352a29232dab3617afe2"} err="failed to get container status \"577ae3619c447d338e7ae675a53e32ff49ee3dded20e352a29232dab3617afe2\": rpc error: code = NotFound desc = could not find container \"577ae3619c447d338e7ae675a53e32ff49ee3dded20e352a29232dab3617afe2\": container with ID starting with 577ae3619c447d338e7ae675a53e32ff49ee3dded20e352a29232dab3617afe2 not found: ID does not exist" Nov 25 15:15:29 crc kubenswrapper[4890]: I1125 15:15:29.705592 4890 scope.go:117] "RemoveContainer" containerID="d91b877ae1823ad0dffcba52d2099b23884c6897a141a7d85be3bdaf16ef9031" Nov 25 15:15:29 crc kubenswrapper[4890]: E1125 15:15:29.705869 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d91b877ae1823ad0dffcba52d2099b23884c6897a141a7d85be3bdaf16ef9031\": container with ID starting with d91b877ae1823ad0dffcba52d2099b23884c6897a141a7d85be3bdaf16ef9031 not found: ID does not exist" containerID="d91b877ae1823ad0dffcba52d2099b23884c6897a141a7d85be3bdaf16ef9031" Nov 25 15:15:29 crc kubenswrapper[4890]: I1125 15:15:29.705895 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d91b877ae1823ad0dffcba52d2099b23884c6897a141a7d85be3bdaf16ef9031"} err="failed to get container status \"d91b877ae1823ad0dffcba52d2099b23884c6897a141a7d85be3bdaf16ef9031\": rpc error: code = NotFound desc = could not find container \"d91b877ae1823ad0dffcba52d2099b23884c6897a141a7d85be3bdaf16ef9031\": container with ID starting with d91b877ae1823ad0dffcba52d2099b23884c6897a141a7d85be3bdaf16ef9031 not found: ID does not exist" Nov 25 15:15:30 crc kubenswrapper[4890]: I1125 15:15:30.180455 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f02bea2-bd7d-4153-9b42-3a898dddea30" path="/var/lib/kubelet/pods/2f02bea2-bd7d-4153-9b42-3a898dddea30/volumes" Nov 25 15:15:30 crc kubenswrapper[4890]: I1125 15:15:30.631296 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vhkk5" event={"ID":"8e8357ad-84de-4f9c-9902-7109dabb2dc3","Type":"ContainerStarted","Data":"34d391804e7af1a69145fcf4e4eda2b537746c2a14e72b9b71fdb9fce45a5f64"} Nov 25 15:15:30 crc kubenswrapper[4890]: I1125 15:15:30.647364 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vhkk5" podStartSLOduration=1.880882514 podStartE2EDuration="4.647345764s" podCreationTimestamp="2025-11-25 15:15:26 +0000 UTC" firstStartedPulling="2025-11-25 15:15:27.601979127 +0000 UTC m=+786.044441737" lastFinishedPulling="2025-11-25 15:15:30.368442367 +0000 UTC m=+788.810904987" observedRunningTime="2025-11-25 15:15:30.646110423 +0000 UTC m=+789.088573053" watchObservedRunningTime="2025-11-25 15:15:30.647345764 +0000 UTC m=+789.089808374" Nov 25 15:15:36 crc kubenswrapper[4890]: I1125 15:15:36.429302 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vhkk5" Nov 25 15:15:36 crc kubenswrapper[4890]: I1125 15:15:36.431006 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vhkk5" Nov 25 15:15:36 crc kubenswrapper[4890]: I1125 15:15:36.479138 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vhkk5" Nov 25 15:15:36 crc kubenswrapper[4890]: I1125 15:15:36.713480 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vhkk5" Nov 25 15:15:37 crc kubenswrapper[4890]: I1125 15:15:37.293382 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vhkk5"] Nov 25 15:15:38 crc kubenswrapper[4890]: I1125 15:15:38.005328 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-6d88ccc4fc-9smnn" Nov 25 15:15:38 crc kubenswrapper[4890]: I1125 15:15:38.684361 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vhkk5" podUID="8e8357ad-84de-4f9c-9902-7109dabb2dc3" containerName="registry-server" containerID="cri-o://34d391804e7af1a69145fcf4e4eda2b537746c2a14e72b9b71fdb9fce45a5f64" gracePeriod=2 Nov 25 15:15:39 crc kubenswrapper[4890]: I1125 15:15:39.072690 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vhkk5" Nov 25 15:15:39 crc kubenswrapper[4890]: I1125 15:15:39.195932 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zxfdc\" (UniqueName: \"kubernetes.io/projected/8e8357ad-84de-4f9c-9902-7109dabb2dc3-kube-api-access-zxfdc\") pod \"8e8357ad-84de-4f9c-9902-7109dabb2dc3\" (UID: \"8e8357ad-84de-4f9c-9902-7109dabb2dc3\") " Nov 25 15:15:39 crc kubenswrapper[4890]: I1125 15:15:39.196007 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e8357ad-84de-4f9c-9902-7109dabb2dc3-utilities\") pod \"8e8357ad-84de-4f9c-9902-7109dabb2dc3\" (UID: \"8e8357ad-84de-4f9c-9902-7109dabb2dc3\") " Nov 25 15:15:39 crc kubenswrapper[4890]: I1125 15:15:39.196029 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e8357ad-84de-4f9c-9902-7109dabb2dc3-catalog-content\") pod \"8e8357ad-84de-4f9c-9902-7109dabb2dc3\" (UID: \"8e8357ad-84de-4f9c-9902-7109dabb2dc3\") " Nov 25 15:15:39 crc kubenswrapper[4890]: I1125 15:15:39.197775 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e8357ad-84de-4f9c-9902-7109dabb2dc3-utilities" (OuterVolumeSpecName: "utilities") pod "8e8357ad-84de-4f9c-9902-7109dabb2dc3" (UID: "8e8357ad-84de-4f9c-9902-7109dabb2dc3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:15:39 crc kubenswrapper[4890]: I1125 15:15:39.200759 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e8357ad-84de-4f9c-9902-7109dabb2dc3-kube-api-access-zxfdc" (OuterVolumeSpecName: "kube-api-access-zxfdc") pod "8e8357ad-84de-4f9c-9902-7109dabb2dc3" (UID: "8e8357ad-84de-4f9c-9902-7109dabb2dc3"). InnerVolumeSpecName "kube-api-access-zxfdc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:15:39 crc kubenswrapper[4890]: I1125 15:15:39.252229 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e8357ad-84de-4f9c-9902-7109dabb2dc3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8e8357ad-84de-4f9c-9902-7109dabb2dc3" (UID: "8e8357ad-84de-4f9c-9902-7109dabb2dc3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:15:39 crc kubenswrapper[4890]: I1125 15:15:39.298692 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e8357ad-84de-4f9c-9902-7109dabb2dc3-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 15:15:39 crc kubenswrapper[4890]: I1125 15:15:39.298721 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e8357ad-84de-4f9c-9902-7109dabb2dc3-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 15:15:39 crc kubenswrapper[4890]: I1125 15:15:39.298735 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zxfdc\" (UniqueName: \"kubernetes.io/projected/8e8357ad-84de-4f9c-9902-7109dabb2dc3-kube-api-access-zxfdc\") on node \"crc\" DevicePath \"\"" Nov 25 15:15:39 crc kubenswrapper[4890]: I1125 15:15:39.702960 4890 generic.go:334] "Generic (PLEG): container finished" podID="8e8357ad-84de-4f9c-9902-7109dabb2dc3" containerID="34d391804e7af1a69145fcf4e4eda2b537746c2a14e72b9b71fdb9fce45a5f64" exitCode=0 Nov 25 15:15:39 crc kubenswrapper[4890]: I1125 15:15:39.703062 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vhkk5" event={"ID":"8e8357ad-84de-4f9c-9902-7109dabb2dc3","Type":"ContainerDied","Data":"34d391804e7af1a69145fcf4e4eda2b537746c2a14e72b9b71fdb9fce45a5f64"} Nov 25 15:15:39 crc kubenswrapper[4890]: I1125 15:15:39.703092 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vhkk5" Nov 25 15:15:39 crc kubenswrapper[4890]: I1125 15:15:39.703313 4890 scope.go:117] "RemoveContainer" containerID="34d391804e7af1a69145fcf4e4eda2b537746c2a14e72b9b71fdb9fce45a5f64" Nov 25 15:15:39 crc kubenswrapper[4890]: I1125 15:15:39.703605 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vhkk5" event={"ID":"8e8357ad-84de-4f9c-9902-7109dabb2dc3","Type":"ContainerDied","Data":"8b6180d7ddb272cca4c2bba33651fa2a371158bcf32c299d49f28397226b1817"} Nov 25 15:15:39 crc kubenswrapper[4890]: I1125 15:15:39.739059 4890 scope.go:117] "RemoveContainer" containerID="43f56df7e9635d8afda56f4558cbfc1334f403e9c5098c2472d94f0b72394fd6" Nov 25 15:15:39 crc kubenswrapper[4890]: I1125 15:15:39.743431 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vhkk5"] Nov 25 15:15:39 crc kubenswrapper[4890]: I1125 15:15:39.748094 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vhkk5"] Nov 25 15:15:39 crc kubenswrapper[4890]: I1125 15:15:39.757917 4890 scope.go:117] "RemoveContainer" containerID="67afd6a182b3cd90bc50c60f35aaf92e4fbbd85de76d773b2c756a91870529b0" Nov 25 15:15:39 crc kubenswrapper[4890]: I1125 15:15:39.778444 4890 scope.go:117] "RemoveContainer" containerID="34d391804e7af1a69145fcf4e4eda2b537746c2a14e72b9b71fdb9fce45a5f64" Nov 25 15:15:39 crc kubenswrapper[4890]: E1125 15:15:39.778977 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34d391804e7af1a69145fcf4e4eda2b537746c2a14e72b9b71fdb9fce45a5f64\": container with ID starting with 34d391804e7af1a69145fcf4e4eda2b537746c2a14e72b9b71fdb9fce45a5f64 not found: ID does not exist" containerID="34d391804e7af1a69145fcf4e4eda2b537746c2a14e72b9b71fdb9fce45a5f64" Nov 25 15:15:39 crc kubenswrapper[4890]: I1125 15:15:39.779026 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34d391804e7af1a69145fcf4e4eda2b537746c2a14e72b9b71fdb9fce45a5f64"} err="failed to get container status \"34d391804e7af1a69145fcf4e4eda2b537746c2a14e72b9b71fdb9fce45a5f64\": rpc error: code = NotFound desc = could not find container \"34d391804e7af1a69145fcf4e4eda2b537746c2a14e72b9b71fdb9fce45a5f64\": container with ID starting with 34d391804e7af1a69145fcf4e4eda2b537746c2a14e72b9b71fdb9fce45a5f64 not found: ID does not exist" Nov 25 15:15:39 crc kubenswrapper[4890]: I1125 15:15:39.779060 4890 scope.go:117] "RemoveContainer" containerID="43f56df7e9635d8afda56f4558cbfc1334f403e9c5098c2472d94f0b72394fd6" Nov 25 15:15:39 crc kubenswrapper[4890]: E1125 15:15:39.779469 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43f56df7e9635d8afda56f4558cbfc1334f403e9c5098c2472d94f0b72394fd6\": container with ID starting with 43f56df7e9635d8afda56f4558cbfc1334f403e9c5098c2472d94f0b72394fd6 not found: ID does not exist" containerID="43f56df7e9635d8afda56f4558cbfc1334f403e9c5098c2472d94f0b72394fd6" Nov 25 15:15:39 crc kubenswrapper[4890]: I1125 15:15:39.779515 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43f56df7e9635d8afda56f4558cbfc1334f403e9c5098c2472d94f0b72394fd6"} err="failed to get container status \"43f56df7e9635d8afda56f4558cbfc1334f403e9c5098c2472d94f0b72394fd6\": rpc error: code = NotFound desc = could not find container \"43f56df7e9635d8afda56f4558cbfc1334f403e9c5098c2472d94f0b72394fd6\": container with ID starting with 43f56df7e9635d8afda56f4558cbfc1334f403e9c5098c2472d94f0b72394fd6 not found: ID does not exist" Nov 25 15:15:39 crc kubenswrapper[4890]: I1125 15:15:39.779540 4890 scope.go:117] "RemoveContainer" containerID="67afd6a182b3cd90bc50c60f35aaf92e4fbbd85de76d773b2c756a91870529b0" Nov 25 15:15:39 crc kubenswrapper[4890]: E1125 15:15:39.779862 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67afd6a182b3cd90bc50c60f35aaf92e4fbbd85de76d773b2c756a91870529b0\": container with ID starting with 67afd6a182b3cd90bc50c60f35aaf92e4fbbd85de76d773b2c756a91870529b0 not found: ID does not exist" containerID="67afd6a182b3cd90bc50c60f35aaf92e4fbbd85de76d773b2c756a91870529b0" Nov 25 15:15:39 crc kubenswrapper[4890]: I1125 15:15:39.779880 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67afd6a182b3cd90bc50c60f35aaf92e4fbbd85de76d773b2c756a91870529b0"} err="failed to get container status \"67afd6a182b3cd90bc50c60f35aaf92e4fbbd85de76d773b2c756a91870529b0\": rpc error: code = NotFound desc = could not find container \"67afd6a182b3cd90bc50c60f35aaf92e4fbbd85de76d773b2c756a91870529b0\": container with ID starting with 67afd6a182b3cd90bc50c60f35aaf92e4fbbd85de76d773b2c756a91870529b0 not found: ID does not exist" Nov 25 15:15:40 crc kubenswrapper[4890]: I1125 15:15:40.186567 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e8357ad-84de-4f9c-9902-7109dabb2dc3" path="/var/lib/kubelet/pods/8e8357ad-84de-4f9c-9902-7109dabb2dc3/volumes" Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.776792 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-86dc4d89c8-pmmf2"] Nov 25 15:16:05 crc kubenswrapper[4890]: E1125 15:16:05.777489 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e8357ad-84de-4f9c-9902-7109dabb2dc3" containerName="registry-server" Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.777501 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e8357ad-84de-4f9c-9902-7109dabb2dc3" containerName="registry-server" Nov 25 15:16:05 crc kubenswrapper[4890]: E1125 15:16:05.777511 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f02bea2-bd7d-4153-9b42-3a898dddea30" containerName="registry-server" Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.777517 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f02bea2-bd7d-4153-9b42-3a898dddea30" containerName="registry-server" Nov 25 15:16:05 crc kubenswrapper[4890]: E1125 15:16:05.777525 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e8357ad-84de-4f9c-9902-7109dabb2dc3" containerName="extract-content" Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.777532 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e8357ad-84de-4f9c-9902-7109dabb2dc3" containerName="extract-content" Nov 25 15:16:05 crc kubenswrapper[4890]: E1125 15:16:05.777541 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e8357ad-84de-4f9c-9902-7109dabb2dc3" containerName="extract-utilities" Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.777547 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e8357ad-84de-4f9c-9902-7109dabb2dc3" containerName="extract-utilities" Nov 25 15:16:05 crc kubenswrapper[4890]: E1125 15:16:05.777557 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f02bea2-bd7d-4153-9b42-3a898dddea30" containerName="extract-content" Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.777562 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f02bea2-bd7d-4153-9b42-3a898dddea30" containerName="extract-content" Nov 25 15:16:05 crc kubenswrapper[4890]: E1125 15:16:05.777578 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f02bea2-bd7d-4153-9b42-3a898dddea30" containerName="extract-utilities" Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.777585 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f02bea2-bd7d-4153-9b42-3a898dddea30" containerName="extract-utilities" Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.777698 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e8357ad-84de-4f9c-9902-7109dabb2dc3" containerName="registry-server" Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.777709 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f02bea2-bd7d-4153-9b42-3a898dddea30" containerName="registry-server" Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.778335 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-pmmf2" Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.781091 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-2dmvv" Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.781854 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-79856dc55c-jcw7x"] Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.783065 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-jcw7x" Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.786054 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-6b6q8" Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.793321 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-79856dc55c-jcw7x"] Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.796670 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-7d695c9b56-stwjm"] Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.797747 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-stwjm" Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.799932 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-86dc4d89c8-pmmf2"] Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.803402 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-nhckl" Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.816185 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-7d695c9b56-stwjm"] Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.839637 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-68b95954c9-lfnwl"] Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.840543 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-lfnwl" Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.847390 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-ldsnl" Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.861389 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-774b86978c-xvfgd"] Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.863179 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-774b86978c-xvfgd" Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.867284 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-68b95954c9-lfnwl"] Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.867321 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-7fwnk" Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.880264 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-774b86978c-xvfgd"] Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.888087 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7sm8v\" (UniqueName: \"kubernetes.io/projected/77a5e8f9-92ab-48b5-be19-29a7a8e5df49-kube-api-access-7sm8v\") pod \"designate-operator-controller-manager-7d695c9b56-stwjm\" (UID: \"77a5e8f9-92ab-48b5-be19-29a7a8e5df49\") " pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-stwjm" Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.888174 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czrgm\" (UniqueName: \"kubernetes.io/projected/72c26c5b-3370-4994-a5f0-82128c62c2df-kube-api-access-czrgm\") pod \"heat-operator-controller-manager-774b86978c-xvfgd\" (UID: \"72c26c5b-3370-4994-a5f0-82128c62c2df\") " pod="openstack-operators/heat-operator-controller-manager-774b86978c-xvfgd" Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.888195 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wssz4\" (UniqueName: \"kubernetes.io/projected/0bdbada3-e08c-4f8a-bd55-bac955d4370d-kube-api-access-wssz4\") pod \"cinder-operator-controller-manager-79856dc55c-jcw7x\" (UID: \"0bdbada3-e08c-4f8a-bd55-bac955d4370d\") " pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-jcw7x" Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.888211 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmwcs\" (UniqueName: \"kubernetes.io/projected/d175a35b-8b9b-49c0-8fc2-02ef62f131d9-kube-api-access-vmwcs\") pod \"barbican-operator-controller-manager-86dc4d89c8-pmmf2\" (UID: \"d175a35b-8b9b-49c0-8fc2-02ef62f131d9\") " pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-pmmf2" Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.888232 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtqfz\" (UniqueName: \"kubernetes.io/projected/076c2a0e-7fdb-42e1-ab73-e296bba2b7a9-kube-api-access-rtqfz\") pod \"glance-operator-controller-manager-68b95954c9-lfnwl\" (UID: \"076c2a0e-7fdb-42e1-ab73-e296bba2b7a9\") " pod="openstack-operators/glance-operator-controller-manager-68b95954c9-lfnwl" Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.903238 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c9694994-j6pqs"] Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.904473 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-j6pqs" Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.911724 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-9nwng" Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.917052 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-d5cc86f4b-stflb"] Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.917986 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-stflb" Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.922429 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c9694994-j6pqs"] Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.925525 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.927223 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-6h9n9" Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.939504 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-d5cc86f4b-stflb"] Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.976076 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5bfcdc958c-hxzf4"] Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.977717 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-hxzf4" Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.982470 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-748dc6576f-rs6hx"] Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.983776 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-rs6hx" Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.984700 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-q5l6w" Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.985616 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-ln68r" Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.989639 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czrgm\" (UniqueName: \"kubernetes.io/projected/72c26c5b-3370-4994-a5f0-82128c62c2df-kube-api-access-czrgm\") pod \"heat-operator-controller-manager-774b86978c-xvfgd\" (UID: \"72c26c5b-3370-4994-a5f0-82128c62c2df\") " pod="openstack-operators/heat-operator-controller-manager-774b86978c-xvfgd" Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.989700 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wssz4\" (UniqueName: \"kubernetes.io/projected/0bdbada3-e08c-4f8a-bd55-bac955d4370d-kube-api-access-wssz4\") pod \"cinder-operator-controller-manager-79856dc55c-jcw7x\" (UID: \"0bdbada3-e08c-4f8a-bd55-bac955d4370d\") " pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-jcw7x" Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.989741 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmwcs\" (UniqueName: \"kubernetes.io/projected/d175a35b-8b9b-49c0-8fc2-02ef62f131d9-kube-api-access-vmwcs\") pod \"barbican-operator-controller-manager-86dc4d89c8-pmmf2\" (UID: \"d175a35b-8b9b-49c0-8fc2-02ef62f131d9\") " pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-pmmf2" Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.989766 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzc25\" (UniqueName: \"kubernetes.io/projected/7e576602-19f5-4656-84dd-5ab80abca895-kube-api-access-lzc25\") pod \"infra-operator-controller-manager-d5cc86f4b-stflb\" (UID: \"7e576602-19f5-4656-84dd-5ab80abca895\") " pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-stflb" Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.989791 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtqfz\" (UniqueName: \"kubernetes.io/projected/076c2a0e-7fdb-42e1-ab73-e296bba2b7a9-kube-api-access-rtqfz\") pod \"glance-operator-controller-manager-68b95954c9-lfnwl\" (UID: \"076c2a0e-7fdb-42e1-ab73-e296bba2b7a9\") " pod="openstack-operators/glance-operator-controller-manager-68b95954c9-lfnwl" Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.989833 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7t5zd\" (UniqueName: \"kubernetes.io/projected/679b3bf3-1ebc-4d19-9ad2-588bd41565dd-kube-api-access-7t5zd\") pod \"ironic-operator-controller-manager-5bfcdc958c-hxzf4\" (UID: \"679b3bf3-1ebc-4d19-9ad2-588bd41565dd\") " pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-hxzf4" Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.989873 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7sm8v\" (UniqueName: \"kubernetes.io/projected/77a5e8f9-92ab-48b5-be19-29a7a8e5df49-kube-api-access-7sm8v\") pod \"designate-operator-controller-manager-7d695c9b56-stwjm\" (UID: \"77a5e8f9-92ab-48b5-be19-29a7a8e5df49\") " pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-stwjm" Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.989915 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9lcx\" (UniqueName: \"kubernetes.io/projected/c528348f-0907-423a-a0c2-f15cf18ff99a-kube-api-access-f9lcx\") pod \"horizon-operator-controller-manager-68c9694994-j6pqs\" (UID: \"c528348f-0907-423a-a0c2-f15cf18ff99a\") " pod="openstack-operators/horizon-operator-controller-manager-68c9694994-j6pqs" Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.989944 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7e576602-19f5-4656-84dd-5ab80abca895-cert\") pod \"infra-operator-controller-manager-d5cc86f4b-stflb\" (UID: \"7e576602-19f5-4656-84dd-5ab80abca895\") " pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-stflb" Nov 25 15:16:05 crc kubenswrapper[4890]: I1125 15:16:05.992095 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5bfcdc958c-hxzf4"] Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.020243 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czrgm\" (UniqueName: \"kubernetes.io/projected/72c26c5b-3370-4994-a5f0-82128c62c2df-kube-api-access-czrgm\") pod \"heat-operator-controller-manager-774b86978c-xvfgd\" (UID: \"72c26c5b-3370-4994-a5f0-82128c62c2df\") " pod="openstack-operators/heat-operator-controller-manager-774b86978c-xvfgd" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.023703 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wssz4\" (UniqueName: \"kubernetes.io/projected/0bdbada3-e08c-4f8a-bd55-bac955d4370d-kube-api-access-wssz4\") pod \"cinder-operator-controller-manager-79856dc55c-jcw7x\" (UID: \"0bdbada3-e08c-4f8a-bd55-bac955d4370d\") " pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-jcw7x" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.034241 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7sm8v\" (UniqueName: \"kubernetes.io/projected/77a5e8f9-92ab-48b5-be19-29a7a8e5df49-kube-api-access-7sm8v\") pod \"designate-operator-controller-manager-7d695c9b56-stwjm\" (UID: \"77a5e8f9-92ab-48b5-be19-29a7a8e5df49\") " pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-stwjm" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.039954 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmwcs\" (UniqueName: \"kubernetes.io/projected/d175a35b-8b9b-49c0-8fc2-02ef62f131d9-kube-api-access-vmwcs\") pod \"barbican-operator-controller-manager-86dc4d89c8-pmmf2\" (UID: \"d175a35b-8b9b-49c0-8fc2-02ef62f131d9\") " pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-pmmf2" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.052269 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-wvjw7"] Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.053527 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-wvjw7" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.054912 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtqfz\" (UniqueName: \"kubernetes.io/projected/076c2a0e-7fdb-42e1-ab73-e296bba2b7a9-kube-api-access-rtqfz\") pod \"glance-operator-controller-manager-68b95954c9-lfnwl\" (UID: \"076c2a0e-7fdb-42e1-ab73-e296bba2b7a9\") " pod="openstack-operators/glance-operator-controller-manager-68b95954c9-lfnwl" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.061201 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-zr9nn" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.079541 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-829nk"] Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.080902 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-829nk" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.083675 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-58bb8d67cc-slpt4"] Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.084778 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-hqgtf" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.086782 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-slpt4" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.089907 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-748dc6576f-rs6hx"] Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.091070 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9lcx\" (UniqueName: \"kubernetes.io/projected/c528348f-0907-423a-a0c2-f15cf18ff99a-kube-api-access-f9lcx\") pod \"horizon-operator-controller-manager-68c9694994-j6pqs\" (UID: \"c528348f-0907-423a-a0c2-f15cf18ff99a\") " pod="openstack-operators/horizon-operator-controller-manager-68c9694994-j6pqs" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.091140 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9snz\" (UniqueName: \"kubernetes.io/projected/03353c8b-91f4-41ae-afc1-c522b1c73bb4-kube-api-access-n9snz\") pod \"mariadb-operator-controller-manager-cb6c4fdb7-wvjw7\" (UID: \"03353c8b-91f4-41ae-afc1-c522b1c73bb4\") " pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-wvjw7" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.091219 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7e576602-19f5-4656-84dd-5ab80abca895-cert\") pod \"infra-operator-controller-manager-d5cc86f4b-stflb\" (UID: \"7e576602-19f5-4656-84dd-5ab80abca895\") " pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-stflb" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.091309 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzc25\" (UniqueName: \"kubernetes.io/projected/7e576602-19f5-4656-84dd-5ab80abca895-kube-api-access-lzc25\") pod \"infra-operator-controller-manager-d5cc86f4b-stflb\" (UID: \"7e576602-19f5-4656-84dd-5ab80abca895\") " pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-stflb" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.091362 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7t5zd\" (UniqueName: \"kubernetes.io/projected/679b3bf3-1ebc-4d19-9ad2-588bd41565dd-kube-api-access-7t5zd\") pod \"ironic-operator-controller-manager-5bfcdc958c-hxzf4\" (UID: \"679b3bf3-1ebc-4d19-9ad2-588bd41565dd\") " pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-hxzf4" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.091394 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2gg7\" (UniqueName: \"kubernetes.io/projected/a5d8a8ba-50c0-4fb6-823f-f28cf0d0f125-kube-api-access-x2gg7\") pod \"keystone-operator-controller-manager-748dc6576f-rs6hx\" (UID: \"a5d8a8ba-50c0-4fb6-823f-f28cf0d0f125\") " pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-rs6hx" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.100067 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-829nk"] Nov 25 15:16:06 crc kubenswrapper[4890]: E1125 15:16:06.104990 4890 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 25 15:16:06 crc kubenswrapper[4890]: E1125 15:16:06.105082 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7e576602-19f5-4656-84dd-5ab80abca895-cert podName:7e576602-19f5-4656-84dd-5ab80abca895 nodeName:}" failed. No retries permitted until 2025-11-25 15:16:06.605065121 +0000 UTC m=+825.047527731 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/7e576602-19f5-4656-84dd-5ab80abca895-cert") pod "infra-operator-controller-manager-d5cc86f4b-stflb" (UID: "7e576602-19f5-4656-84dd-5ab80abca895") : secret "infra-operator-webhook-server-cert" not found Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.107382 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-pmmf2" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.110353 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-cn445" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.116390 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-jcw7x" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.118383 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-wvjw7"] Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.120316 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-stwjm" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.145391 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7t5zd\" (UniqueName: \"kubernetes.io/projected/679b3bf3-1ebc-4d19-9ad2-588bd41565dd-kube-api-access-7t5zd\") pod \"ironic-operator-controller-manager-5bfcdc958c-hxzf4\" (UID: \"679b3bf3-1ebc-4d19-9ad2-588bd41565dd\") " pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-hxzf4" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.157066 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-lfnwl" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.160035 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9lcx\" (UniqueName: \"kubernetes.io/projected/c528348f-0907-423a-a0c2-f15cf18ff99a-kube-api-access-f9lcx\") pod \"horizon-operator-controller-manager-68c9694994-j6pqs\" (UID: \"c528348f-0907-423a-a0c2-f15cf18ff99a\") " pod="openstack-operators/horizon-operator-controller-manager-68c9694994-j6pqs" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.169032 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzc25\" (UniqueName: \"kubernetes.io/projected/7e576602-19f5-4656-84dd-5ab80abca895-kube-api-access-lzc25\") pod \"infra-operator-controller-manager-d5cc86f4b-stflb\" (UID: \"7e576602-19f5-4656-84dd-5ab80abca895\") " pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-stflb" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.194797 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2gg7\" (UniqueName: \"kubernetes.io/projected/a5d8a8ba-50c0-4fb6-823f-f28cf0d0f125-kube-api-access-x2gg7\") pod \"keystone-operator-controller-manager-748dc6576f-rs6hx\" (UID: \"a5d8a8ba-50c0-4fb6-823f-f28cf0d0f125\") " pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-rs6hx" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.206670 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-774b86978c-xvfgd" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.209852 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9snz\" (UniqueName: \"kubernetes.io/projected/03353c8b-91f4-41ae-afc1-c522b1c73bb4-kube-api-access-n9snz\") pod \"mariadb-operator-controller-manager-cb6c4fdb7-wvjw7\" (UID: \"03353c8b-91f4-41ae-afc1-c522b1c73bb4\") " pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-wvjw7" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.222507 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-j6pqs" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.235766 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9snz\" (UniqueName: \"kubernetes.io/projected/03353c8b-91f4-41ae-afc1-c522b1c73bb4-kube-api-access-n9snz\") pod \"mariadb-operator-controller-manager-cb6c4fdb7-wvjw7\" (UID: \"03353c8b-91f4-41ae-afc1-c522b1c73bb4\") " pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-wvjw7" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.237190 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-6jq7q"] Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.238338 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-58bb8d67cc-slpt4"] Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.238357 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-fd75fd47d-56bcr"] Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.239014 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-6jq7q"] Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.239026 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-5lr7r"] Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.242836 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2gg7\" (UniqueName: \"kubernetes.io/projected/a5d8a8ba-50c0-4fb6-823f-f28cf0d0f125-kube-api-access-x2gg7\") pod \"keystone-operator-controller-manager-748dc6576f-rs6hx\" (UID: \"a5d8a8ba-50c0-4fb6-823f-f28cf0d0f125\") " pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-rs6hx" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.243705 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6jq7q" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.243715 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-56bcr" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.244657 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-fd75fd47d-56bcr"] Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.244681 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-5lr7r"] Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.244692 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-66cf5c67ff-vx75t"] Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.245453 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-5db546f9d9-nshw7"] Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.246130 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5db546f9d9-nshw7"] Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.248414 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-5lr7r" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.248990 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-vx75t" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.249264 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-xznbp" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.250494 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.251095 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-xj5wg" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.251310 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-68dvd" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.252003 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-9z5kf" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.253484 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-nshw7" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.256975 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-cvp79" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.261119 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-66cf5c67ff-vx75t"] Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.300861 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-hxzf4" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.312475 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-6fdc4fcf86-dzkgt"] Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.314554 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-rs6hx" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.314681 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-dzkgt" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.315589 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpprx\" (UniqueName: \"kubernetes.io/projected/9f0e3cf9-dcc0-405f-a2de-9148844ec3ae-kube-api-access-vpprx\") pod \"placement-operator-controller-manager-5db546f9d9-nshw7\" (UID: \"9f0e3cf9-dcc0-405f-a2de-9148844ec3ae\") " pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-nshw7" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.315628 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ffnn\" (UniqueName: \"kubernetes.io/projected/f272410f-230d-43f2-946e-e34e4d9c05ea-kube-api-access-9ffnn\") pod \"manila-operator-controller-manager-58bb8d67cc-slpt4\" (UID: \"f272410f-230d-43f2-946e-e34e4d9c05ea\") " pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-slpt4" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.315693 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwkd7\" (UniqueName: \"kubernetes.io/projected/0769cde4-b49f-41c5-b720-e0dbc667eb4a-kube-api-access-nwkd7\") pod \"nova-operator-controller-manager-79556f57fc-6jq7q\" (UID: \"0769cde4-b49f-41c5-b720-e0dbc667eb4a\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6jq7q" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.315734 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6q58\" (UniqueName: \"kubernetes.io/projected/29f250db-2a96-41ad-bb8f-5b503a9288d8-kube-api-access-b6q58\") pod \"ovn-operator-controller-manager-66cf5c67ff-vx75t\" (UID: \"29f250db-2a96-41ad-bb8f-5b503a9288d8\") " pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-vx75t" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.315761 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgbqr\" (UniqueName: \"kubernetes.io/projected/4364eec5-444a-41f6-af1c-fe16b24bd044-kube-api-access-lgbqr\") pod \"neutron-operator-controller-manager-7c57c8bbc4-829nk\" (UID: \"4364eec5-444a-41f6-af1c-fe16b24bd044\") " pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-829nk" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.315786 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jntg\" (UniqueName: \"kubernetes.io/projected/10534635-efe1-4e20-8c9b-9cbea73fbe87-kube-api-access-9jntg\") pod \"openstack-baremetal-operator-controller-manager-544b9bb9-5lr7r\" (UID: \"10534635-efe1-4e20-8c9b-9cbea73fbe87\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-5lr7r" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.315834 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/10534635-efe1-4e20-8c9b-9cbea73fbe87-cert\") pod \"openstack-baremetal-operator-controller-manager-544b9bb9-5lr7r\" (UID: \"10534635-efe1-4e20-8c9b-9cbea73fbe87\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-5lr7r" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.315870 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n88zf\" (UniqueName: \"kubernetes.io/projected/a9efeabb-bf1c-4de2-8322-961b284e95cc-kube-api-access-n88zf\") pod \"octavia-operator-controller-manager-fd75fd47d-56bcr\" (UID: \"a9efeabb-bf1c-4de2-8322-961b284e95cc\") " pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-56bcr" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.317443 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-fdhzs" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.355404 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6fdc4fcf86-dzkgt"] Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.383318 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-567f98c9d-6zr4x"] Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.384482 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-6zr4x" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.406891 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-ncrf8" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.416612 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/10534635-efe1-4e20-8c9b-9cbea73fbe87-cert\") pod \"openstack-baremetal-operator-controller-manager-544b9bb9-5lr7r\" (UID: \"10534635-efe1-4e20-8c9b-9cbea73fbe87\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-5lr7r" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.416650 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n88zf\" (UniqueName: \"kubernetes.io/projected/a9efeabb-bf1c-4de2-8322-961b284e95cc-kube-api-access-n88zf\") pod \"octavia-operator-controller-manager-fd75fd47d-56bcr\" (UID: \"a9efeabb-bf1c-4de2-8322-961b284e95cc\") " pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-56bcr" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.416681 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpprx\" (UniqueName: \"kubernetes.io/projected/9f0e3cf9-dcc0-405f-a2de-9148844ec3ae-kube-api-access-vpprx\") pod \"placement-operator-controller-manager-5db546f9d9-nshw7\" (UID: \"9f0e3cf9-dcc0-405f-a2de-9148844ec3ae\") " pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-nshw7" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.416698 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ffnn\" (UniqueName: \"kubernetes.io/projected/f272410f-230d-43f2-946e-e34e4d9c05ea-kube-api-access-9ffnn\") pod \"manila-operator-controller-manager-58bb8d67cc-slpt4\" (UID: \"f272410f-230d-43f2-946e-e34e4d9c05ea\") " pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-slpt4" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.416725 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tl2qz\" (UniqueName: \"kubernetes.io/projected/d50aec9d-1673-4b81-bb57-04cae418843e-kube-api-access-tl2qz\") pod \"telemetry-operator-controller-manager-567f98c9d-6zr4x\" (UID: \"d50aec9d-1673-4b81-bb57-04cae418843e\") " pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-6zr4x" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.416757 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w29mm\" (UniqueName: \"kubernetes.io/projected/861317e1-222d-4f14-b931-e9e7d100ebd6-kube-api-access-w29mm\") pod \"swift-operator-controller-manager-6fdc4fcf86-dzkgt\" (UID: \"861317e1-222d-4f14-b931-e9e7d100ebd6\") " pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-dzkgt" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.416778 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwkd7\" (UniqueName: \"kubernetes.io/projected/0769cde4-b49f-41c5-b720-e0dbc667eb4a-kube-api-access-nwkd7\") pod \"nova-operator-controller-manager-79556f57fc-6jq7q\" (UID: \"0769cde4-b49f-41c5-b720-e0dbc667eb4a\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6jq7q" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.416807 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6q58\" (UniqueName: \"kubernetes.io/projected/29f250db-2a96-41ad-bb8f-5b503a9288d8-kube-api-access-b6q58\") pod \"ovn-operator-controller-manager-66cf5c67ff-vx75t\" (UID: \"29f250db-2a96-41ad-bb8f-5b503a9288d8\") " pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-vx75t" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.416823 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgbqr\" (UniqueName: \"kubernetes.io/projected/4364eec5-444a-41f6-af1c-fe16b24bd044-kube-api-access-lgbqr\") pod \"neutron-operator-controller-manager-7c57c8bbc4-829nk\" (UID: \"4364eec5-444a-41f6-af1c-fe16b24bd044\") " pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-829nk" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.416841 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jntg\" (UniqueName: \"kubernetes.io/projected/10534635-efe1-4e20-8c9b-9cbea73fbe87-kube-api-access-9jntg\") pod \"openstack-baremetal-operator-controller-manager-544b9bb9-5lr7r\" (UID: \"10534635-efe1-4e20-8c9b-9cbea73fbe87\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-5lr7r" Nov 25 15:16:06 crc kubenswrapper[4890]: E1125 15:16:06.417443 4890 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 25 15:16:06 crc kubenswrapper[4890]: E1125 15:16:06.417481 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/10534635-efe1-4e20-8c9b-9cbea73fbe87-cert podName:10534635-efe1-4e20-8c9b-9cbea73fbe87 nodeName:}" failed. No retries permitted until 2025-11-25 15:16:06.917468295 +0000 UTC m=+825.359930905 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/10534635-efe1-4e20-8c9b-9cbea73fbe87-cert") pod "openstack-baremetal-operator-controller-manager-544b9bb9-5lr7r" (UID: "10534635-efe1-4e20-8c9b-9cbea73fbe87") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.420527 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-567f98c9d-6zr4x"] Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.436176 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-wvjw7" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.460869 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpprx\" (UniqueName: \"kubernetes.io/projected/9f0e3cf9-dcc0-405f-a2de-9148844ec3ae-kube-api-access-vpprx\") pod \"placement-operator-controller-manager-5db546f9d9-nshw7\" (UID: \"9f0e3cf9-dcc0-405f-a2de-9148844ec3ae\") " pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-nshw7" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.478915 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6q58\" (UniqueName: \"kubernetes.io/projected/29f250db-2a96-41ad-bb8f-5b503a9288d8-kube-api-access-b6q58\") pod \"ovn-operator-controller-manager-66cf5c67ff-vx75t\" (UID: \"29f250db-2a96-41ad-bb8f-5b503a9288d8\") " pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-vx75t" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.487806 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ffnn\" (UniqueName: \"kubernetes.io/projected/f272410f-230d-43f2-946e-e34e4d9c05ea-kube-api-access-9ffnn\") pod \"manila-operator-controller-manager-58bb8d67cc-slpt4\" (UID: \"f272410f-230d-43f2-946e-e34e4d9c05ea\") " pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-slpt4" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.491767 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n88zf\" (UniqueName: \"kubernetes.io/projected/a9efeabb-bf1c-4de2-8322-961b284e95cc-kube-api-access-n88zf\") pod \"octavia-operator-controller-manager-fd75fd47d-56bcr\" (UID: \"a9efeabb-bf1c-4de2-8322-961b284e95cc\") " pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-56bcr" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.495411 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jntg\" (UniqueName: \"kubernetes.io/projected/10534635-efe1-4e20-8c9b-9cbea73fbe87-kube-api-access-9jntg\") pod \"openstack-baremetal-operator-controller-manager-544b9bb9-5lr7r\" (UID: \"10534635-efe1-4e20-8c9b-9cbea73fbe87\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-5lr7r" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.514689 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-slpt4" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.518853 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tl2qz\" (UniqueName: \"kubernetes.io/projected/d50aec9d-1673-4b81-bb57-04cae418843e-kube-api-access-tl2qz\") pod \"telemetry-operator-controller-manager-567f98c9d-6zr4x\" (UID: \"d50aec9d-1673-4b81-bb57-04cae418843e\") " pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-6zr4x" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.518908 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w29mm\" (UniqueName: \"kubernetes.io/projected/861317e1-222d-4f14-b931-e9e7d100ebd6-kube-api-access-w29mm\") pod \"swift-operator-controller-manager-6fdc4fcf86-dzkgt\" (UID: \"861317e1-222d-4f14-b931-e9e7d100ebd6\") " pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-dzkgt" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.535312 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5cb74df96-6mqt2"] Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.536423 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cb74df96-6mqt2" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.544081 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-g5vrc" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.549044 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwkd7\" (UniqueName: \"kubernetes.io/projected/0769cde4-b49f-41c5-b720-e0dbc667eb4a-kube-api-access-nwkd7\") pod \"nova-operator-controller-manager-79556f57fc-6jq7q\" (UID: \"0769cde4-b49f-41c5-b720-e0dbc667eb4a\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6jq7q" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.554718 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgbqr\" (UniqueName: \"kubernetes.io/projected/4364eec5-444a-41f6-af1c-fe16b24bd044-kube-api-access-lgbqr\") pod \"neutron-operator-controller-manager-7c57c8bbc4-829nk\" (UID: \"4364eec5-444a-41f6-af1c-fe16b24bd044\") " pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-829nk" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.557915 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w29mm\" (UniqueName: \"kubernetes.io/projected/861317e1-222d-4f14-b931-e9e7d100ebd6-kube-api-access-w29mm\") pod \"swift-operator-controller-manager-6fdc4fcf86-dzkgt\" (UID: \"861317e1-222d-4f14-b931-e9e7d100ebd6\") " pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-dzkgt" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.558729 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tl2qz\" (UniqueName: \"kubernetes.io/projected/d50aec9d-1673-4b81-bb57-04cae418843e-kube-api-access-tl2qz\") pod \"telemetry-operator-controller-manager-567f98c9d-6zr4x\" (UID: \"d50aec9d-1673-4b81-bb57-04cae418843e\") " pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-6zr4x" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.566845 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-6zr4x" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.571132 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cb74df96-6mqt2"] Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.609139 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-864885998-62ddz"] Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.611329 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-864885998-62ddz" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.612221 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6jq7q" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.621993 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-s7rlv" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.624011 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7e576602-19f5-4656-84dd-5ab80abca895-cert\") pod \"infra-operator-controller-manager-d5cc86f4b-stflb\" (UID: \"7e576602-19f5-4656-84dd-5ab80abca895\") " pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-stflb" Nov 25 15:16:06 crc kubenswrapper[4890]: E1125 15:16:06.624257 4890 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 25 15:16:06 crc kubenswrapper[4890]: E1125 15:16:06.624314 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7e576602-19f5-4656-84dd-5ab80abca895-cert podName:7e576602-19f5-4656-84dd-5ab80abca895 nodeName:}" failed. No retries permitted until 2025-11-25 15:16:07.624296323 +0000 UTC m=+826.066758933 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/7e576602-19f5-4656-84dd-5ab80abca895-cert") pod "infra-operator-controller-manager-d5cc86f4b-stflb" (UID: "7e576602-19f5-4656-84dd-5ab80abca895") : secret "infra-operator-webhook-server-cert" not found Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.632913 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-864885998-62ddz"] Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.647556 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-56bcr" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.718092 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-vx75t" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.720261 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-68587559f4-9b4cf"] Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.721779 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-68587559f4-9b4cf" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.734819 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.735936 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-44bfh" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.736092 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.755138 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3e496c3d-9485-44bc-b3b1-dc1682a88434-webhook-certs\") pod \"openstack-operator-controller-manager-68587559f4-9b4cf\" (UID: \"3e496c3d-9485-44bc-b3b1-dc1682a88434\") " pod="openstack-operators/openstack-operator-controller-manager-68587559f4-9b4cf" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.755216 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w22cz\" (UniqueName: \"kubernetes.io/projected/3e496c3d-9485-44bc-b3b1-dc1682a88434-kube-api-access-w22cz\") pod \"openstack-operator-controller-manager-68587559f4-9b4cf\" (UID: \"3e496c3d-9485-44bc-b3b1-dc1682a88434\") " pod="openstack-operators/openstack-operator-controller-manager-68587559f4-9b4cf" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.755248 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqp5s\" (UniqueName: \"kubernetes.io/projected/9637fded-765d-4037-be8f-82e59ffdfc14-kube-api-access-wqp5s\") pod \"test-operator-controller-manager-5cb74df96-6mqt2\" (UID: \"9637fded-765d-4037-be8f-82e59ffdfc14\") " pod="openstack-operators/test-operator-controller-manager-5cb74df96-6mqt2" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.755294 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5r2pb\" (UniqueName: \"kubernetes.io/projected/a9a69828-3f0d-4d55-a840-10c2c791444c-kube-api-access-5r2pb\") pod \"watcher-operator-controller-manager-864885998-62ddz\" (UID: \"a9a69828-3f0d-4d55-a840-10c2c791444c\") " pod="openstack-operators/watcher-operator-controller-manager-864885998-62ddz" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.755821 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-nshw7" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.758751 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3e496c3d-9485-44bc-b3b1-dc1682a88434-metrics-certs\") pod \"openstack-operator-controller-manager-68587559f4-9b4cf\" (UID: \"3e496c3d-9485-44bc-b3b1-dc1682a88434\") " pod="openstack-operators/openstack-operator-controller-manager-68587559f4-9b4cf" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.761061 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-68587559f4-9b4cf"] Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.774372 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-dzkgt" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.789362 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-829nk" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.798307 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-dmlx4"] Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.799561 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-dmlx4" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.802708 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-xngkr" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.805935 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-dmlx4"] Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.864017 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swsbh\" (UniqueName: \"kubernetes.io/projected/e26594ec-4bc2-4051-8a49-fc490bd3b6c9-kube-api-access-swsbh\") pod \"rabbitmq-cluster-operator-manager-668c99d594-dmlx4\" (UID: \"e26594ec-4bc2-4051-8a49-fc490bd3b6c9\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-dmlx4" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.864081 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3e496c3d-9485-44bc-b3b1-dc1682a88434-webhook-certs\") pod \"openstack-operator-controller-manager-68587559f4-9b4cf\" (UID: \"3e496c3d-9485-44bc-b3b1-dc1682a88434\") " pod="openstack-operators/openstack-operator-controller-manager-68587559f4-9b4cf" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.864106 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w22cz\" (UniqueName: \"kubernetes.io/projected/3e496c3d-9485-44bc-b3b1-dc1682a88434-kube-api-access-w22cz\") pod \"openstack-operator-controller-manager-68587559f4-9b4cf\" (UID: \"3e496c3d-9485-44bc-b3b1-dc1682a88434\") " pod="openstack-operators/openstack-operator-controller-manager-68587559f4-9b4cf" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.864148 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqp5s\" (UniqueName: \"kubernetes.io/projected/9637fded-765d-4037-be8f-82e59ffdfc14-kube-api-access-wqp5s\") pod \"test-operator-controller-manager-5cb74df96-6mqt2\" (UID: \"9637fded-765d-4037-be8f-82e59ffdfc14\") " pod="openstack-operators/test-operator-controller-manager-5cb74df96-6mqt2" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.864205 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5r2pb\" (UniqueName: \"kubernetes.io/projected/a9a69828-3f0d-4d55-a840-10c2c791444c-kube-api-access-5r2pb\") pod \"watcher-operator-controller-manager-864885998-62ddz\" (UID: \"a9a69828-3f0d-4d55-a840-10c2c791444c\") " pod="openstack-operators/watcher-operator-controller-manager-864885998-62ddz" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.864227 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3e496c3d-9485-44bc-b3b1-dc1682a88434-metrics-certs\") pod \"openstack-operator-controller-manager-68587559f4-9b4cf\" (UID: \"3e496c3d-9485-44bc-b3b1-dc1682a88434\") " pod="openstack-operators/openstack-operator-controller-manager-68587559f4-9b4cf" Nov 25 15:16:06 crc kubenswrapper[4890]: E1125 15:16:06.864340 4890 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 25 15:16:06 crc kubenswrapper[4890]: E1125 15:16:06.864345 4890 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 25 15:16:06 crc kubenswrapper[4890]: E1125 15:16:06.864378 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3e496c3d-9485-44bc-b3b1-dc1682a88434-metrics-certs podName:3e496c3d-9485-44bc-b3b1-dc1682a88434 nodeName:}" failed. No retries permitted until 2025-11-25 15:16:07.364365041 +0000 UTC m=+825.806827641 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3e496c3d-9485-44bc-b3b1-dc1682a88434-metrics-certs") pod "openstack-operator-controller-manager-68587559f4-9b4cf" (UID: "3e496c3d-9485-44bc-b3b1-dc1682a88434") : secret "metrics-server-cert" not found Nov 25 15:16:06 crc kubenswrapper[4890]: E1125 15:16:06.864432 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3e496c3d-9485-44bc-b3b1-dc1682a88434-webhook-certs podName:3e496c3d-9485-44bc-b3b1-dc1682a88434 nodeName:}" failed. No retries permitted until 2025-11-25 15:16:07.364410312 +0000 UTC m=+825.806872922 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/3e496c3d-9485-44bc-b3b1-dc1682a88434-webhook-certs") pod "openstack-operator-controller-manager-68587559f4-9b4cf" (UID: "3e496c3d-9485-44bc-b3b1-dc1682a88434") : secret "webhook-server-cert" not found Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.885797 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqp5s\" (UniqueName: \"kubernetes.io/projected/9637fded-765d-4037-be8f-82e59ffdfc14-kube-api-access-wqp5s\") pod \"test-operator-controller-manager-5cb74df96-6mqt2\" (UID: \"9637fded-765d-4037-be8f-82e59ffdfc14\") " pod="openstack-operators/test-operator-controller-manager-5cb74df96-6mqt2" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.885931 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w22cz\" (UniqueName: \"kubernetes.io/projected/3e496c3d-9485-44bc-b3b1-dc1682a88434-kube-api-access-w22cz\") pod \"openstack-operator-controller-manager-68587559f4-9b4cf\" (UID: \"3e496c3d-9485-44bc-b3b1-dc1682a88434\") " pod="openstack-operators/openstack-operator-controller-manager-68587559f4-9b4cf" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.886080 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5r2pb\" (UniqueName: \"kubernetes.io/projected/a9a69828-3f0d-4d55-a840-10c2c791444c-kube-api-access-5r2pb\") pod \"watcher-operator-controller-manager-864885998-62ddz\" (UID: \"a9a69828-3f0d-4d55-a840-10c2c791444c\") " pod="openstack-operators/watcher-operator-controller-manager-864885998-62ddz" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.949549 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cb74df96-6mqt2" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.965102 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/10534635-efe1-4e20-8c9b-9cbea73fbe87-cert\") pod \"openstack-baremetal-operator-controller-manager-544b9bb9-5lr7r\" (UID: \"10534635-efe1-4e20-8c9b-9cbea73fbe87\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-5lr7r" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.965155 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swsbh\" (UniqueName: \"kubernetes.io/projected/e26594ec-4bc2-4051-8a49-fc490bd3b6c9-kube-api-access-swsbh\") pod \"rabbitmq-cluster-operator-manager-668c99d594-dmlx4\" (UID: \"e26594ec-4bc2-4051-8a49-fc490bd3b6c9\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-dmlx4" Nov 25 15:16:06 crc kubenswrapper[4890]: E1125 15:16:06.965768 4890 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 25 15:16:06 crc kubenswrapper[4890]: E1125 15:16:06.965850 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/10534635-efe1-4e20-8c9b-9cbea73fbe87-cert podName:10534635-efe1-4e20-8c9b-9cbea73fbe87 nodeName:}" failed. No retries permitted until 2025-11-25 15:16:07.965830846 +0000 UTC m=+826.408293456 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/10534635-efe1-4e20-8c9b-9cbea73fbe87-cert") pod "openstack-baremetal-operator-controller-manager-544b9bb9-5lr7r" (UID: "10534635-efe1-4e20-8c9b-9cbea73fbe87") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.980907 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-864885998-62ddz" Nov 25 15:16:06 crc kubenswrapper[4890]: I1125 15:16:06.997273 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swsbh\" (UniqueName: \"kubernetes.io/projected/e26594ec-4bc2-4051-8a49-fc490bd3b6c9-kube-api-access-swsbh\") pod \"rabbitmq-cluster-operator-manager-668c99d594-dmlx4\" (UID: \"e26594ec-4bc2-4051-8a49-fc490bd3b6c9\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-dmlx4" Nov 25 15:16:07 crc kubenswrapper[4890]: I1125 15:16:07.031911 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-7d695c9b56-stwjm"] Nov 25 15:16:07 crc kubenswrapper[4890]: I1125 15:16:07.046366 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-86dc4d89c8-pmmf2"] Nov 25 15:16:07 crc kubenswrapper[4890]: I1125 15:16:07.053219 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-79856dc55c-jcw7x"] Nov 25 15:16:07 crc kubenswrapper[4890]: W1125 15:16:07.137408 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0bdbada3_e08c_4f8a_bd55_bac955d4370d.slice/crio-7c54f8baa66686ab63ed1820820b472455892f5ce9f1a7ea7886d8350ff77523 WatchSource:0}: Error finding container 7c54f8baa66686ab63ed1820820b472455892f5ce9f1a7ea7886d8350ff77523: Status 404 returned error can't find the container with id 7c54f8baa66686ab63ed1820820b472455892f5ce9f1a7ea7886d8350ff77523 Nov 25 15:16:07 crc kubenswrapper[4890]: I1125 15:16:07.190154 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-dmlx4" Nov 25 15:16:07 crc kubenswrapper[4890]: I1125 15:16:07.282294 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-774b86978c-xvfgd"] Nov 25 15:16:07 crc kubenswrapper[4890]: I1125 15:16:07.289016 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-68b95954c9-lfnwl"] Nov 25 15:16:07 crc kubenswrapper[4890]: I1125 15:16:07.369953 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3e496c3d-9485-44bc-b3b1-dc1682a88434-metrics-certs\") pod \"openstack-operator-controller-manager-68587559f4-9b4cf\" (UID: \"3e496c3d-9485-44bc-b3b1-dc1682a88434\") " pod="openstack-operators/openstack-operator-controller-manager-68587559f4-9b4cf" Nov 25 15:16:07 crc kubenswrapper[4890]: I1125 15:16:07.370151 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3e496c3d-9485-44bc-b3b1-dc1682a88434-webhook-certs\") pod \"openstack-operator-controller-manager-68587559f4-9b4cf\" (UID: \"3e496c3d-9485-44bc-b3b1-dc1682a88434\") " pod="openstack-operators/openstack-operator-controller-manager-68587559f4-9b4cf" Nov 25 15:16:07 crc kubenswrapper[4890]: E1125 15:16:07.370317 4890 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 25 15:16:07 crc kubenswrapper[4890]: E1125 15:16:07.370383 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3e496c3d-9485-44bc-b3b1-dc1682a88434-webhook-certs podName:3e496c3d-9485-44bc-b3b1-dc1682a88434 nodeName:}" failed. No retries permitted until 2025-11-25 15:16:08.370365644 +0000 UTC m=+826.812828254 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/3e496c3d-9485-44bc-b3b1-dc1682a88434-webhook-certs") pod "openstack-operator-controller-manager-68587559f4-9b4cf" (UID: "3e496c3d-9485-44bc-b3b1-dc1682a88434") : secret "webhook-server-cert" not found Nov 25 15:16:07 crc kubenswrapper[4890]: E1125 15:16:07.370614 4890 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 25 15:16:07 crc kubenswrapper[4890]: E1125 15:16:07.370680 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3e496c3d-9485-44bc-b3b1-dc1682a88434-metrics-certs podName:3e496c3d-9485-44bc-b3b1-dc1682a88434 nodeName:}" failed. No retries permitted until 2025-11-25 15:16:08.370669911 +0000 UTC m=+826.813132521 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3e496c3d-9485-44bc-b3b1-dc1682a88434-metrics-certs") pod "openstack-operator-controller-manager-68587559f4-9b4cf" (UID: "3e496c3d-9485-44bc-b3b1-dc1682a88434") : secret "metrics-server-cert" not found Nov 25 15:16:07 crc kubenswrapper[4890]: I1125 15:16:07.471289 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5bfcdc958c-hxzf4"] Nov 25 15:16:07 crc kubenswrapper[4890]: W1125 15:16:07.480234 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod679b3bf3_1ebc_4d19_9ad2_588bd41565dd.slice/crio-2f1c6eb5fa46b8036c0fb9ad4bb86ea98ec4e04723b249cab34c2a96041fcc67 WatchSource:0}: Error finding container 2f1c6eb5fa46b8036c0fb9ad4bb86ea98ec4e04723b249cab34c2a96041fcc67: Status 404 returned error can't find the container with id 2f1c6eb5fa46b8036c0fb9ad4bb86ea98ec4e04723b249cab34c2a96041fcc67 Nov 25 15:16:07 crc kubenswrapper[4890]: I1125 15:16:07.548129 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-748dc6576f-rs6hx"] Nov 25 15:16:07 crc kubenswrapper[4890]: I1125 15:16:07.572295 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c9694994-j6pqs"] Nov 25 15:16:07 crc kubenswrapper[4890]: I1125 15:16:07.576630 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-567f98c9d-6zr4x"] Nov 25 15:16:07 crc kubenswrapper[4890]: I1125 15:16:07.593106 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-58bb8d67cc-slpt4"] Nov 25 15:16:07 crc kubenswrapper[4890]: W1125 15:16:07.596939 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4364eec5_444a_41f6_af1c_fe16b24bd044.slice/crio-9a20bac8ac4f6e4dd5af7374d21f2407afa170dfdabe8fe9012b4cb0e4c17483 WatchSource:0}: Error finding container 9a20bac8ac4f6e4dd5af7374d21f2407afa170dfdabe8fe9012b4cb0e4c17483: Status 404 returned error can't find the container with id 9a20bac8ac4f6e4dd5af7374d21f2407afa170dfdabe8fe9012b4cb0e4c17483 Nov 25 15:16:07 crc kubenswrapper[4890]: W1125 15:16:07.599707 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9f0e3cf9_dcc0_405f_a2de_9148844ec3ae.slice/crio-0b10fa541f7f344027f68d6088079ce4cef8b68881cbf4e00aae7ba981eda3ce WatchSource:0}: Error finding container 0b10fa541f7f344027f68d6088079ce4cef8b68881cbf4e00aae7ba981eda3ce: Status 404 returned error can't find the container with id 0b10fa541f7f344027f68d6088079ce4cef8b68881cbf4e00aae7ba981eda3ce Nov 25 15:16:07 crc kubenswrapper[4890]: I1125 15:16:07.602249 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5db546f9d9-nshw7"] Nov 25 15:16:07 crc kubenswrapper[4890]: W1125 15:16:07.602948 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod861317e1_222d_4f14_b931_e9e7d100ebd6.slice/crio-da5811753bd8ae8bead057521cdbff92ac04632178536d0e42d3c8755068ba8e WatchSource:0}: Error finding container da5811753bd8ae8bead057521cdbff92ac04632178536d0e42d3c8755068ba8e: Status 404 returned error can't find the container with id da5811753bd8ae8bead057521cdbff92ac04632178536d0e42d3c8755068ba8e Nov 25 15:16:07 crc kubenswrapper[4890]: E1125 15:16:07.605577 4890 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-w29mm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-6fdc4fcf86-dzkgt_openstack-operators(861317e1-222d-4f14-b931-e9e7d100ebd6): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 15:16:07 crc kubenswrapper[4890]: E1125 15:16:07.605740 4890 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:7b90521b9e9cb4eb43c2f1c3bf85dbd068d684315f4f705b07708dd078df9d04,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-n9snz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-cb6c4fdb7-wvjw7_openstack-operators(03353c8b-91f4-41ae-afc1-c522b1c73bb4): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 15:16:07 crc kubenswrapper[4890]: E1125 15:16:07.607523 4890 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-n9snz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-cb6c4fdb7-wvjw7_openstack-operators(03353c8b-91f4-41ae-afc1-c522b1c73bb4): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 15:16:07 crc kubenswrapper[4890]: E1125 15:16:07.607635 4890 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-w29mm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-6fdc4fcf86-dzkgt_openstack-operators(861317e1-222d-4f14-b931-e9e7d100ebd6): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 15:16:07 crc kubenswrapper[4890]: I1125 15:16:07.608020 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6fdc4fcf86-dzkgt"] Nov 25 15:16:07 crc kubenswrapper[4890]: E1125 15:16:07.608600 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-wvjw7" podUID="03353c8b-91f4-41ae-afc1-c522b1c73bb4" Nov 25 15:16:07 crc kubenswrapper[4890]: E1125 15:16:07.608678 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-dzkgt" podUID="861317e1-222d-4f14-b931-e9e7d100ebd6" Nov 25 15:16:07 crc kubenswrapper[4890]: I1125 15:16:07.614150 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-829nk"] Nov 25 15:16:07 crc kubenswrapper[4890]: I1125 15:16:07.621485 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-wvjw7"] Nov 25 15:16:07 crc kubenswrapper[4890]: I1125 15:16:07.673132 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7e576602-19f5-4656-84dd-5ab80abca895-cert\") pod \"infra-operator-controller-manager-d5cc86f4b-stflb\" (UID: \"7e576602-19f5-4656-84dd-5ab80abca895\") " pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-stflb" Nov 25 15:16:07 crc kubenswrapper[4890]: I1125 15:16:07.679730 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7e576602-19f5-4656-84dd-5ab80abca895-cert\") pod \"infra-operator-controller-manager-d5cc86f4b-stflb\" (UID: \"7e576602-19f5-4656-84dd-5ab80abca895\") " pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-stflb" Nov 25 15:16:07 crc kubenswrapper[4890]: I1125 15:16:07.726412 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-66cf5c67ff-vx75t"] Nov 25 15:16:07 crc kubenswrapper[4890]: W1125 15:16:07.728968 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod29f250db_2a96_41ad_bb8f_5b503a9288d8.slice/crio-c1e1bd0f8a7e6167aa2d0f66fa944267684c56e2f506a8d9e00094c50d1a4de3 WatchSource:0}: Error finding container c1e1bd0f8a7e6167aa2d0f66fa944267684c56e2f506a8d9e00094c50d1a4de3: Status 404 returned error can't find the container with id c1e1bd0f8a7e6167aa2d0f66fa944267684c56e2f506a8d9e00094c50d1a4de3 Nov 25 15:16:07 crc kubenswrapper[4890]: I1125 15:16:07.732281 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-stflb" Nov 25 15:16:07 crc kubenswrapper[4890]: I1125 15:16:07.756979 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-dmlx4"] Nov 25 15:16:07 crc kubenswrapper[4890]: I1125 15:16:07.761254 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-fd75fd47d-56bcr"] Nov 25 15:16:07 crc kubenswrapper[4890]: E1125 15:16:07.777864 4890 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-swsbh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-dmlx4_openstack-operators(e26594ec-4bc2-4051-8a49-fc490bd3b6c9): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 15:16:07 crc kubenswrapper[4890]: W1125 15:16:07.778244 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda9efeabb_bf1c_4de2_8322_961b284e95cc.slice/crio-09ed8bd82ba1777b270c9e9089ea2a64dfd007d6f14117b249ae4124b8324224 WatchSource:0}: Error finding container 09ed8bd82ba1777b270c9e9089ea2a64dfd007d6f14117b249ae4124b8324224: Status 404 returned error can't find the container with id 09ed8bd82ba1777b270c9e9089ea2a64dfd007d6f14117b249ae4124b8324224 Nov 25 15:16:07 crc kubenswrapper[4890]: E1125 15:16:07.779534 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-dmlx4" podUID="e26594ec-4bc2-4051-8a49-fc490bd3b6c9" Nov 25 15:16:07 crc kubenswrapper[4890]: I1125 15:16:07.781454 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-6jq7q"] Nov 25 15:16:07 crc kubenswrapper[4890]: W1125 15:16:07.786670 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0769cde4_b49f_41c5_b720_e0dbc667eb4a.slice/crio-f0ffd78f393061bb26f75df64b55df422e06c632057fb08f001463c871fd1d51 WatchSource:0}: Error finding container f0ffd78f393061bb26f75df64b55df422e06c632057fb08f001463c871fd1d51: Status 404 returned error can't find the container with id f0ffd78f393061bb26f75df64b55df422e06c632057fb08f001463c871fd1d51 Nov 25 15:16:07 crc kubenswrapper[4890]: I1125 15:16:07.788498 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cb74df96-6mqt2"] Nov 25 15:16:07 crc kubenswrapper[4890]: W1125 15:16:07.788676 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9637fded_765d_4037_be8f_82e59ffdfc14.slice/crio-9adffdc7cc493f7e2c9057e90549de041fa22f51bf98de2d3c2e67e2d18afbb0 WatchSource:0}: Error finding container 9adffdc7cc493f7e2c9057e90549de041fa22f51bf98de2d3c2e67e2d18afbb0: Status 404 returned error can't find the container with id 9adffdc7cc493f7e2c9057e90549de041fa22f51bf98de2d3c2e67e2d18afbb0 Nov 25 15:16:07 crc kubenswrapper[4890]: E1125 15:16:07.790983 4890 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nwkd7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-79556f57fc-6jq7q_openstack-operators(0769cde4-b49f-41c5-b720-e0dbc667eb4a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 15:16:07 crc kubenswrapper[4890]: W1125 15:16:07.791449 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda9a69828_3f0d_4d55_a840_10c2c791444c.slice/crio-eb69fabbc8d6df7c8ce3c8802db231e31fd300c5dc20786c185d467985247195 WatchSource:0}: Error finding container eb69fabbc8d6df7c8ce3c8802db231e31fd300c5dc20786c185d467985247195: Status 404 returned error can't find the container with id eb69fabbc8d6df7c8ce3c8802db231e31fd300c5dc20786c185d467985247195 Nov 25 15:16:07 crc kubenswrapper[4890]: E1125 15:16:07.791651 4890 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wqp5s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5cb74df96-6mqt2_openstack-operators(9637fded-765d-4037-be8f-82e59ffdfc14): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 15:16:07 crc kubenswrapper[4890]: E1125 15:16:07.795006 4890 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:442c269d79163f8da75505019c02e9f0815837aaadcaddacb8e6c12df297ca13,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-n88zf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-fd75fd47d-56bcr_openstack-operators(a9efeabb-bf1c-4de2-8322-961b284e95cc): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 15:16:07 crc kubenswrapper[4890]: I1125 15:16:07.798425 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-864885998-62ddz"] Nov 25 15:16:07 crc kubenswrapper[4890]: E1125 15:16:07.800426 4890 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-n88zf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-fd75fd47d-56bcr_openstack-operators(a9efeabb-bf1c-4de2-8322-961b284e95cc): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 15:16:07 crc kubenswrapper[4890]: E1125 15:16:07.800530 4890 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nwkd7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-79556f57fc-6jq7q_openstack-operators(0769cde4-b49f-41c5-b720-e0dbc667eb4a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 15:16:07 crc kubenswrapper[4890]: E1125 15:16:07.800647 4890 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wqp5s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5cb74df96-6mqt2_openstack-operators(9637fded-765d-4037-be8f-82e59ffdfc14): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 15:16:07 crc kubenswrapper[4890]: E1125 15:16:07.800756 4890 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5r2pb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-864885998-62ddz_openstack-operators(a9a69828-3f0d-4d55-a840-10c2c791444c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 15:16:07 crc kubenswrapper[4890]: E1125 15:16:07.802298 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5cb74df96-6mqt2" podUID="9637fded-765d-4037-be8f-82e59ffdfc14" Nov 25 15:16:07 crc kubenswrapper[4890]: E1125 15:16:07.802312 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6jq7q" podUID="0769cde4-b49f-41c5-b720-e0dbc667eb4a" Nov 25 15:16:07 crc kubenswrapper[4890]: E1125 15:16:07.802351 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-56bcr" podUID="a9efeabb-bf1c-4de2-8322-961b284e95cc" Nov 25 15:16:07 crc kubenswrapper[4890]: E1125 15:16:07.803307 4890 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5r2pb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-864885998-62ddz_openstack-operators(a9a69828-3f0d-4d55-a840-10c2c791444c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 15:16:07 crc kubenswrapper[4890]: E1125 15:16:07.804704 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/watcher-operator-controller-manager-864885998-62ddz" podUID="a9a69828-3f0d-4d55-a840-10c2c791444c" Nov 25 15:16:07 crc kubenswrapper[4890]: I1125 15:16:07.902916 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-lfnwl" event={"ID":"076c2a0e-7fdb-42e1-ab73-e296bba2b7a9","Type":"ContainerStarted","Data":"d58069a8ab17e672838c27d0237d901cb2495c1991a4963e4459de73005fceca"} Nov 25 15:16:07 crc kubenswrapper[4890]: I1125 15:16:07.904260 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-wvjw7" event={"ID":"03353c8b-91f4-41ae-afc1-c522b1c73bb4","Type":"ContainerStarted","Data":"8076b10ee1192227bde62db4e788c9310da6f2f29feaac9b0a69aa1f058e810f"} Nov 25 15:16:07 crc kubenswrapper[4890]: E1125 15:16:07.906804 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:7b90521b9e9cb4eb43c2f1c3bf85dbd068d684315f4f705b07708dd078df9d04\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-wvjw7" podUID="03353c8b-91f4-41ae-afc1-c522b1c73bb4" Nov 25 15:16:07 crc kubenswrapper[4890]: I1125 15:16:07.907566 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-56bcr" event={"ID":"a9efeabb-bf1c-4de2-8322-961b284e95cc","Type":"ContainerStarted","Data":"09ed8bd82ba1777b270c9e9089ea2a64dfd007d6f14117b249ae4124b8324224"} Nov 25 15:16:07 crc kubenswrapper[4890]: E1125 15:16:07.910711 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:442c269d79163f8da75505019c02e9f0815837aaadcaddacb8e6c12df297ca13\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-56bcr" podUID="a9efeabb-bf1c-4de2-8322-961b284e95cc" Nov 25 15:16:07 crc kubenswrapper[4890]: I1125 15:16:07.911178 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-774b86978c-xvfgd" event={"ID":"72c26c5b-3370-4994-a5f0-82128c62c2df","Type":"ContainerStarted","Data":"fd8691107fe265de6077fa14335805f9b53adcc58aedeb31308ada4c81aac90f"} Nov 25 15:16:07 crc kubenswrapper[4890]: I1125 15:16:07.912094 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-rs6hx" event={"ID":"a5d8a8ba-50c0-4fb6-823f-f28cf0d0f125","Type":"ContainerStarted","Data":"41794c055d0453ed183c5e24ef237da16d18ca29c39aadd87f8c5a8052ecbdae"} Nov 25 15:16:07 crc kubenswrapper[4890]: I1125 15:16:07.913453 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-j6pqs" event={"ID":"c528348f-0907-423a-a0c2-f15cf18ff99a","Type":"ContainerStarted","Data":"8df4a270d35a14c962b6941c6976d497fa55fd2b0e48cd0a3018db9c52bfb746"} Nov 25 15:16:07 crc kubenswrapper[4890]: I1125 15:16:07.921005 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-pmmf2" event={"ID":"d175a35b-8b9b-49c0-8fc2-02ef62f131d9","Type":"ContainerStarted","Data":"26ddfc01803e9ddbb21fe25338a5599b1e1fb75c26d3f3d3330741c8196b250a"} Nov 25 15:16:07 crc kubenswrapper[4890]: I1125 15:16:07.931687 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-dmlx4" event={"ID":"e26594ec-4bc2-4051-8a49-fc490bd3b6c9","Type":"ContainerStarted","Data":"578dc0fe2ab733745c6a8f224624bc18045529707cc18e383b449c3dd14babc4"} Nov 25 15:16:07 crc kubenswrapper[4890]: E1125 15:16:07.933571 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-dmlx4" podUID="e26594ec-4bc2-4051-8a49-fc490bd3b6c9" Nov 25 15:16:07 crc kubenswrapper[4890]: I1125 15:16:07.941144 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6jq7q" event={"ID":"0769cde4-b49f-41c5-b720-e0dbc667eb4a","Type":"ContainerStarted","Data":"f0ffd78f393061bb26f75df64b55df422e06c632057fb08f001463c871fd1d51"} Nov 25 15:16:07 crc kubenswrapper[4890]: I1125 15:16:07.943247 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-829nk" event={"ID":"4364eec5-444a-41f6-af1c-fe16b24bd044","Type":"ContainerStarted","Data":"9a20bac8ac4f6e4dd5af7374d21f2407afa170dfdabe8fe9012b4cb0e4c17483"} Nov 25 15:16:07 crc kubenswrapper[4890]: E1125 15:16:07.943307 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6jq7q" podUID="0769cde4-b49f-41c5-b720-e0dbc667eb4a" Nov 25 15:16:07 crc kubenswrapper[4890]: I1125 15:16:07.944801 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-slpt4" event={"ID":"f272410f-230d-43f2-946e-e34e4d9c05ea","Type":"ContainerStarted","Data":"d9b2e1b96e68aa1adedb9c59544fedfd11a29653fdac17ed9b29669f3dcbc973"} Nov 25 15:16:07 crc kubenswrapper[4890]: I1125 15:16:07.945956 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-864885998-62ddz" event={"ID":"a9a69828-3f0d-4d55-a840-10c2c791444c","Type":"ContainerStarted","Data":"eb69fabbc8d6df7c8ce3c8802db231e31fd300c5dc20786c185d467985247195"} Nov 25 15:16:07 crc kubenswrapper[4890]: E1125 15:16:07.947661 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-864885998-62ddz" podUID="a9a69828-3f0d-4d55-a840-10c2c791444c" Nov 25 15:16:07 crc kubenswrapper[4890]: I1125 15:16:07.948211 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-stwjm" event={"ID":"77a5e8f9-92ab-48b5-be19-29a7a8e5df49","Type":"ContainerStarted","Data":"c670322c8b301270d81ade0bf0d59c2df0734a9f7c15b2c230edb6ed69c46593"} Nov 25 15:16:07 crc kubenswrapper[4890]: I1125 15:16:07.956570 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-6zr4x" event={"ID":"d50aec9d-1673-4b81-bb57-04cae418843e","Type":"ContainerStarted","Data":"2bb4d7b304c14137ae878e216c438f7d9aa9eae89adfe75c3a64cfd5ca611cc6"} Nov 25 15:16:07 crc kubenswrapper[4890]: I1125 15:16:07.962222 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-jcw7x" event={"ID":"0bdbada3-e08c-4f8a-bd55-bac955d4370d","Type":"ContainerStarted","Data":"7c54f8baa66686ab63ed1820820b472455892f5ce9f1a7ea7886d8350ff77523"} Nov 25 15:16:07 crc kubenswrapper[4890]: I1125 15:16:07.963525 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-nshw7" event={"ID":"9f0e3cf9-dcc0-405f-a2de-9148844ec3ae","Type":"ContainerStarted","Data":"0b10fa541f7f344027f68d6088079ce4cef8b68881cbf4e00aae7ba981eda3ce"} Nov 25 15:16:07 crc kubenswrapper[4890]: I1125 15:16:07.964423 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cb74df96-6mqt2" event={"ID":"9637fded-765d-4037-be8f-82e59ffdfc14","Type":"ContainerStarted","Data":"9adffdc7cc493f7e2c9057e90549de041fa22f51bf98de2d3c2e67e2d18afbb0"} Nov 25 15:16:07 crc kubenswrapper[4890]: E1125 15:16:07.967015 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5cb74df96-6mqt2" podUID="9637fded-765d-4037-be8f-82e59ffdfc14" Nov 25 15:16:07 crc kubenswrapper[4890]: I1125 15:16:07.967192 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-dzkgt" event={"ID":"861317e1-222d-4f14-b931-e9e7d100ebd6","Type":"ContainerStarted","Data":"da5811753bd8ae8bead057521cdbff92ac04632178536d0e42d3c8755068ba8e"} Nov 25 15:16:07 crc kubenswrapper[4890]: I1125 15:16:07.968719 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-d5cc86f4b-stflb"] Nov 25 15:16:07 crc kubenswrapper[4890]: I1125 15:16:07.969138 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-hxzf4" event={"ID":"679b3bf3-1ebc-4d19-9ad2-588bd41565dd","Type":"ContainerStarted","Data":"2f1c6eb5fa46b8036c0fb9ad4bb86ea98ec4e04723b249cab34c2a96041fcc67"} Nov 25 15:16:07 crc kubenswrapper[4890]: E1125 15:16:07.969290 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-dzkgt" podUID="861317e1-222d-4f14-b931-e9e7d100ebd6" Nov 25 15:16:07 crc kubenswrapper[4890]: I1125 15:16:07.970496 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-vx75t" event={"ID":"29f250db-2a96-41ad-bb8f-5b503a9288d8","Type":"ContainerStarted","Data":"c1e1bd0f8a7e6167aa2d0f66fa944267684c56e2f506a8d9e00094c50d1a4de3"} Nov 25 15:16:07 crc kubenswrapper[4890]: W1125 15:16:07.972042 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7e576602_19f5_4656_84dd_5ab80abca895.slice/crio-96715f0a454cbbf097bc680d3df451ded8bdb4b4805a06ec9d1d39c1989e4da4 WatchSource:0}: Error finding container 96715f0a454cbbf097bc680d3df451ded8bdb4b4805a06ec9d1d39c1989e4da4: Status 404 returned error can't find the container with id 96715f0a454cbbf097bc680d3df451ded8bdb4b4805a06ec9d1d39c1989e4da4 Nov 25 15:16:07 crc kubenswrapper[4890]: I1125 15:16:07.978891 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/10534635-efe1-4e20-8c9b-9cbea73fbe87-cert\") pod \"openstack-baremetal-operator-controller-manager-544b9bb9-5lr7r\" (UID: \"10534635-efe1-4e20-8c9b-9cbea73fbe87\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-5lr7r" Nov 25 15:16:07 crc kubenswrapper[4890]: I1125 15:16:07.990818 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/10534635-efe1-4e20-8c9b-9cbea73fbe87-cert\") pod \"openstack-baremetal-operator-controller-manager-544b9bb9-5lr7r\" (UID: \"10534635-efe1-4e20-8c9b-9cbea73fbe87\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-5lr7r" Nov 25 15:16:08 crc kubenswrapper[4890]: I1125 15:16:08.163252 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-5lr7r" Nov 25 15:16:08 crc kubenswrapper[4890]: I1125 15:16:08.385922 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3e496c3d-9485-44bc-b3b1-dc1682a88434-webhook-certs\") pod \"openstack-operator-controller-manager-68587559f4-9b4cf\" (UID: \"3e496c3d-9485-44bc-b3b1-dc1682a88434\") " pod="openstack-operators/openstack-operator-controller-manager-68587559f4-9b4cf" Nov 25 15:16:08 crc kubenswrapper[4890]: I1125 15:16:08.385995 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3e496c3d-9485-44bc-b3b1-dc1682a88434-metrics-certs\") pod \"openstack-operator-controller-manager-68587559f4-9b4cf\" (UID: \"3e496c3d-9485-44bc-b3b1-dc1682a88434\") " pod="openstack-operators/openstack-operator-controller-manager-68587559f4-9b4cf" Nov 25 15:16:08 crc kubenswrapper[4890]: E1125 15:16:08.386213 4890 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 25 15:16:08 crc kubenswrapper[4890]: E1125 15:16:08.386253 4890 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 25 15:16:08 crc kubenswrapper[4890]: E1125 15:16:08.386326 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3e496c3d-9485-44bc-b3b1-dc1682a88434-webhook-certs podName:3e496c3d-9485-44bc-b3b1-dc1682a88434 nodeName:}" failed. No retries permitted until 2025-11-25 15:16:10.386305947 +0000 UTC m=+828.828768557 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/3e496c3d-9485-44bc-b3b1-dc1682a88434-webhook-certs") pod "openstack-operator-controller-manager-68587559f4-9b4cf" (UID: "3e496c3d-9485-44bc-b3b1-dc1682a88434") : secret "webhook-server-cert" not found Nov 25 15:16:08 crc kubenswrapper[4890]: E1125 15:16:08.386345 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3e496c3d-9485-44bc-b3b1-dc1682a88434-metrics-certs podName:3e496c3d-9485-44bc-b3b1-dc1682a88434 nodeName:}" failed. No retries permitted until 2025-11-25 15:16:10.386339128 +0000 UTC m=+828.828801738 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3e496c3d-9485-44bc-b3b1-dc1682a88434-metrics-certs") pod "openstack-operator-controller-manager-68587559f4-9b4cf" (UID: "3e496c3d-9485-44bc-b3b1-dc1682a88434") : secret "metrics-server-cert" not found Nov 25 15:16:08 crc kubenswrapper[4890]: I1125 15:16:08.598006 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-5lr7r"] Nov 25 15:16:08 crc kubenswrapper[4890]: I1125 15:16:08.982456 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-stflb" event={"ID":"7e576602-19f5-4656-84dd-5ab80abca895","Type":"ContainerStarted","Data":"96715f0a454cbbf097bc680d3df451ded8bdb4b4805a06ec9d1d39c1989e4da4"} Nov 25 15:16:08 crc kubenswrapper[4890]: I1125 15:16:08.990503 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-5lr7r" event={"ID":"10534635-efe1-4e20-8c9b-9cbea73fbe87","Type":"ContainerStarted","Data":"eb1a677088d0f6096c12df96e8461b92f7ad747202cc1fc956d3eeb0850c0339"} Nov 25 15:16:08 crc kubenswrapper[4890]: E1125 15:16:08.998023 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:7b90521b9e9cb4eb43c2f1c3bf85dbd068d684315f4f705b07708dd078df9d04\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-wvjw7" podUID="03353c8b-91f4-41ae-afc1-c522b1c73bb4" Nov 25 15:16:08 crc kubenswrapper[4890]: E1125 15:16:08.998097 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-dmlx4" podUID="e26594ec-4bc2-4051-8a49-fc490bd3b6c9" Nov 25 15:16:08 crc kubenswrapper[4890]: E1125 15:16:08.998141 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5cb74df96-6mqt2" podUID="9637fded-765d-4037-be8f-82e59ffdfc14" Nov 25 15:16:08 crc kubenswrapper[4890]: E1125 15:16:08.998659 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6jq7q" podUID="0769cde4-b49f-41c5-b720-e0dbc667eb4a" Nov 25 15:16:08 crc kubenswrapper[4890]: E1125 15:16:08.998798 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-864885998-62ddz" podUID="a9a69828-3f0d-4d55-a840-10c2c791444c" Nov 25 15:16:08 crc kubenswrapper[4890]: E1125 15:16:08.998806 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:442c269d79163f8da75505019c02e9f0815837aaadcaddacb8e6c12df297ca13\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-56bcr" podUID="a9efeabb-bf1c-4de2-8322-961b284e95cc" Nov 25 15:16:09 crc kubenswrapper[4890]: E1125 15:16:09.000346 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-dzkgt" podUID="861317e1-222d-4f14-b931-e9e7d100ebd6" Nov 25 15:16:10 crc kubenswrapper[4890]: I1125 15:16:10.413991 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3e496c3d-9485-44bc-b3b1-dc1682a88434-webhook-certs\") pod \"openstack-operator-controller-manager-68587559f4-9b4cf\" (UID: \"3e496c3d-9485-44bc-b3b1-dc1682a88434\") " pod="openstack-operators/openstack-operator-controller-manager-68587559f4-9b4cf" Nov 25 15:16:10 crc kubenswrapper[4890]: I1125 15:16:10.414088 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3e496c3d-9485-44bc-b3b1-dc1682a88434-metrics-certs\") pod \"openstack-operator-controller-manager-68587559f4-9b4cf\" (UID: \"3e496c3d-9485-44bc-b3b1-dc1682a88434\") " pod="openstack-operators/openstack-operator-controller-manager-68587559f4-9b4cf" Nov 25 15:16:10 crc kubenswrapper[4890]: I1125 15:16:10.422131 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3e496c3d-9485-44bc-b3b1-dc1682a88434-metrics-certs\") pod \"openstack-operator-controller-manager-68587559f4-9b4cf\" (UID: \"3e496c3d-9485-44bc-b3b1-dc1682a88434\") " pod="openstack-operators/openstack-operator-controller-manager-68587559f4-9b4cf" Nov 25 15:16:10 crc kubenswrapper[4890]: I1125 15:16:10.426523 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3e496c3d-9485-44bc-b3b1-dc1682a88434-webhook-certs\") pod \"openstack-operator-controller-manager-68587559f4-9b4cf\" (UID: \"3e496c3d-9485-44bc-b3b1-dc1682a88434\") " pod="openstack-operators/openstack-operator-controller-manager-68587559f4-9b4cf" Nov 25 15:16:10 crc kubenswrapper[4890]: I1125 15:16:10.662984 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-68587559f4-9b4cf" Nov 25 15:16:16 crc kubenswrapper[4890]: I1125 15:16:16.936879 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-68587559f4-9b4cf"] Nov 25 15:16:16 crc kubenswrapper[4890]: W1125 15:16:16.963332 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3e496c3d_9485_44bc_b3b1_dc1682a88434.slice/crio-dd99d1437b7f4e57c070103d4e5153e39071455a28019683db86cb529fd6b9f1 WatchSource:0}: Error finding container dd99d1437b7f4e57c070103d4e5153e39071455a28019683db86cb529fd6b9f1: Status 404 returned error can't find the container with id dd99d1437b7f4e57c070103d4e5153e39071455a28019683db86cb529fd6b9f1 Nov 25 15:16:17 crc kubenswrapper[4890]: I1125 15:16:17.062113 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-lfnwl" event={"ID":"076c2a0e-7fdb-42e1-ab73-e296bba2b7a9","Type":"ContainerStarted","Data":"5d003110096be8e35d99eab0930c5aa6a3915752a25f20c2a6536a69af3d3afa"} Nov 25 15:16:17 crc kubenswrapper[4890]: I1125 15:16:17.063789 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-68587559f4-9b4cf" event={"ID":"3e496c3d-9485-44bc-b3b1-dc1682a88434","Type":"ContainerStarted","Data":"dd99d1437b7f4e57c070103d4e5153e39071455a28019683db86cb529fd6b9f1"} Nov 25 15:16:17 crc kubenswrapper[4890]: I1125 15:16:17.071082 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-jcw7x" event={"ID":"0bdbada3-e08c-4f8a-bd55-bac955d4370d","Type":"ContainerStarted","Data":"a3d62940c9f4d8d0e7c33babf91f1142db44aa39a25637bc44e7832b2e2c5c86"} Nov 25 15:16:17 crc kubenswrapper[4890]: I1125 15:16:17.074549 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-pmmf2" event={"ID":"d175a35b-8b9b-49c0-8fc2-02ef62f131d9","Type":"ContainerStarted","Data":"1eda77f255a83e5efb319f627a4f7cd146d6e94ac18d2411cf044b47387b7986"} Nov 25 15:16:17 crc kubenswrapper[4890]: I1125 15:16:17.075862 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-vx75t" event={"ID":"29f250db-2a96-41ad-bb8f-5b503a9288d8","Type":"ContainerStarted","Data":"75820db98fdfa144b9503a1d7e8750f8d454f7258242e932a8932d40b2582263"} Nov 25 15:16:17 crc kubenswrapper[4890]: I1125 15:16:17.077186 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-stwjm" event={"ID":"77a5e8f9-92ab-48b5-be19-29a7a8e5df49","Type":"ContainerStarted","Data":"d9880e46a52dd26c73643cdd5427eeee12138ceffa656e9d825b0e88a70d2335"} Nov 25 15:16:17 crc kubenswrapper[4890]: E1125 15:16:17.190971 4890 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7t5zd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-5bfcdc958c-hxzf4_openstack-operators(679b3bf3-1ebc-4d19-9ad2-588bd41565dd): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 15:16:17 crc kubenswrapper[4890]: E1125 15:16:17.195202 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-hxzf4" podUID="679b3bf3-1ebc-4d19-9ad2-588bd41565dd" Nov 25 15:16:17 crc kubenswrapper[4890]: E1125 15:16:17.212894 4890 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vpprx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-5db546f9d9-nshw7_openstack-operators(9f0e3cf9-dcc0-405f-a2de-9148844ec3ae): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 15:16:17 crc kubenswrapper[4890]: E1125 15:16:17.214337 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-nshw7" podUID="9f0e3cf9-dcc0-405f-a2de-9148844ec3ae" Nov 25 15:16:17 crc kubenswrapper[4890]: E1125 15:16:17.218612 4890 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9jntg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-544b9bb9-5lr7r_openstack-operators(10534635-efe1-4e20-8c9b-9cbea73fbe87): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 15:16:17 crc kubenswrapper[4890]: E1125 15:16:17.222382 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-5lr7r" podUID="10534635-efe1-4e20-8c9b-9cbea73fbe87" Nov 25 15:16:17 crc kubenswrapper[4890]: E1125 15:16:17.239767 4890 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lgbqr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-7c57c8bbc4-829nk_openstack-operators(4364eec5-444a-41f6-af1c-fe16b24bd044): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 15:16:17 crc kubenswrapper[4890]: E1125 15:16:17.243252 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-829nk" podUID="4364eec5-444a-41f6-af1c-fe16b24bd044" Nov 25 15:16:18 crc kubenswrapper[4890]: I1125 15:16:18.086411 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-774b86978c-xvfgd" event={"ID":"72c26c5b-3370-4994-a5f0-82128c62c2df","Type":"ContainerStarted","Data":"544c8c93f950a78e83efe0178f366fc587812386594cb297859d064206d045e1"} Nov 25 15:16:18 crc kubenswrapper[4890]: I1125 15:16:18.088191 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-slpt4" event={"ID":"f272410f-230d-43f2-946e-e34e4d9c05ea","Type":"ContainerStarted","Data":"8e7f0def045216a15a744d26de933e25b9a2e25ac620c03d95240aa41cc03038"} Nov 25 15:16:18 crc kubenswrapper[4890]: I1125 15:16:18.089783 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-6zr4x" event={"ID":"d50aec9d-1673-4b81-bb57-04cae418843e","Type":"ContainerStarted","Data":"b281ef61e39f8d5a43f4b56119cbfae16eb18c923bfe412c39c180018be2370e"} Nov 25 15:16:18 crc kubenswrapper[4890]: I1125 15:16:18.091689 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-5lr7r" event={"ID":"10534635-efe1-4e20-8c9b-9cbea73fbe87","Type":"ContainerStarted","Data":"eb7398300eba18c949ded5060fd4b31292e20d51a55d721c612ff62b864206f4"} Nov 25 15:16:18 crc kubenswrapper[4890]: I1125 15:16:18.091860 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-5lr7r" Nov 25 15:16:18 crc kubenswrapper[4890]: E1125 15:16:18.093234 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-5lr7r" podUID="10534635-efe1-4e20-8c9b-9cbea73fbe87" Nov 25 15:16:18 crc kubenswrapper[4890]: I1125 15:16:18.093785 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-nshw7" event={"ID":"9f0e3cf9-dcc0-405f-a2de-9148844ec3ae","Type":"ContainerStarted","Data":"cc0132b2f896b99524175db61d9c1524df2e54f587019f94d706da2cdea210cc"} Nov 25 15:16:18 crc kubenswrapper[4890]: I1125 15:16:18.093917 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-nshw7" Nov 25 15:16:18 crc kubenswrapper[4890]: E1125 15:16:18.106174 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-nshw7" podUID="9f0e3cf9-dcc0-405f-a2de-9148844ec3ae" Nov 25 15:16:18 crc kubenswrapper[4890]: I1125 15:16:18.107862 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-j6pqs" event={"ID":"c528348f-0907-423a-a0c2-f15cf18ff99a","Type":"ContainerStarted","Data":"dbe71c9a5936928af6baf15ad268bc19748f775a26475ca9db52499631dbeef9"} Nov 25 15:16:18 crc kubenswrapper[4890]: I1125 15:16:18.111152 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-68587559f4-9b4cf" event={"ID":"3e496c3d-9485-44bc-b3b1-dc1682a88434","Type":"ContainerStarted","Data":"25d97eb8b0662c62e96b5e278d3fba5c2fb55235f072e6116280714f791a4c15"} Nov 25 15:16:18 crc kubenswrapper[4890]: I1125 15:16:18.112994 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-68587559f4-9b4cf" Nov 25 15:16:18 crc kubenswrapper[4890]: I1125 15:16:18.129013 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-rs6hx" event={"ID":"a5d8a8ba-50c0-4fb6-823f-f28cf0d0f125","Type":"ContainerStarted","Data":"8756ce10612e1443f132d473e033dbcdfe66560ddde56b5f30ba360b8c363acb"} Nov 25 15:16:18 crc kubenswrapper[4890]: I1125 15:16:18.138585 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-829nk" event={"ID":"4364eec5-444a-41f6-af1c-fe16b24bd044","Type":"ContainerStarted","Data":"cc917ad88eaa4725e8cdac41f0da61db2690d8f49ca913c86a6c76f895cef181"} Nov 25 15:16:18 crc kubenswrapper[4890]: I1125 15:16:18.140248 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-829nk" Nov 25 15:16:18 crc kubenswrapper[4890]: E1125 15:16:18.145389 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-829nk" podUID="4364eec5-444a-41f6-af1c-fe16b24bd044" Nov 25 15:16:18 crc kubenswrapper[4890]: I1125 15:16:18.149901 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-hxzf4" event={"ID":"679b3bf3-1ebc-4d19-9ad2-588bd41565dd","Type":"ContainerStarted","Data":"9cb6b91fbb6bd830583da0085dd152e38f130ea75d9d3d4475674031434bf81c"} Nov 25 15:16:18 crc kubenswrapper[4890]: I1125 15:16:18.149976 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-hxzf4" Nov 25 15:16:18 crc kubenswrapper[4890]: E1125 15:16:18.151522 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-hxzf4" podUID="679b3bf3-1ebc-4d19-9ad2-588bd41565dd" Nov 25 15:16:18 crc kubenswrapper[4890]: I1125 15:16:18.151809 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-stflb" event={"ID":"7e576602-19f5-4656-84dd-5ab80abca895","Type":"ContainerStarted","Data":"aa72b2ab7e7b9beabe0d4dc521c692a9eb7688f8f0273d14f7a29938b97646c9"} Nov 25 15:16:18 crc kubenswrapper[4890]: I1125 15:16:18.171591 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-68587559f4-9b4cf" podStartSLOduration=12.17157025 podStartE2EDuration="12.17157025s" podCreationTimestamp="2025-11-25 15:16:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:16:18.168149705 +0000 UTC m=+836.610612305" watchObservedRunningTime="2025-11-25 15:16:18.17157025 +0000 UTC m=+836.614032860" Nov 25 15:16:19 crc kubenswrapper[4890]: E1125 15:16:19.160978 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-829nk" podUID="4364eec5-444a-41f6-af1c-fe16b24bd044" Nov 25 15:16:19 crc kubenswrapper[4890]: E1125 15:16:19.164048 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-5lr7r" podUID="10534635-efe1-4e20-8c9b-9cbea73fbe87" Nov 25 15:16:19 crc kubenswrapper[4890]: E1125 15:16:19.164090 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-nshw7" podUID="9f0e3cf9-dcc0-405f-a2de-9148844ec3ae" Nov 25 15:16:19 crc kubenswrapper[4890]: E1125 15:16:19.164287 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-hxzf4" podUID="679b3bf3-1ebc-4d19-9ad2-588bd41565dd" Nov 25 15:16:21 crc kubenswrapper[4890]: I1125 15:16:21.192411 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-lfnwl" event={"ID":"076c2a0e-7fdb-42e1-ab73-e296bba2b7a9","Type":"ContainerStarted","Data":"680d4899991f527843b87d49cf262e3535d1771c94fe9b96a804306e756b34a4"} Nov 25 15:16:21 crc kubenswrapper[4890]: I1125 15:16:21.193824 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-lfnwl" Nov 25 15:16:21 crc kubenswrapper[4890]: I1125 15:16:21.199713 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-lfnwl" Nov 25 15:16:21 crc kubenswrapper[4890]: I1125 15:16:21.202105 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-6zr4x" event={"ID":"d50aec9d-1673-4b81-bb57-04cae418843e","Type":"ContainerStarted","Data":"ad71ab6fa34e456775b3f6dbdc4c37250f5e23718b2ce23510092cafbe9303b3"} Nov 25 15:16:21 crc kubenswrapper[4890]: I1125 15:16:21.202257 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-6zr4x" Nov 25 15:16:21 crc kubenswrapper[4890]: I1125 15:16:21.205197 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-j6pqs" event={"ID":"c528348f-0907-423a-a0c2-f15cf18ff99a","Type":"ContainerStarted","Data":"e4e29e1d87b1d1d763aa5b6f21ae8e753eeb9079a8d3940c3b0d9ce9680c661f"} Nov 25 15:16:21 crc kubenswrapper[4890]: I1125 15:16:21.206259 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-j6pqs" Nov 25 15:16:21 crc kubenswrapper[4890]: I1125 15:16:21.217388 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-pmmf2" event={"ID":"d175a35b-8b9b-49c0-8fc2-02ef62f131d9","Type":"ContainerStarted","Data":"c8632fdcc2f09e024b9c4187699522fa290e53c8121563d32db8af414277a030"} Nov 25 15:16:21 crc kubenswrapper[4890]: I1125 15:16:21.218141 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-pmmf2" Nov 25 15:16:21 crc kubenswrapper[4890]: I1125 15:16:21.224110 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-vx75t" event={"ID":"29f250db-2a96-41ad-bb8f-5b503a9288d8","Type":"ContainerStarted","Data":"e30d7e6eefd2a297f4a88de212eebf843d555db098d142ed7d46a3ac149ae5b0"} Nov 25 15:16:21 crc kubenswrapper[4890]: I1125 15:16:21.224818 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-vx75t" Nov 25 15:16:21 crc kubenswrapper[4890]: I1125 15:16:21.226551 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-lfnwl" podStartSLOduration=2.648615485 podStartE2EDuration="16.226540498s" podCreationTimestamp="2025-11-25 15:16:05 +0000 UTC" firstStartedPulling="2025-11-25 15:16:07.303502043 +0000 UTC m=+825.745964653" lastFinishedPulling="2025-11-25 15:16:20.881427056 +0000 UTC m=+839.323889666" observedRunningTime="2025-11-25 15:16:21.212009545 +0000 UTC m=+839.654472165" watchObservedRunningTime="2025-11-25 15:16:21.226540498 +0000 UTC m=+839.669003108" Nov 25 15:16:21 crc kubenswrapper[4890]: I1125 15:16:21.231875 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-6zr4x" podStartSLOduration=2.178566416 podStartE2EDuration="15.231863771s" podCreationTimestamp="2025-11-25 15:16:06 +0000 UTC" firstStartedPulling="2025-11-25 15:16:07.594370351 +0000 UTC m=+826.036832961" lastFinishedPulling="2025-11-25 15:16:20.647667706 +0000 UTC m=+839.090130316" observedRunningTime="2025-11-25 15:16:21.228865166 +0000 UTC m=+839.671327776" watchObservedRunningTime="2025-11-25 15:16:21.231863771 +0000 UTC m=+839.674326381" Nov 25 15:16:21 crc kubenswrapper[4890]: I1125 15:16:21.237760 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-pmmf2" Nov 25 15:16:21 crc kubenswrapper[4890]: I1125 15:16:21.240302 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-774b86978c-xvfgd" event={"ID":"72c26c5b-3370-4994-a5f0-82128c62c2df","Type":"ContainerStarted","Data":"a483e6c27cf03325a911faf0ab774fd1743585c34aef5b74463cdc5883dfbebb"} Nov 25 15:16:21 crc kubenswrapper[4890]: I1125 15:16:21.240929 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-774b86978c-xvfgd" Nov 25 15:16:21 crc kubenswrapper[4890]: I1125 15:16:21.261901 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-stwjm" event={"ID":"77a5e8f9-92ab-48b5-be19-29a7a8e5df49","Type":"ContainerStarted","Data":"588ab6580e0ed0cfb3ab90e3e11fcffaad45240ae27ee34957feaee107de4a9c"} Nov 25 15:16:21 crc kubenswrapper[4890]: I1125 15:16:21.262520 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-stwjm" Nov 25 15:16:21 crc kubenswrapper[4890]: I1125 15:16:21.278386 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-slpt4" event={"ID":"f272410f-230d-43f2-946e-e34e4d9c05ea","Type":"ContainerStarted","Data":"03d0dafe03c3316ba502d6958eee7b18626df85587b1d29602175302378a32d3"} Nov 25 15:16:21 crc kubenswrapper[4890]: I1125 15:16:21.279060 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-slpt4" Nov 25 15:16:21 crc kubenswrapper[4890]: I1125 15:16:21.304225 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-j6pqs" podStartSLOduration=3.125578303 podStartE2EDuration="16.304201999s" podCreationTimestamp="2025-11-25 15:16:05 +0000 UTC" firstStartedPulling="2025-11-25 15:16:07.592676348 +0000 UTC m=+826.035138958" lastFinishedPulling="2025-11-25 15:16:20.771300044 +0000 UTC m=+839.213762654" observedRunningTime="2025-11-25 15:16:21.279384399 +0000 UTC m=+839.721847009" watchObservedRunningTime="2025-11-25 15:16:21.304201999 +0000 UTC m=+839.746664599" Nov 25 15:16:21 crc kubenswrapper[4890]: I1125 15:16:21.347762 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-vx75t" podStartSLOduration=2.512812768 podStartE2EDuration="15.347738807s" podCreationTimestamp="2025-11-25 15:16:06 +0000 UTC" firstStartedPulling="2025-11-25 15:16:07.731416695 +0000 UTC m=+826.173879305" lastFinishedPulling="2025-11-25 15:16:20.566342734 +0000 UTC m=+839.008805344" observedRunningTime="2025-11-25 15:16:21.315282116 +0000 UTC m=+839.757744726" watchObservedRunningTime="2025-11-25 15:16:21.347738807 +0000 UTC m=+839.790201417" Nov 25 15:16:21 crc kubenswrapper[4890]: I1125 15:16:21.349152 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-pmmf2" podStartSLOduration=2.849348501 podStartE2EDuration="16.349142092s" podCreationTimestamp="2025-11-25 15:16:05 +0000 UTC" firstStartedPulling="2025-11-25 15:16:07.11846812 +0000 UTC m=+825.560930730" lastFinishedPulling="2025-11-25 15:16:20.618261711 +0000 UTC m=+839.060724321" observedRunningTime="2025-11-25 15:16:21.337529341 +0000 UTC m=+839.779991971" watchObservedRunningTime="2025-11-25 15:16:21.349142092 +0000 UTC m=+839.791604702" Nov 25 15:16:21 crc kubenswrapper[4890]: I1125 15:16:21.370844 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-slpt4" podStartSLOduration=3.257069138 podStartE2EDuration="16.370827733s" podCreationTimestamp="2025-11-25 15:16:05 +0000 UTC" firstStartedPulling="2025-11-25 15:16:07.597580551 +0000 UTC m=+826.040043151" lastFinishedPulling="2025-11-25 15:16:20.711339126 +0000 UTC m=+839.153801746" observedRunningTime="2025-11-25 15:16:21.357425809 +0000 UTC m=+839.799888419" watchObservedRunningTime="2025-11-25 15:16:21.370827733 +0000 UTC m=+839.813290343" Nov 25 15:16:21 crc kubenswrapper[4890]: I1125 15:16:21.397440 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-774b86978c-xvfgd" podStartSLOduration=2.90171887 podStartE2EDuration="16.397415458s" podCreationTimestamp="2025-11-25 15:16:05 +0000 UTC" firstStartedPulling="2025-11-25 15:16:07.300949459 +0000 UTC m=+825.743412069" lastFinishedPulling="2025-11-25 15:16:20.796646047 +0000 UTC m=+839.239108657" observedRunningTime="2025-11-25 15:16:21.392953426 +0000 UTC m=+839.835416036" watchObservedRunningTime="2025-11-25 15:16:21.397415458 +0000 UTC m=+839.839878068" Nov 25 15:16:21 crc kubenswrapper[4890]: I1125 15:16:21.413832 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-stwjm" podStartSLOduration=2.554619637 podStartE2EDuration="16.413808077s" podCreationTimestamp="2025-11-25 15:16:05 +0000 UTC" firstStartedPulling="2025-11-25 15:16:07.133989948 +0000 UTC m=+825.576452558" lastFinishedPulling="2025-11-25 15:16:20.993178388 +0000 UTC m=+839.435640998" observedRunningTime="2025-11-25 15:16:21.409665734 +0000 UTC m=+839.852128344" watchObservedRunningTime="2025-11-25 15:16:21.413808077 +0000 UTC m=+839.856270687" Nov 25 15:16:22 crc kubenswrapper[4890]: I1125 15:16:22.294584 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-jcw7x" event={"ID":"0bdbada3-e08c-4f8a-bd55-bac955d4370d","Type":"ContainerStarted","Data":"925076a8867c27e12ac378d166b3d872f0fa421c1e7ff934b1add83190541eee"} Nov 25 15:16:22 crc kubenswrapper[4890]: I1125 15:16:22.295573 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-jcw7x" Nov 25 15:16:22 crc kubenswrapper[4890]: I1125 15:16:22.297193 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-jcw7x" Nov 25 15:16:22 crc kubenswrapper[4890]: I1125 15:16:22.300185 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-rs6hx" event={"ID":"a5d8a8ba-50c0-4fb6-823f-f28cf0d0f125","Type":"ContainerStarted","Data":"84624c8ad94f89a9f962b50febe1a7a7f52edfe5fddfe9cdee4109d7b890a13b"} Nov 25 15:16:22 crc kubenswrapper[4890]: I1125 15:16:22.300608 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-rs6hx" Nov 25 15:16:22 crc kubenswrapper[4890]: I1125 15:16:22.303864 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-rs6hx" Nov 25 15:16:22 crc kubenswrapper[4890]: I1125 15:16:22.305447 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-stflb" event={"ID":"7e576602-19f5-4656-84dd-5ab80abca895","Type":"ContainerStarted","Data":"f63bad099eb225b83bbb2fec0ca67b0f889323f1ea9ee86ec5979cf8c1793e0e"} Nov 25 15:16:22 crc kubenswrapper[4890]: I1125 15:16:22.308753 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-vx75t" Nov 25 15:16:22 crc kubenswrapper[4890]: I1125 15:16:22.308814 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-774b86978c-xvfgd" Nov 25 15:16:22 crc kubenswrapper[4890]: I1125 15:16:22.309805 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-slpt4" Nov 25 15:16:22 crc kubenswrapper[4890]: I1125 15:16:22.309892 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-6zr4x" Nov 25 15:16:22 crc kubenswrapper[4890]: I1125 15:16:22.310322 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-j6pqs" Nov 25 15:16:22 crc kubenswrapper[4890]: I1125 15:16:22.310693 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-stwjm" Nov 25 15:16:22 crc kubenswrapper[4890]: I1125 15:16:22.322661 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-jcw7x" podStartSLOduration=3.37321173 podStartE2EDuration="17.322641125s" podCreationTimestamp="2025-11-25 15:16:05 +0000 UTC" firstStartedPulling="2025-11-25 15:16:07.141487125 +0000 UTC m=+825.583949735" lastFinishedPulling="2025-11-25 15:16:21.09091652 +0000 UTC m=+839.533379130" observedRunningTime="2025-11-25 15:16:22.316410559 +0000 UTC m=+840.758873189" watchObservedRunningTime="2025-11-25 15:16:22.322641125 +0000 UTC m=+840.765103735" Nov 25 15:16:22 crc kubenswrapper[4890]: I1125 15:16:22.339704 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-rs6hx" podStartSLOduration=3.505274349 podStartE2EDuration="17.33968251s" podCreationTimestamp="2025-11-25 15:16:05 +0000 UTC" firstStartedPulling="2025-11-25 15:16:07.56833329 +0000 UTC m=+826.010795900" lastFinishedPulling="2025-11-25 15:16:21.402741451 +0000 UTC m=+839.845204061" observedRunningTime="2025-11-25 15:16:22.332933212 +0000 UTC m=+840.775395852" watchObservedRunningTime="2025-11-25 15:16:22.33968251 +0000 UTC m=+840.782145120" Nov 25 15:16:22 crc kubenswrapper[4890]: I1125 15:16:22.355805 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-stflb" podStartSLOduration=4.162714716 podStartE2EDuration="17.355789563s" podCreationTimestamp="2025-11-25 15:16:05 +0000 UTC" firstStartedPulling="2025-11-25 15:16:07.973850492 +0000 UTC m=+826.416313102" lastFinishedPulling="2025-11-25 15:16:21.166925339 +0000 UTC m=+839.609387949" observedRunningTime="2025-11-25 15:16:22.354633554 +0000 UTC m=+840.797096184" watchObservedRunningTime="2025-11-25 15:16:22.355789563 +0000 UTC m=+840.798252183" Nov 25 15:16:23 crc kubenswrapper[4890]: I1125 15:16:23.315239 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cb74df96-6mqt2" event={"ID":"9637fded-765d-4037-be8f-82e59ffdfc14","Type":"ContainerStarted","Data":"b67c37eff45a02fbf56c92843ff4f46029b678915f3b9724d7fe7e4e4ea50ef1"} Nov 25 15:16:23 crc kubenswrapper[4890]: I1125 15:16:23.315903 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cb74df96-6mqt2" event={"ID":"9637fded-765d-4037-be8f-82e59ffdfc14","Type":"ContainerStarted","Data":"8fb4dbced7063fa7a276b9f33f0f1924462d950c851895f4c391b9a313b109b4"} Nov 25 15:16:23 crc kubenswrapper[4890]: I1125 15:16:23.323420 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-stflb" Nov 25 15:16:23 crc kubenswrapper[4890]: I1125 15:16:23.330397 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-stflb" Nov 25 15:16:23 crc kubenswrapper[4890]: I1125 15:16:23.341645 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5cb74df96-6mqt2" podStartSLOduration=2.186182977 podStartE2EDuration="17.341622344s" podCreationTimestamp="2025-11-25 15:16:06 +0000 UTC" firstStartedPulling="2025-11-25 15:16:07.791572148 +0000 UTC m=+826.234034758" lastFinishedPulling="2025-11-25 15:16:22.947011515 +0000 UTC m=+841.389474125" observedRunningTime="2025-11-25 15:16:23.337434429 +0000 UTC m=+841.779897039" watchObservedRunningTime="2025-11-25 15:16:23.341622344 +0000 UTC m=+841.784084954" Nov 25 15:16:26 crc kubenswrapper[4890]: I1125 15:16:26.304755 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-hxzf4" Nov 25 15:16:26 crc kubenswrapper[4890]: I1125 15:16:26.759736 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-nshw7" Nov 25 15:16:26 crc kubenswrapper[4890]: I1125 15:16:26.793478 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-829nk" Nov 25 15:16:26 crc kubenswrapper[4890]: I1125 15:16:26.950973 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5cb74df96-6mqt2" Nov 25 15:16:27 crc kubenswrapper[4890]: I1125 15:16:27.357341 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-864885998-62ddz" event={"ID":"a9a69828-3f0d-4d55-a840-10c2c791444c","Type":"ContainerStarted","Data":"13f22d92a97bcc7278184b44c61653f64e5812c74ac3e58470ed92e7af1cc92b"} Nov 25 15:16:27 crc kubenswrapper[4890]: I1125 15:16:27.357394 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-864885998-62ddz" event={"ID":"a9a69828-3f0d-4d55-a840-10c2c791444c","Type":"ContainerStarted","Data":"76d8244364a59157c2b9cbffef85f40bfbb1b0424e7c06c66a9fdf3f856a8843"} Nov 25 15:16:27 crc kubenswrapper[4890]: I1125 15:16:27.359421 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-864885998-62ddz" Nov 25 15:16:27 crc kubenswrapper[4890]: I1125 15:16:27.367076 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-hxzf4" event={"ID":"679b3bf3-1ebc-4d19-9ad2-588bd41565dd","Type":"ContainerStarted","Data":"4ee0c49fb540c7d684da8abf3885a0d9544d9019f99f70662881aa410a480980"} Nov 25 15:16:27 crc kubenswrapper[4890]: I1125 15:16:27.377981 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-864885998-62ddz" podStartSLOduration=2.956285408 podStartE2EDuration="21.377964211s" podCreationTimestamp="2025-11-25 15:16:06 +0000 UTC" firstStartedPulling="2025-11-25 15:16:07.800685885 +0000 UTC m=+826.243148505" lastFinishedPulling="2025-11-25 15:16:26.222364708 +0000 UTC m=+844.664827308" observedRunningTime="2025-11-25 15:16:27.374671839 +0000 UTC m=+845.817134459" watchObservedRunningTime="2025-11-25 15:16:27.377964211 +0000 UTC m=+845.820426821" Nov 25 15:16:27 crc kubenswrapper[4890]: I1125 15:16:27.400206 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-hxzf4" podStartSLOduration=13.34208183 podStartE2EDuration="22.400181106s" podCreationTimestamp="2025-11-25 15:16:05 +0000 UTC" firstStartedPulling="2025-11-25 15:16:07.487628084 +0000 UTC m=+825.930090694" lastFinishedPulling="2025-11-25 15:16:16.54572734 +0000 UTC m=+834.988189970" observedRunningTime="2025-11-25 15:16:27.395989332 +0000 UTC m=+845.838451942" watchObservedRunningTime="2025-11-25 15:16:27.400181106 +0000 UTC m=+845.842643716" Nov 25 15:16:28 crc kubenswrapper[4890]: I1125 15:16:28.171846 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-5lr7r" Nov 25 15:16:30 crc kubenswrapper[4890]: I1125 15:16:30.671868 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-68587559f4-9b4cf" Nov 25 15:16:36 crc kubenswrapper[4890]: I1125 15:16:36.954229 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5cb74df96-6mqt2" Nov 25 15:16:36 crc kubenswrapper[4890]: I1125 15:16:36.985649 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-864885998-62ddz" Nov 25 15:16:42 crc kubenswrapper[4890]: E1125 15:16:42.246730 4890 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7" Nov 25 15:16:42 crc kubenswrapper[4890]: E1125 15:16:42.247320 4890 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nwkd7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-79556f57fc-6jq7q_openstack-operators(0769cde4-b49f-41c5-b720-e0dbc667eb4a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 15:16:43 crc kubenswrapper[4890]: E1125 15:16:43.627358 4890 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:7b90521b9e9cb4eb43c2f1c3bf85dbd068d684315f4f705b07708dd078df9d04" Nov 25 15:16:43 crc kubenswrapper[4890]: E1125 15:16:43.627764 4890 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:7b90521b9e9cb4eb43c2f1c3bf85dbd068d684315f4f705b07708dd078df9d04,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-n9snz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-cb6c4fdb7-wvjw7_openstack-operators(03353c8b-91f4-41ae-afc1-c522b1c73bb4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 15:16:44 crc kubenswrapper[4890]: E1125 15:16:44.081538 4890 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:442c269d79163f8da75505019c02e9f0815837aaadcaddacb8e6c12df297ca13" Nov 25 15:16:44 crc kubenswrapper[4890]: E1125 15:16:44.081734 4890 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:442c269d79163f8da75505019c02e9f0815837aaadcaddacb8e6c12df297ca13,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-n88zf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-fd75fd47d-56bcr_openstack-operators(a9efeabb-bf1c-4de2-8322-961b284e95cc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 15:16:44 crc kubenswrapper[4890]: E1125 15:16:44.567456 4890 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0" Nov 25 15:16:44 crc kubenswrapper[4890]: E1125 15:16:44.567634 4890 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-w29mm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-6fdc4fcf86-dzkgt_openstack-operators(861317e1-222d-4f14-b931-e9e7d100ebd6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 15:16:46 crc kubenswrapper[4890]: E1125 15:16:46.294641 4890 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Nov 25 15:16:46 crc kubenswrapper[4890]: E1125 15:16:46.295082 4890 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-swsbh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-dmlx4_openstack-operators(e26594ec-4bc2-4051-8a49-fc490bd3b6c9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 15:16:46 crc kubenswrapper[4890]: E1125 15:16:46.296226 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-dmlx4" podUID="e26594ec-4bc2-4051-8a49-fc490bd3b6c9" Nov 25 15:16:46 crc kubenswrapper[4890]: I1125 15:16:46.504711 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6jq7q" event={"ID":"0769cde4-b49f-41c5-b720-e0dbc667eb4a","Type":"ContainerStarted","Data":"4b2489af249fe5d7ba1a3ae5bf149c84e5432029f80ac3f5bec05f8f16b06870"} Nov 25 15:16:46 crc kubenswrapper[4890]: E1125 15:16:46.507338 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6jq7q" podUID="0769cde4-b49f-41c5-b720-e0dbc667eb4a" Nov 25 15:16:46 crc kubenswrapper[4890]: E1125 15:16:46.565831 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-56bcr" podUID="a9efeabb-bf1c-4de2-8322-961b284e95cc" Nov 25 15:16:46 crc kubenswrapper[4890]: E1125 15:16:46.566362 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-dzkgt" podUID="861317e1-222d-4f14-b931-e9e7d100ebd6" Nov 25 15:16:46 crc kubenswrapper[4890]: E1125 15:16:46.566537 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-wvjw7" podUID="03353c8b-91f4-41ae-afc1-c522b1c73bb4" Nov 25 15:16:47 crc kubenswrapper[4890]: I1125 15:16:47.511129 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-nshw7" event={"ID":"9f0e3cf9-dcc0-405f-a2de-9148844ec3ae","Type":"ContainerStarted","Data":"90c7f70116f68db4c061cb31dd394d28122c16e72ac1938367a1fd14805651fa"} Nov 25 15:16:47 crc kubenswrapper[4890]: I1125 15:16:47.513044 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-829nk" event={"ID":"4364eec5-444a-41f6-af1c-fe16b24bd044","Type":"ContainerStarted","Data":"e0d452c4e793c0c083c0922a99b52527dd8e04a55df59233623a49e76a84f41f"} Nov 25 15:16:47 crc kubenswrapper[4890]: I1125 15:16:47.515480 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-wvjw7" event={"ID":"03353c8b-91f4-41ae-afc1-c522b1c73bb4","Type":"ContainerStarted","Data":"ee277e95f7d5f457f32aeb9968baa9c29929cabb07d949add1fc946b4081aba4"} Nov 25 15:16:47 crc kubenswrapper[4890]: E1125 15:16:47.516896 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:7b90521b9e9cb4eb43c2f1c3bf85dbd068d684315f4f705b07708dd078df9d04\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-wvjw7" podUID="03353c8b-91f4-41ae-afc1-c522b1c73bb4" Nov 25 15:16:47 crc kubenswrapper[4890]: I1125 15:16:47.517848 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-dzkgt" event={"ID":"861317e1-222d-4f14-b931-e9e7d100ebd6","Type":"ContainerStarted","Data":"1b993b4e7949869bdad9d51d3f411c5fcce2e92dcda369aa6970e40d8690ed6e"} Nov 25 15:16:47 crc kubenswrapper[4890]: E1125 15:16:47.518977 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0\\\"\"" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-dzkgt" podUID="861317e1-222d-4f14-b931-e9e7d100ebd6" Nov 25 15:16:47 crc kubenswrapper[4890]: I1125 15:16:47.519533 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-56bcr" event={"ID":"a9efeabb-bf1c-4de2-8322-961b284e95cc","Type":"ContainerStarted","Data":"ca9596222c1ac988a1b8a8d5acb5ef66df5c1f0a25e1a5474633495dc53377d9"} Nov 25 15:16:47 crc kubenswrapper[4890]: E1125 15:16:47.520986 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:442c269d79163f8da75505019c02e9f0815837aaadcaddacb8e6c12df297ca13\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-56bcr" podUID="a9efeabb-bf1c-4de2-8322-961b284e95cc" Nov 25 15:16:47 crc kubenswrapper[4890]: I1125 15:16:47.521685 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-5lr7r" event={"ID":"10534635-efe1-4e20-8c9b-9cbea73fbe87","Type":"ContainerStarted","Data":"63770d9caa8d70f71ac7a3c576959bba332c39c13bdcde72f2bea54e4199e585"} Nov 25 15:16:47 crc kubenswrapper[4890]: E1125 15:16:47.523671 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7\\\"\"" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6jq7q" podUID="0769cde4-b49f-41c5-b720-e0dbc667eb4a" Nov 25 15:16:47 crc kubenswrapper[4890]: I1125 15:16:47.536951 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-nshw7" podStartSLOduration=32.533638819 podStartE2EDuration="41.536936796s" podCreationTimestamp="2025-11-25 15:16:06 +0000 UTC" firstStartedPulling="2025-11-25 15:16:07.602949995 +0000 UTC m=+826.045412605" lastFinishedPulling="2025-11-25 15:16:16.606247972 +0000 UTC m=+835.048710582" observedRunningTime="2025-11-25 15:16:47.531260954 +0000 UTC m=+865.973723564" watchObservedRunningTime="2025-11-25 15:16:47.536936796 +0000 UTC m=+865.979399406" Nov 25 15:16:47 crc kubenswrapper[4890]: I1125 15:16:47.613867 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-829nk" podStartSLOduration=33.662116209 podStartE2EDuration="42.613845817s" podCreationTimestamp="2025-11-25 15:16:05 +0000 UTC" firstStartedPulling="2025-11-25 15:16:07.602571896 +0000 UTC m=+826.045034506" lastFinishedPulling="2025-11-25 15:16:16.554301504 +0000 UTC m=+834.996764114" observedRunningTime="2025-11-25 15:16:47.610851063 +0000 UTC m=+866.053313673" watchObservedRunningTime="2025-11-25 15:16:47.613845817 +0000 UTC m=+866.056308427" Nov 25 15:16:47 crc kubenswrapper[4890]: I1125 15:16:47.689275 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-5lr7r" podStartSLOduration=33.755936718 podStartE2EDuration="41.689252841s" podCreationTimestamp="2025-11-25 15:16:06 +0000 UTC" firstStartedPulling="2025-11-25 15:16:08.617219777 +0000 UTC m=+827.059682387" lastFinishedPulling="2025-11-25 15:16:16.5505359 +0000 UTC m=+834.992998510" observedRunningTime="2025-11-25 15:16:47.684151174 +0000 UTC m=+866.126613794" watchObservedRunningTime="2025-11-25 15:16:47.689252841 +0000 UTC m=+866.131715451" Nov 25 15:16:56 crc kubenswrapper[4890]: I1125 15:16:56.447136 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 15:16:56 crc kubenswrapper[4890]: I1125 15:16:56.448434 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 15:16:59 crc kubenswrapper[4890]: E1125 15:16:59.175845 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-dmlx4" podUID="e26594ec-4bc2-4051-8a49-fc490bd3b6c9" Nov 25 15:16:59 crc kubenswrapper[4890]: E1125 15:16:59.176675 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7\\\"\"" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6jq7q" podUID="0769cde4-b49f-41c5-b720-e0dbc667eb4a" Nov 25 15:17:00 crc kubenswrapper[4890]: E1125 15:17:00.175034 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:442c269d79163f8da75505019c02e9f0815837aaadcaddacb8e6c12df297ca13\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-56bcr" podUID="a9efeabb-bf1c-4de2-8322-961b284e95cc" Nov 25 15:17:02 crc kubenswrapper[4890]: E1125 15:17:02.180093 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:7b90521b9e9cb4eb43c2f1c3bf85dbd068d684315f4f705b07708dd078df9d04\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-wvjw7" podUID="03353c8b-91f4-41ae-afc1-c522b1c73bb4" Nov 25 15:17:02 crc kubenswrapper[4890]: E1125 15:17:02.181007 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0\\\"\"" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-dzkgt" podUID="861317e1-222d-4f14-b931-e9e7d100ebd6" Nov 25 15:17:13 crc kubenswrapper[4890]: I1125 15:17:13.175531 4890 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 15:17:14 crc kubenswrapper[4890]: I1125 15:17:14.713791 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-56bcr" event={"ID":"a9efeabb-bf1c-4de2-8322-961b284e95cc","Type":"ContainerStarted","Data":"360f2cd67207fde2598cf2336d7ee654c1e558557dae8116644fbbad97f92bbe"} Nov 25 15:17:14 crc kubenswrapper[4890]: I1125 15:17:14.715398 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-dmlx4" event={"ID":"e26594ec-4bc2-4051-8a49-fc490bd3b6c9","Type":"ContainerStarted","Data":"61b9be5d90ce8a666acdd758f569aaa247bc6bf33719297ce9c65e886c773f90"} Nov 25 15:17:14 crc kubenswrapper[4890]: I1125 15:17:14.715464 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-56bcr" Nov 25 15:17:14 crc kubenswrapper[4890]: I1125 15:17:14.717320 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6jq7q" event={"ID":"0769cde4-b49f-41c5-b720-e0dbc667eb4a","Type":"ContainerStarted","Data":"97e02800cd95b6b08806006fece844f452d79638de429148891a1ede18c31c8f"} Nov 25 15:17:14 crc kubenswrapper[4890]: I1125 15:17:14.717526 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6jq7q" Nov 25 15:17:14 crc kubenswrapper[4890]: I1125 15:17:14.739959 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-56bcr" podStartSLOduration=2.755447278 podStartE2EDuration="1m8.739924879s" podCreationTimestamp="2025-11-25 15:16:06 +0000 UTC" firstStartedPulling="2025-11-25 15:16:07.794545872 +0000 UTC m=+826.237008502" lastFinishedPulling="2025-11-25 15:17:13.779023493 +0000 UTC m=+892.221486103" observedRunningTime="2025-11-25 15:17:14.732711339 +0000 UTC m=+893.175173969" watchObservedRunningTime="2025-11-25 15:17:14.739924879 +0000 UTC m=+893.182387489" Nov 25 15:17:14 crc kubenswrapper[4890]: I1125 15:17:14.752233 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6jq7q" podStartSLOduration=3.858423492 podStartE2EDuration="1m9.752212167s" podCreationTimestamp="2025-11-25 15:16:05 +0000 UTC" firstStartedPulling="2025-11-25 15:16:07.79085573 +0000 UTC m=+826.233318340" lastFinishedPulling="2025-11-25 15:17:13.684644405 +0000 UTC m=+892.127107015" observedRunningTime="2025-11-25 15:17:14.751619402 +0000 UTC m=+893.194082012" watchObservedRunningTime="2025-11-25 15:17:14.752212167 +0000 UTC m=+893.194674777" Nov 25 15:17:14 crc kubenswrapper[4890]: I1125 15:17:14.770119 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-dmlx4" podStartSLOduration=2.940733398 podStartE2EDuration="1m8.770096163s" podCreationTimestamp="2025-11-25 15:16:06 +0000 UTC" firstStartedPulling="2025-11-25 15:16:07.777726212 +0000 UTC m=+826.220188822" lastFinishedPulling="2025-11-25 15:17:13.607088967 +0000 UTC m=+892.049551587" observedRunningTime="2025-11-25 15:17:14.766036222 +0000 UTC m=+893.208498832" watchObservedRunningTime="2025-11-25 15:17:14.770096163 +0000 UTC m=+893.212558773" Nov 25 15:17:15 crc kubenswrapper[4890]: I1125 15:17:15.727474 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-wvjw7" event={"ID":"03353c8b-91f4-41ae-afc1-c522b1c73bb4","Type":"ContainerStarted","Data":"820303aa9ed50019579596b6edb70067e3cf2b1af539efec84d656ec8ffdf5f5"} Nov 25 15:17:15 crc kubenswrapper[4890]: I1125 15:17:15.727776 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-wvjw7" Nov 25 15:17:15 crc kubenswrapper[4890]: I1125 15:17:15.745569 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-wvjw7" podStartSLOduration=3.444605144 podStartE2EDuration="1m10.745549345s" podCreationTimestamp="2025-11-25 15:16:05 +0000 UTC" firstStartedPulling="2025-11-25 15:16:07.605626602 +0000 UTC m=+826.048089212" lastFinishedPulling="2025-11-25 15:17:14.906570803 +0000 UTC m=+893.349033413" observedRunningTime="2025-11-25 15:17:15.743835652 +0000 UTC m=+894.186298282" watchObservedRunningTime="2025-11-25 15:17:15.745549345 +0000 UTC m=+894.188011965" Nov 25 15:17:18 crc kubenswrapper[4890]: I1125 15:17:18.747513 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-dzkgt" event={"ID":"861317e1-222d-4f14-b931-e9e7d100ebd6","Type":"ContainerStarted","Data":"b8f781558438b476e77ca8c2495fc65050de7477ff54cb3185dc7dd498e07866"} Nov 25 15:17:18 crc kubenswrapper[4890]: I1125 15:17:18.748360 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-dzkgt" Nov 25 15:17:18 crc kubenswrapper[4890]: I1125 15:17:18.766722 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-dzkgt" podStartSLOduration=2.3470982879999998 podStartE2EDuration="1m12.766702299s" podCreationTimestamp="2025-11-25 15:16:06 +0000 UTC" firstStartedPulling="2025-11-25 15:16:07.605404196 +0000 UTC m=+826.047866806" lastFinishedPulling="2025-11-25 15:17:18.025008197 +0000 UTC m=+896.467470817" observedRunningTime="2025-11-25 15:17:18.763740915 +0000 UTC m=+897.206203525" watchObservedRunningTime="2025-11-25 15:17:18.766702299 +0000 UTC m=+897.209164909" Nov 25 15:17:26 crc kubenswrapper[4890]: I1125 15:17:26.439632 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-wvjw7" Nov 25 15:17:26 crc kubenswrapper[4890]: I1125 15:17:26.447446 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 15:17:26 crc kubenswrapper[4890]: I1125 15:17:26.447498 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 15:17:26 crc kubenswrapper[4890]: I1125 15:17:26.625728 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6jq7q" Nov 25 15:17:26 crc kubenswrapper[4890]: I1125 15:17:26.650880 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-56bcr" Nov 25 15:17:26 crc kubenswrapper[4890]: I1125 15:17:26.777999 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-dzkgt" Nov 25 15:17:40 crc kubenswrapper[4890]: I1125 15:17:40.725071 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-7pn5f"] Nov 25 15:17:40 crc kubenswrapper[4890]: I1125 15:17:40.727065 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-7pn5f" Nov 25 15:17:40 crc kubenswrapper[4890]: I1125 15:17:40.729216 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-wtxp4" Nov 25 15:17:40 crc kubenswrapper[4890]: I1125 15:17:40.730722 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Nov 25 15:17:40 crc kubenswrapper[4890]: I1125 15:17:40.731223 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Nov 25 15:17:40 crc kubenswrapper[4890]: I1125 15:17:40.731644 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Nov 25 15:17:40 crc kubenswrapper[4890]: I1125 15:17:40.743728 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-7pn5f"] Nov 25 15:17:40 crc kubenswrapper[4890]: I1125 15:17:40.818705 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-6fkdh"] Nov 25 15:17:40 crc kubenswrapper[4890]: I1125 15:17:40.823535 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-6fkdh" Nov 25 15:17:40 crc kubenswrapper[4890]: I1125 15:17:40.827302 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Nov 25 15:17:40 crc kubenswrapper[4890]: I1125 15:17:40.830438 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-6fkdh"] Nov 25 15:17:40 crc kubenswrapper[4890]: I1125 15:17:40.885413 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtbps\" (UniqueName: \"kubernetes.io/projected/a6c5bcbf-f90f-4d40-819d-c0bc8a5a47fa-kube-api-access-jtbps\") pod \"dnsmasq-dns-675f4bcbfc-7pn5f\" (UID: \"a6c5bcbf-f90f-4d40-819d-c0bc8a5a47fa\") " pod="openstack/dnsmasq-dns-675f4bcbfc-7pn5f" Nov 25 15:17:40 crc kubenswrapper[4890]: I1125 15:17:40.885551 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6c5bcbf-f90f-4d40-819d-c0bc8a5a47fa-config\") pod \"dnsmasq-dns-675f4bcbfc-7pn5f\" (UID: \"a6c5bcbf-f90f-4d40-819d-c0bc8a5a47fa\") " pod="openstack/dnsmasq-dns-675f4bcbfc-7pn5f" Nov 25 15:17:40 crc kubenswrapper[4890]: I1125 15:17:40.987589 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2f7ef20-6da9-4d06-a733-4a529aecffb4-config\") pod \"dnsmasq-dns-78dd6ddcc-6fkdh\" (UID: \"c2f7ef20-6da9-4d06-a733-4a529aecffb4\") " pod="openstack/dnsmasq-dns-78dd6ddcc-6fkdh" Nov 25 15:17:40 crc kubenswrapper[4890]: I1125 15:17:40.987645 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtbps\" (UniqueName: \"kubernetes.io/projected/a6c5bcbf-f90f-4d40-819d-c0bc8a5a47fa-kube-api-access-jtbps\") pod \"dnsmasq-dns-675f4bcbfc-7pn5f\" (UID: \"a6c5bcbf-f90f-4d40-819d-c0bc8a5a47fa\") " pod="openstack/dnsmasq-dns-675f4bcbfc-7pn5f" Nov 25 15:17:40 crc kubenswrapper[4890]: I1125 15:17:40.987693 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c2f7ef20-6da9-4d06-a733-4a529aecffb4-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-6fkdh\" (UID: \"c2f7ef20-6da9-4d06-a733-4a529aecffb4\") " pod="openstack/dnsmasq-dns-78dd6ddcc-6fkdh" Nov 25 15:17:40 crc kubenswrapper[4890]: I1125 15:17:40.987711 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6c5bcbf-f90f-4d40-819d-c0bc8a5a47fa-config\") pod \"dnsmasq-dns-675f4bcbfc-7pn5f\" (UID: \"a6c5bcbf-f90f-4d40-819d-c0bc8a5a47fa\") " pod="openstack/dnsmasq-dns-675f4bcbfc-7pn5f" Nov 25 15:17:40 crc kubenswrapper[4890]: I1125 15:17:40.987749 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l26cv\" (UniqueName: \"kubernetes.io/projected/c2f7ef20-6da9-4d06-a733-4a529aecffb4-kube-api-access-l26cv\") pod \"dnsmasq-dns-78dd6ddcc-6fkdh\" (UID: \"c2f7ef20-6da9-4d06-a733-4a529aecffb4\") " pod="openstack/dnsmasq-dns-78dd6ddcc-6fkdh" Nov 25 15:17:40 crc kubenswrapper[4890]: I1125 15:17:40.988812 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6c5bcbf-f90f-4d40-819d-c0bc8a5a47fa-config\") pod \"dnsmasq-dns-675f4bcbfc-7pn5f\" (UID: \"a6c5bcbf-f90f-4d40-819d-c0bc8a5a47fa\") " pod="openstack/dnsmasq-dns-675f4bcbfc-7pn5f" Nov 25 15:17:41 crc kubenswrapper[4890]: I1125 15:17:41.005899 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtbps\" (UniqueName: \"kubernetes.io/projected/a6c5bcbf-f90f-4d40-819d-c0bc8a5a47fa-kube-api-access-jtbps\") pod \"dnsmasq-dns-675f4bcbfc-7pn5f\" (UID: \"a6c5bcbf-f90f-4d40-819d-c0bc8a5a47fa\") " pod="openstack/dnsmasq-dns-675f4bcbfc-7pn5f" Nov 25 15:17:41 crc kubenswrapper[4890]: I1125 15:17:41.043599 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-7pn5f" Nov 25 15:17:41 crc kubenswrapper[4890]: I1125 15:17:41.088846 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2f7ef20-6da9-4d06-a733-4a529aecffb4-config\") pod \"dnsmasq-dns-78dd6ddcc-6fkdh\" (UID: \"c2f7ef20-6da9-4d06-a733-4a529aecffb4\") " pod="openstack/dnsmasq-dns-78dd6ddcc-6fkdh" Nov 25 15:17:41 crc kubenswrapper[4890]: I1125 15:17:41.089241 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c2f7ef20-6da9-4d06-a733-4a529aecffb4-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-6fkdh\" (UID: \"c2f7ef20-6da9-4d06-a733-4a529aecffb4\") " pod="openstack/dnsmasq-dns-78dd6ddcc-6fkdh" Nov 25 15:17:41 crc kubenswrapper[4890]: I1125 15:17:41.089284 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l26cv\" (UniqueName: \"kubernetes.io/projected/c2f7ef20-6da9-4d06-a733-4a529aecffb4-kube-api-access-l26cv\") pod \"dnsmasq-dns-78dd6ddcc-6fkdh\" (UID: \"c2f7ef20-6da9-4d06-a733-4a529aecffb4\") " pod="openstack/dnsmasq-dns-78dd6ddcc-6fkdh" Nov 25 15:17:41 crc kubenswrapper[4890]: I1125 15:17:41.089862 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2f7ef20-6da9-4d06-a733-4a529aecffb4-config\") pod \"dnsmasq-dns-78dd6ddcc-6fkdh\" (UID: \"c2f7ef20-6da9-4d06-a733-4a529aecffb4\") " pod="openstack/dnsmasq-dns-78dd6ddcc-6fkdh" Nov 25 15:17:41 crc kubenswrapper[4890]: I1125 15:17:41.090119 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c2f7ef20-6da9-4d06-a733-4a529aecffb4-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-6fkdh\" (UID: \"c2f7ef20-6da9-4d06-a733-4a529aecffb4\") " pod="openstack/dnsmasq-dns-78dd6ddcc-6fkdh" Nov 25 15:17:41 crc kubenswrapper[4890]: I1125 15:17:41.109677 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l26cv\" (UniqueName: \"kubernetes.io/projected/c2f7ef20-6da9-4d06-a733-4a529aecffb4-kube-api-access-l26cv\") pod \"dnsmasq-dns-78dd6ddcc-6fkdh\" (UID: \"c2f7ef20-6da9-4d06-a733-4a529aecffb4\") " pod="openstack/dnsmasq-dns-78dd6ddcc-6fkdh" Nov 25 15:17:41 crc kubenswrapper[4890]: I1125 15:17:41.145870 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-6fkdh" Nov 25 15:17:41 crc kubenswrapper[4890]: I1125 15:17:41.499204 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-7pn5f"] Nov 25 15:17:41 crc kubenswrapper[4890]: W1125 15:17:41.502613 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6c5bcbf_f90f_4d40_819d_c0bc8a5a47fa.slice/crio-391ae6f417a13745e5a1e5e36836a2f2c902909d118b7a39eef9a309b7f9466a WatchSource:0}: Error finding container 391ae6f417a13745e5a1e5e36836a2f2c902909d118b7a39eef9a309b7f9466a: Status 404 returned error can't find the container with id 391ae6f417a13745e5a1e5e36836a2f2c902909d118b7a39eef9a309b7f9466a Nov 25 15:17:41 crc kubenswrapper[4890]: I1125 15:17:41.600724 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-6fkdh"] Nov 25 15:17:41 crc kubenswrapper[4890]: W1125 15:17:41.609271 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc2f7ef20_6da9_4d06_a733_4a529aecffb4.slice/crio-dfa7ec5ff51ebc8667cd597ac8e2559da9365067699eabf6c10f258657d85c42 WatchSource:0}: Error finding container dfa7ec5ff51ebc8667cd597ac8e2559da9365067699eabf6c10f258657d85c42: Status 404 returned error can't find the container with id dfa7ec5ff51ebc8667cd597ac8e2559da9365067699eabf6c10f258657d85c42 Nov 25 15:17:41 crc kubenswrapper[4890]: I1125 15:17:41.907775 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-7pn5f" event={"ID":"a6c5bcbf-f90f-4d40-819d-c0bc8a5a47fa","Type":"ContainerStarted","Data":"391ae6f417a13745e5a1e5e36836a2f2c902909d118b7a39eef9a309b7f9466a"} Nov 25 15:17:41 crc kubenswrapper[4890]: I1125 15:17:41.909371 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-6fkdh" event={"ID":"c2f7ef20-6da9-4d06-a733-4a529aecffb4","Type":"ContainerStarted","Data":"dfa7ec5ff51ebc8667cd597ac8e2559da9365067699eabf6c10f258657d85c42"} Nov 25 15:17:43 crc kubenswrapper[4890]: I1125 15:17:43.928742 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-7pn5f"] Nov 25 15:17:43 crc kubenswrapper[4890]: I1125 15:17:43.961060 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-7nfql"] Nov 25 15:17:43 crc kubenswrapper[4890]: I1125 15:17:43.962489 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-7nfql" Nov 25 15:17:43 crc kubenswrapper[4890]: I1125 15:17:43.981892 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-7nfql"] Nov 25 15:17:44 crc kubenswrapper[4890]: I1125 15:17:44.054212 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6qg9\" (UniqueName: \"kubernetes.io/projected/536d18a7-2067-46d8-864c-8c84be2a87a8-kube-api-access-x6qg9\") pod \"dnsmasq-dns-666b6646f7-7nfql\" (UID: \"536d18a7-2067-46d8-864c-8c84be2a87a8\") " pod="openstack/dnsmasq-dns-666b6646f7-7nfql" Nov 25 15:17:44 crc kubenswrapper[4890]: I1125 15:17:44.054300 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/536d18a7-2067-46d8-864c-8c84be2a87a8-dns-svc\") pod \"dnsmasq-dns-666b6646f7-7nfql\" (UID: \"536d18a7-2067-46d8-864c-8c84be2a87a8\") " pod="openstack/dnsmasq-dns-666b6646f7-7nfql" Nov 25 15:17:44 crc kubenswrapper[4890]: I1125 15:17:44.054346 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/536d18a7-2067-46d8-864c-8c84be2a87a8-config\") pod \"dnsmasq-dns-666b6646f7-7nfql\" (UID: \"536d18a7-2067-46d8-864c-8c84be2a87a8\") " pod="openstack/dnsmasq-dns-666b6646f7-7nfql" Nov 25 15:17:44 crc kubenswrapper[4890]: I1125 15:17:44.155691 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/536d18a7-2067-46d8-864c-8c84be2a87a8-dns-svc\") pod \"dnsmasq-dns-666b6646f7-7nfql\" (UID: \"536d18a7-2067-46d8-864c-8c84be2a87a8\") " pod="openstack/dnsmasq-dns-666b6646f7-7nfql" Nov 25 15:17:44 crc kubenswrapper[4890]: I1125 15:17:44.155796 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/536d18a7-2067-46d8-864c-8c84be2a87a8-config\") pod \"dnsmasq-dns-666b6646f7-7nfql\" (UID: \"536d18a7-2067-46d8-864c-8c84be2a87a8\") " pod="openstack/dnsmasq-dns-666b6646f7-7nfql" Nov 25 15:17:44 crc kubenswrapper[4890]: I1125 15:17:44.155841 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6qg9\" (UniqueName: \"kubernetes.io/projected/536d18a7-2067-46d8-864c-8c84be2a87a8-kube-api-access-x6qg9\") pod \"dnsmasq-dns-666b6646f7-7nfql\" (UID: \"536d18a7-2067-46d8-864c-8c84be2a87a8\") " pod="openstack/dnsmasq-dns-666b6646f7-7nfql" Nov 25 15:17:44 crc kubenswrapper[4890]: I1125 15:17:44.157318 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/536d18a7-2067-46d8-864c-8c84be2a87a8-dns-svc\") pod \"dnsmasq-dns-666b6646f7-7nfql\" (UID: \"536d18a7-2067-46d8-864c-8c84be2a87a8\") " pod="openstack/dnsmasq-dns-666b6646f7-7nfql" Nov 25 15:17:44 crc kubenswrapper[4890]: I1125 15:17:44.157347 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/536d18a7-2067-46d8-864c-8c84be2a87a8-config\") pod \"dnsmasq-dns-666b6646f7-7nfql\" (UID: \"536d18a7-2067-46d8-864c-8c84be2a87a8\") " pod="openstack/dnsmasq-dns-666b6646f7-7nfql" Nov 25 15:17:44 crc kubenswrapper[4890]: I1125 15:17:44.199246 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6qg9\" (UniqueName: \"kubernetes.io/projected/536d18a7-2067-46d8-864c-8c84be2a87a8-kube-api-access-x6qg9\") pod \"dnsmasq-dns-666b6646f7-7nfql\" (UID: \"536d18a7-2067-46d8-864c-8c84be2a87a8\") " pod="openstack/dnsmasq-dns-666b6646f7-7nfql" Nov 25 15:17:44 crc kubenswrapper[4890]: I1125 15:17:44.274950 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-6fkdh"] Nov 25 15:17:44 crc kubenswrapper[4890]: I1125 15:17:44.298415 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-27vl6"] Nov 25 15:17:44 crc kubenswrapper[4890]: I1125 15:17:44.299831 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-27vl6" Nov 25 15:17:44 crc kubenswrapper[4890]: I1125 15:17:44.300780 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-7nfql" Nov 25 15:17:44 crc kubenswrapper[4890]: I1125 15:17:44.331831 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-27vl6"] Nov 25 15:17:44 crc kubenswrapper[4890]: I1125 15:17:44.358914 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4w4l\" (UniqueName: \"kubernetes.io/projected/f467b253-fb57-465a-8e67-5168b2dc0e55-kube-api-access-t4w4l\") pod \"dnsmasq-dns-57d769cc4f-27vl6\" (UID: \"f467b253-fb57-465a-8e67-5168b2dc0e55\") " pod="openstack/dnsmasq-dns-57d769cc4f-27vl6" Nov 25 15:17:44 crc kubenswrapper[4890]: I1125 15:17:44.358966 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f467b253-fb57-465a-8e67-5168b2dc0e55-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-27vl6\" (UID: \"f467b253-fb57-465a-8e67-5168b2dc0e55\") " pod="openstack/dnsmasq-dns-57d769cc4f-27vl6" Nov 25 15:17:44 crc kubenswrapper[4890]: I1125 15:17:44.359023 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f467b253-fb57-465a-8e67-5168b2dc0e55-config\") pod \"dnsmasq-dns-57d769cc4f-27vl6\" (UID: \"f467b253-fb57-465a-8e67-5168b2dc0e55\") " pod="openstack/dnsmasq-dns-57d769cc4f-27vl6" Nov 25 15:17:44 crc kubenswrapper[4890]: I1125 15:17:44.460327 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4w4l\" (UniqueName: \"kubernetes.io/projected/f467b253-fb57-465a-8e67-5168b2dc0e55-kube-api-access-t4w4l\") pod \"dnsmasq-dns-57d769cc4f-27vl6\" (UID: \"f467b253-fb57-465a-8e67-5168b2dc0e55\") " pod="openstack/dnsmasq-dns-57d769cc4f-27vl6" Nov 25 15:17:44 crc kubenswrapper[4890]: I1125 15:17:44.460801 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f467b253-fb57-465a-8e67-5168b2dc0e55-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-27vl6\" (UID: \"f467b253-fb57-465a-8e67-5168b2dc0e55\") " pod="openstack/dnsmasq-dns-57d769cc4f-27vl6" Nov 25 15:17:44 crc kubenswrapper[4890]: I1125 15:17:44.460872 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f467b253-fb57-465a-8e67-5168b2dc0e55-config\") pod \"dnsmasq-dns-57d769cc4f-27vl6\" (UID: \"f467b253-fb57-465a-8e67-5168b2dc0e55\") " pod="openstack/dnsmasq-dns-57d769cc4f-27vl6" Nov 25 15:17:44 crc kubenswrapper[4890]: I1125 15:17:44.462037 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f467b253-fb57-465a-8e67-5168b2dc0e55-config\") pod \"dnsmasq-dns-57d769cc4f-27vl6\" (UID: \"f467b253-fb57-465a-8e67-5168b2dc0e55\") " pod="openstack/dnsmasq-dns-57d769cc4f-27vl6" Nov 25 15:17:44 crc kubenswrapper[4890]: I1125 15:17:44.462542 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f467b253-fb57-465a-8e67-5168b2dc0e55-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-27vl6\" (UID: \"f467b253-fb57-465a-8e67-5168b2dc0e55\") " pod="openstack/dnsmasq-dns-57d769cc4f-27vl6" Nov 25 15:17:44 crc kubenswrapper[4890]: I1125 15:17:44.499327 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4w4l\" (UniqueName: \"kubernetes.io/projected/f467b253-fb57-465a-8e67-5168b2dc0e55-kube-api-access-t4w4l\") pod \"dnsmasq-dns-57d769cc4f-27vl6\" (UID: \"f467b253-fb57-465a-8e67-5168b2dc0e55\") " pod="openstack/dnsmasq-dns-57d769cc4f-27vl6" Nov 25 15:17:44 crc kubenswrapper[4890]: I1125 15:17:44.634055 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-27vl6" Nov 25 15:17:44 crc kubenswrapper[4890]: I1125 15:17:44.968340 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-7nfql"] Nov 25 15:17:45 crc kubenswrapper[4890]: W1125 15:17:45.006576 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod536d18a7_2067_46d8_864c_8c84be2a87a8.slice/crio-cd2f43755d4af48a478a8b9adda57cb42a52662faf40d5d6938059b090f834ff WatchSource:0}: Error finding container cd2f43755d4af48a478a8b9adda57cb42a52662faf40d5d6938059b090f834ff: Status 404 returned error can't find the container with id cd2f43755d4af48a478a8b9adda57cb42a52662faf40d5d6938059b090f834ff Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.107211 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-27vl6"] Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.135254 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.137324 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.139861 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.149510 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-9cgrd" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.149625 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.149861 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.150058 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.150088 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.150308 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.151125 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.280841 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-config-data\") pod \"rabbitmq-server-0\" (UID: \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\") " pod="openstack/rabbitmq-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.280983 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\") " pod="openstack/rabbitmq-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.281022 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\") " pod="openstack/rabbitmq-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.281051 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\") " pod="openstack/rabbitmq-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.281089 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-server-conf\") pod \"rabbitmq-server-0\" (UID: \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\") " pod="openstack/rabbitmq-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.281130 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\") " pod="openstack/rabbitmq-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.281253 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\") " pod="openstack/rabbitmq-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.281381 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\") " pod="openstack/rabbitmq-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.281467 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\") " pod="openstack/rabbitmq-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.281547 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-pod-info\") pod \"rabbitmq-server-0\" (UID: \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\") " pod="openstack/rabbitmq-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.281653 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbr45\" (UniqueName: \"kubernetes.io/projected/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-kube-api-access-wbr45\") pod \"rabbitmq-server-0\" (UID: \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\") " pod="openstack/rabbitmq-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.382865 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\") " pod="openstack/rabbitmq-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.382932 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-pod-info\") pod \"rabbitmq-server-0\" (UID: \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\") " pod="openstack/rabbitmq-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.382966 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbr45\" (UniqueName: \"kubernetes.io/projected/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-kube-api-access-wbr45\") pod \"rabbitmq-server-0\" (UID: \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\") " pod="openstack/rabbitmq-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.383036 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-config-data\") pod \"rabbitmq-server-0\" (UID: \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\") " pod="openstack/rabbitmq-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.383071 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\") " pod="openstack/rabbitmq-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.383104 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\") " pod="openstack/rabbitmq-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.383130 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\") " pod="openstack/rabbitmq-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.383198 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-server-conf\") pod \"rabbitmq-server-0\" (UID: \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\") " pod="openstack/rabbitmq-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.383239 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\") " pod="openstack/rabbitmq-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.383268 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\") " pod="openstack/rabbitmq-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.383296 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\") " pod="openstack/rabbitmq-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.385509 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\") " pod="openstack/rabbitmq-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.385720 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\") " pod="openstack/rabbitmq-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.385889 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\") " pod="openstack/rabbitmq-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.386321 4890 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.386589 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-server-conf\") pod \"rabbitmq-server-0\" (UID: \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\") " pod="openstack/rabbitmq-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.387443 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-config-data\") pod \"rabbitmq-server-0\" (UID: \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\") " pod="openstack/rabbitmq-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.391219 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\") " pod="openstack/rabbitmq-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.391621 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-pod-info\") pod \"rabbitmq-server-0\" (UID: \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\") " pod="openstack/rabbitmq-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.391703 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\") " pod="openstack/rabbitmq-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.401411 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbr45\" (UniqueName: \"kubernetes.io/projected/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-kube-api-access-wbr45\") pod \"rabbitmq-server-0\" (UID: \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\") " pod="openstack/rabbitmq-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.405517 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\") " pod="openstack/rabbitmq-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.416314 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\") " pod="openstack/rabbitmq-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.428445 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.445031 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.459258 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.460621 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.461878 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.462407 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.462882 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.464769 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.466042 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-2k6hj" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.473942 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.491138 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.592990 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c14ce73e-1a95-4f41-b6b6-33478079806f-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c14ce73e-1a95-4f41-b6b6-33478079806f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.593581 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c14ce73e-1a95-4f41-b6b6-33478079806f-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c14ce73e-1a95-4f41-b6b6-33478079806f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.593605 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c14ce73e-1a95-4f41-b6b6-33478079806f-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c14ce73e-1a95-4f41-b6b6-33478079806f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.593625 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c14ce73e-1a95-4f41-b6b6-33478079806f-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c14ce73e-1a95-4f41-b6b6-33478079806f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.593692 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c14ce73e-1a95-4f41-b6b6-33478079806f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.593718 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tcxvg\" (UniqueName: \"kubernetes.io/projected/c14ce73e-1a95-4f41-b6b6-33478079806f-kube-api-access-tcxvg\") pod \"rabbitmq-cell1-server-0\" (UID: \"c14ce73e-1a95-4f41-b6b6-33478079806f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.593748 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c14ce73e-1a95-4f41-b6b6-33478079806f-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c14ce73e-1a95-4f41-b6b6-33478079806f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.593772 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c14ce73e-1a95-4f41-b6b6-33478079806f-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c14ce73e-1a95-4f41-b6b6-33478079806f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.593807 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c14ce73e-1a95-4f41-b6b6-33478079806f-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c14ce73e-1a95-4f41-b6b6-33478079806f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.593835 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c14ce73e-1a95-4f41-b6b6-33478079806f-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c14ce73e-1a95-4f41-b6b6-33478079806f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.593856 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c14ce73e-1a95-4f41-b6b6-33478079806f-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c14ce73e-1a95-4f41-b6b6-33478079806f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.699100 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c14ce73e-1a95-4f41-b6b6-33478079806f-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c14ce73e-1a95-4f41-b6b6-33478079806f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.699264 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c14ce73e-1a95-4f41-b6b6-33478079806f-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c14ce73e-1a95-4f41-b6b6-33478079806f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.699294 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c14ce73e-1a95-4f41-b6b6-33478079806f-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c14ce73e-1a95-4f41-b6b6-33478079806f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.699317 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c14ce73e-1a95-4f41-b6b6-33478079806f-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c14ce73e-1a95-4f41-b6b6-33478079806f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.699339 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c14ce73e-1a95-4f41-b6b6-33478079806f-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c14ce73e-1a95-4f41-b6b6-33478079806f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.699410 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c14ce73e-1a95-4f41-b6b6-33478079806f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.699441 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tcxvg\" (UniqueName: \"kubernetes.io/projected/c14ce73e-1a95-4f41-b6b6-33478079806f-kube-api-access-tcxvg\") pod \"rabbitmq-cell1-server-0\" (UID: \"c14ce73e-1a95-4f41-b6b6-33478079806f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.699478 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c14ce73e-1a95-4f41-b6b6-33478079806f-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c14ce73e-1a95-4f41-b6b6-33478079806f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.699501 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c14ce73e-1a95-4f41-b6b6-33478079806f-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c14ce73e-1a95-4f41-b6b6-33478079806f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.699541 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c14ce73e-1a95-4f41-b6b6-33478079806f-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c14ce73e-1a95-4f41-b6b6-33478079806f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.699576 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c14ce73e-1a95-4f41-b6b6-33478079806f-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c14ce73e-1a95-4f41-b6b6-33478079806f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.701175 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c14ce73e-1a95-4f41-b6b6-33478079806f-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c14ce73e-1a95-4f41-b6b6-33478079806f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.701453 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c14ce73e-1a95-4f41-b6b6-33478079806f-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c14ce73e-1a95-4f41-b6b6-33478079806f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.701681 4890 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c14ce73e-1a95-4f41-b6b6-33478079806f\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.702090 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c14ce73e-1a95-4f41-b6b6-33478079806f-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c14ce73e-1a95-4f41-b6b6-33478079806f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.703040 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c14ce73e-1a95-4f41-b6b6-33478079806f-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c14ce73e-1a95-4f41-b6b6-33478079806f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.703433 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c14ce73e-1a95-4f41-b6b6-33478079806f-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c14ce73e-1a95-4f41-b6b6-33478079806f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.708505 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c14ce73e-1a95-4f41-b6b6-33478079806f-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c14ce73e-1a95-4f41-b6b6-33478079806f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.710993 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c14ce73e-1a95-4f41-b6b6-33478079806f-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c14ce73e-1a95-4f41-b6b6-33478079806f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.712184 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c14ce73e-1a95-4f41-b6b6-33478079806f-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c14ce73e-1a95-4f41-b6b6-33478079806f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.718525 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c14ce73e-1a95-4f41-b6b6-33478079806f-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c14ce73e-1a95-4f41-b6b6-33478079806f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.721637 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tcxvg\" (UniqueName: \"kubernetes.io/projected/c14ce73e-1a95-4f41-b6b6-33478079806f-kube-api-access-tcxvg\") pod \"rabbitmq-cell1-server-0\" (UID: \"c14ce73e-1a95-4f41-b6b6-33478079806f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.726020 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c14ce73e-1a95-4f41-b6b6-33478079806f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.784788 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.843419 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 15:17:45 crc kubenswrapper[4890]: W1125 15:17:45.849347 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podae7f09bf_3fc2_4da7_9b41_9f4564dd7562.slice/crio-c4b27d4bef264cb091befd2683b8d338325300986c622b7abf1025f011730555 WatchSource:0}: Error finding container c4b27d4bef264cb091befd2683b8d338325300986c622b7abf1025f011730555: Status 404 returned error can't find the container with id c4b27d4bef264cb091befd2683b8d338325300986c622b7abf1025f011730555 Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.980134 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-27vl6" event={"ID":"f467b253-fb57-465a-8e67-5168b2dc0e55","Type":"ContainerStarted","Data":"987bab745511b1ff1afe4fb20a103a10631d862c6389768fc840303ba7fca9a2"} Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.981246 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-7nfql" event={"ID":"536d18a7-2067-46d8-864c-8c84be2a87a8","Type":"ContainerStarted","Data":"cd2f43755d4af48a478a8b9adda57cb42a52662faf40d5d6938059b090f834ff"} Nov 25 15:17:45 crc kubenswrapper[4890]: I1125 15:17:45.983268 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562","Type":"ContainerStarted","Data":"c4b27d4bef264cb091befd2683b8d338325300986c622b7abf1025f011730555"} Nov 25 15:17:46 crc kubenswrapper[4890]: I1125 15:17:46.341911 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 15:17:46 crc kubenswrapper[4890]: W1125 15:17:46.375021 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc14ce73e_1a95_4f41_b6b6_33478079806f.slice/crio-251d05656390cb8a8d176016a4b9a2e1bdd2cb455eac2988bd187eb2e89c8336 WatchSource:0}: Error finding container 251d05656390cb8a8d176016a4b9a2e1bdd2cb455eac2988bd187eb2e89c8336: Status 404 returned error can't find the container with id 251d05656390cb8a8d176016a4b9a2e1bdd2cb455eac2988bd187eb2e89c8336 Nov 25 15:17:46 crc kubenswrapper[4890]: I1125 15:17:46.995348 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c14ce73e-1a95-4f41-b6b6-33478079806f","Type":"ContainerStarted","Data":"251d05656390cb8a8d176016a4b9a2e1bdd2cb455eac2988bd187eb2e89c8336"} Nov 25 15:17:47 crc kubenswrapper[4890]: I1125 15:17:47.125495 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Nov 25 15:17:47 crc kubenswrapper[4890]: I1125 15:17:47.126947 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 25 15:17:47 crc kubenswrapper[4890]: I1125 15:17:47.130530 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Nov 25 15:17:47 crc kubenswrapper[4890]: I1125 15:17:47.130819 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Nov 25 15:17:47 crc kubenswrapper[4890]: I1125 15:17:47.132331 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Nov 25 15:17:47 crc kubenswrapper[4890]: I1125 15:17:47.132534 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-4gbz6" Nov 25 15:17:47 crc kubenswrapper[4890]: I1125 15:17:47.144927 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Nov 25 15:17:47 crc kubenswrapper[4890]: I1125 15:17:47.149153 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 25 15:17:47 crc kubenswrapper[4890]: I1125 15:17:47.325997 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90881a25-bc18-4b43-a226-a44e3165cb3a-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"90881a25-bc18-4b43-a226-a44e3165cb3a\") " pod="openstack/openstack-galera-0" Nov 25 15:17:47 crc kubenswrapper[4890]: I1125 15:17:47.326075 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"90881a25-bc18-4b43-a226-a44e3165cb3a\") " pod="openstack/openstack-galera-0" Nov 25 15:17:47 crc kubenswrapper[4890]: I1125 15:17:47.326108 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmrjv\" (UniqueName: \"kubernetes.io/projected/90881a25-bc18-4b43-a226-a44e3165cb3a-kube-api-access-dmrjv\") pod \"openstack-galera-0\" (UID: \"90881a25-bc18-4b43-a226-a44e3165cb3a\") " pod="openstack/openstack-galera-0" Nov 25 15:17:47 crc kubenswrapper[4890]: I1125 15:17:47.326142 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/90881a25-bc18-4b43-a226-a44e3165cb3a-kolla-config\") pod \"openstack-galera-0\" (UID: \"90881a25-bc18-4b43-a226-a44e3165cb3a\") " pod="openstack/openstack-galera-0" Nov 25 15:17:47 crc kubenswrapper[4890]: I1125 15:17:47.327425 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/90881a25-bc18-4b43-a226-a44e3165cb3a-config-data-generated\") pod \"openstack-galera-0\" (UID: \"90881a25-bc18-4b43-a226-a44e3165cb3a\") " pod="openstack/openstack-galera-0" Nov 25 15:17:47 crc kubenswrapper[4890]: I1125 15:17:47.327462 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/90881a25-bc18-4b43-a226-a44e3165cb3a-config-data-default\") pod \"openstack-galera-0\" (UID: \"90881a25-bc18-4b43-a226-a44e3165cb3a\") " pod="openstack/openstack-galera-0" Nov 25 15:17:47 crc kubenswrapper[4890]: I1125 15:17:47.327538 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/90881a25-bc18-4b43-a226-a44e3165cb3a-operator-scripts\") pod \"openstack-galera-0\" (UID: \"90881a25-bc18-4b43-a226-a44e3165cb3a\") " pod="openstack/openstack-galera-0" Nov 25 15:17:47 crc kubenswrapper[4890]: I1125 15:17:47.327583 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/90881a25-bc18-4b43-a226-a44e3165cb3a-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"90881a25-bc18-4b43-a226-a44e3165cb3a\") " pod="openstack/openstack-galera-0" Nov 25 15:17:47 crc kubenswrapper[4890]: I1125 15:17:47.429316 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/90881a25-bc18-4b43-a226-a44e3165cb3a-config-data-generated\") pod \"openstack-galera-0\" (UID: \"90881a25-bc18-4b43-a226-a44e3165cb3a\") " pod="openstack/openstack-galera-0" Nov 25 15:17:47 crc kubenswrapper[4890]: I1125 15:17:47.429374 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/90881a25-bc18-4b43-a226-a44e3165cb3a-config-data-default\") pod \"openstack-galera-0\" (UID: \"90881a25-bc18-4b43-a226-a44e3165cb3a\") " pod="openstack/openstack-galera-0" Nov 25 15:17:47 crc kubenswrapper[4890]: I1125 15:17:47.429404 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/90881a25-bc18-4b43-a226-a44e3165cb3a-operator-scripts\") pod \"openstack-galera-0\" (UID: \"90881a25-bc18-4b43-a226-a44e3165cb3a\") " pod="openstack/openstack-galera-0" Nov 25 15:17:47 crc kubenswrapper[4890]: I1125 15:17:47.429456 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/90881a25-bc18-4b43-a226-a44e3165cb3a-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"90881a25-bc18-4b43-a226-a44e3165cb3a\") " pod="openstack/openstack-galera-0" Nov 25 15:17:47 crc kubenswrapper[4890]: I1125 15:17:47.429559 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90881a25-bc18-4b43-a226-a44e3165cb3a-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"90881a25-bc18-4b43-a226-a44e3165cb3a\") " pod="openstack/openstack-galera-0" Nov 25 15:17:47 crc kubenswrapper[4890]: I1125 15:17:47.429626 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"90881a25-bc18-4b43-a226-a44e3165cb3a\") " pod="openstack/openstack-galera-0" Nov 25 15:17:47 crc kubenswrapper[4890]: I1125 15:17:47.429651 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmrjv\" (UniqueName: \"kubernetes.io/projected/90881a25-bc18-4b43-a226-a44e3165cb3a-kube-api-access-dmrjv\") pod \"openstack-galera-0\" (UID: \"90881a25-bc18-4b43-a226-a44e3165cb3a\") " pod="openstack/openstack-galera-0" Nov 25 15:17:47 crc kubenswrapper[4890]: I1125 15:17:47.429683 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/90881a25-bc18-4b43-a226-a44e3165cb3a-kolla-config\") pod \"openstack-galera-0\" (UID: \"90881a25-bc18-4b43-a226-a44e3165cb3a\") " pod="openstack/openstack-galera-0" Nov 25 15:17:47 crc kubenswrapper[4890]: I1125 15:17:47.429812 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/90881a25-bc18-4b43-a226-a44e3165cb3a-config-data-generated\") pod \"openstack-galera-0\" (UID: \"90881a25-bc18-4b43-a226-a44e3165cb3a\") " pod="openstack/openstack-galera-0" Nov 25 15:17:47 crc kubenswrapper[4890]: I1125 15:17:47.430450 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/90881a25-bc18-4b43-a226-a44e3165cb3a-kolla-config\") pod \"openstack-galera-0\" (UID: \"90881a25-bc18-4b43-a226-a44e3165cb3a\") " pod="openstack/openstack-galera-0" Nov 25 15:17:47 crc kubenswrapper[4890]: I1125 15:17:47.431634 4890 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"90881a25-bc18-4b43-a226-a44e3165cb3a\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/openstack-galera-0" Nov 25 15:17:47 crc kubenswrapper[4890]: I1125 15:17:47.431686 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/90881a25-bc18-4b43-a226-a44e3165cb3a-config-data-default\") pod \"openstack-galera-0\" (UID: \"90881a25-bc18-4b43-a226-a44e3165cb3a\") " pod="openstack/openstack-galera-0" Nov 25 15:17:47 crc kubenswrapper[4890]: I1125 15:17:47.434620 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/90881a25-bc18-4b43-a226-a44e3165cb3a-operator-scripts\") pod \"openstack-galera-0\" (UID: \"90881a25-bc18-4b43-a226-a44e3165cb3a\") " pod="openstack/openstack-galera-0" Nov 25 15:17:47 crc kubenswrapper[4890]: I1125 15:17:47.438024 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90881a25-bc18-4b43-a226-a44e3165cb3a-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"90881a25-bc18-4b43-a226-a44e3165cb3a\") " pod="openstack/openstack-galera-0" Nov 25 15:17:47 crc kubenswrapper[4890]: I1125 15:17:47.460521 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/90881a25-bc18-4b43-a226-a44e3165cb3a-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"90881a25-bc18-4b43-a226-a44e3165cb3a\") " pod="openstack/openstack-galera-0" Nov 25 15:17:47 crc kubenswrapper[4890]: I1125 15:17:47.460520 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmrjv\" (UniqueName: \"kubernetes.io/projected/90881a25-bc18-4b43-a226-a44e3165cb3a-kube-api-access-dmrjv\") pod \"openstack-galera-0\" (UID: \"90881a25-bc18-4b43-a226-a44e3165cb3a\") " pod="openstack/openstack-galera-0" Nov 25 15:17:47 crc kubenswrapper[4890]: I1125 15:17:47.489495 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"90881a25-bc18-4b43-a226-a44e3165cb3a\") " pod="openstack/openstack-galera-0" Nov 25 15:17:47 crc kubenswrapper[4890]: I1125 15:17:47.503112 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.042982 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.496437 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.499905 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.503034 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.503387 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.503488 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.503965 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-wxzlp" Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.521105 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.654081 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab\") " pod="openstack/openstack-cell1-galera-0" Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.654134 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab\") " pod="openstack/openstack-cell1-galera-0" Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.654208 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab\") " pod="openstack/openstack-cell1-galera-0" Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.654235 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab\") " pod="openstack/openstack-cell1-galera-0" Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.654268 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab\") " pod="openstack/openstack-cell1-galera-0" Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.659075 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab\") " pod="openstack/openstack-cell1-galera-0" Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.659149 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab\") " pod="openstack/openstack-cell1-galera-0" Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.659265 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92q8t\" (UniqueName: \"kubernetes.io/projected/9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab-kube-api-access-92q8t\") pod \"openstack-cell1-galera-0\" (UID: \"9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab\") " pod="openstack/openstack-cell1-galera-0" Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.705077 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.706508 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.710558 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.714534 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-xtbnh" Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.714848 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.742065 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.763321 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab\") " pod="openstack/openstack-cell1-galera-0" Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.763381 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab\") " pod="openstack/openstack-cell1-galera-0" Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.763428 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab\") " pod="openstack/openstack-cell1-galera-0" Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.763453 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab\") " pod="openstack/openstack-cell1-galera-0" Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.763482 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab\") " pod="openstack/openstack-cell1-galera-0" Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.763523 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab\") " pod="openstack/openstack-cell1-galera-0" Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.763542 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab\") " pod="openstack/openstack-cell1-galera-0" Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.763568 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92q8t\" (UniqueName: \"kubernetes.io/projected/9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab-kube-api-access-92q8t\") pod \"openstack-cell1-galera-0\" (UID: \"9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab\") " pod="openstack/openstack-cell1-galera-0" Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.764235 4890 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/openstack-cell1-galera-0" Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.765606 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab\") " pod="openstack/openstack-cell1-galera-0" Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.766631 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab\") " pod="openstack/openstack-cell1-galera-0" Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.766680 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab\") " pod="openstack/openstack-cell1-galera-0" Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.767029 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab\") " pod="openstack/openstack-cell1-galera-0" Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.775314 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab\") " pod="openstack/openstack-cell1-galera-0" Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.780716 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab\") " pod="openstack/openstack-cell1-galera-0" Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.785832 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92q8t\" (UniqueName: \"kubernetes.io/projected/9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab-kube-api-access-92q8t\") pod \"openstack-cell1-galera-0\" (UID: \"9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab\") " pod="openstack/openstack-cell1-galera-0" Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.810471 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab\") " pod="openstack/openstack-cell1-galera-0" Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.845884 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.865535 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxcfj\" (UniqueName: \"kubernetes.io/projected/c5630332-c335-4b9a-8486-e845c6baaa9c-kube-api-access-vxcfj\") pod \"memcached-0\" (UID: \"c5630332-c335-4b9a-8486-e845c6baaa9c\") " pod="openstack/memcached-0" Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.865604 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c5630332-c335-4b9a-8486-e845c6baaa9c-kolla-config\") pod \"memcached-0\" (UID: \"c5630332-c335-4b9a-8486-e845c6baaa9c\") " pod="openstack/memcached-0" Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.865727 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5630332-c335-4b9a-8486-e845c6baaa9c-combined-ca-bundle\") pod \"memcached-0\" (UID: \"c5630332-c335-4b9a-8486-e845c6baaa9c\") " pod="openstack/memcached-0" Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.865793 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/c5630332-c335-4b9a-8486-e845c6baaa9c-memcached-tls-certs\") pod \"memcached-0\" (UID: \"c5630332-c335-4b9a-8486-e845c6baaa9c\") " pod="openstack/memcached-0" Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.865816 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c5630332-c335-4b9a-8486-e845c6baaa9c-config-data\") pod \"memcached-0\" (UID: \"c5630332-c335-4b9a-8486-e845c6baaa9c\") " pod="openstack/memcached-0" Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.967460 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c5630332-c335-4b9a-8486-e845c6baaa9c-config-data\") pod \"memcached-0\" (UID: \"c5630332-c335-4b9a-8486-e845c6baaa9c\") " pod="openstack/memcached-0" Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.967515 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/c5630332-c335-4b9a-8486-e845c6baaa9c-memcached-tls-certs\") pod \"memcached-0\" (UID: \"c5630332-c335-4b9a-8486-e845c6baaa9c\") " pod="openstack/memcached-0" Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.967580 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxcfj\" (UniqueName: \"kubernetes.io/projected/c5630332-c335-4b9a-8486-e845c6baaa9c-kube-api-access-vxcfj\") pod \"memcached-0\" (UID: \"c5630332-c335-4b9a-8486-e845c6baaa9c\") " pod="openstack/memcached-0" Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.967606 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c5630332-c335-4b9a-8486-e845c6baaa9c-kolla-config\") pod \"memcached-0\" (UID: \"c5630332-c335-4b9a-8486-e845c6baaa9c\") " pod="openstack/memcached-0" Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.967654 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5630332-c335-4b9a-8486-e845c6baaa9c-combined-ca-bundle\") pod \"memcached-0\" (UID: \"c5630332-c335-4b9a-8486-e845c6baaa9c\") " pod="openstack/memcached-0" Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.970989 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c5630332-c335-4b9a-8486-e845c6baaa9c-config-data\") pod \"memcached-0\" (UID: \"c5630332-c335-4b9a-8486-e845c6baaa9c\") " pod="openstack/memcached-0" Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.971381 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c5630332-c335-4b9a-8486-e845c6baaa9c-kolla-config\") pod \"memcached-0\" (UID: \"c5630332-c335-4b9a-8486-e845c6baaa9c\") " pod="openstack/memcached-0" Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.973608 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5630332-c335-4b9a-8486-e845c6baaa9c-combined-ca-bundle\") pod \"memcached-0\" (UID: \"c5630332-c335-4b9a-8486-e845c6baaa9c\") " pod="openstack/memcached-0" Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.974348 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/c5630332-c335-4b9a-8486-e845c6baaa9c-memcached-tls-certs\") pod \"memcached-0\" (UID: \"c5630332-c335-4b9a-8486-e845c6baaa9c\") " pod="openstack/memcached-0" Nov 25 15:17:48 crc kubenswrapper[4890]: I1125 15:17:48.988194 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxcfj\" (UniqueName: \"kubernetes.io/projected/c5630332-c335-4b9a-8486-e845c6baaa9c-kube-api-access-vxcfj\") pod \"memcached-0\" (UID: \"c5630332-c335-4b9a-8486-e845c6baaa9c\") " pod="openstack/memcached-0" Nov 25 15:17:49 crc kubenswrapper[4890]: I1125 15:17:49.041003 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 25 15:17:50 crc kubenswrapper[4890]: I1125 15:17:50.818715 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 15:17:50 crc kubenswrapper[4890]: I1125 15:17:50.820391 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 25 15:17:50 crc kubenswrapper[4890]: I1125 15:17:50.827731 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-lwcrw" Nov 25 15:17:50 crc kubenswrapper[4890]: I1125 15:17:50.834012 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 15:17:50 crc kubenswrapper[4890]: I1125 15:17:50.999515 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mtcq\" (UniqueName: \"kubernetes.io/projected/24000ee2-b285-4f1a-afe1-75e643c1e172-kube-api-access-6mtcq\") pod \"kube-state-metrics-0\" (UID: \"24000ee2-b285-4f1a-afe1-75e643c1e172\") " pod="openstack/kube-state-metrics-0" Nov 25 15:17:51 crc kubenswrapper[4890]: I1125 15:17:51.101086 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mtcq\" (UniqueName: \"kubernetes.io/projected/24000ee2-b285-4f1a-afe1-75e643c1e172-kube-api-access-6mtcq\") pod \"kube-state-metrics-0\" (UID: \"24000ee2-b285-4f1a-afe1-75e643c1e172\") " pod="openstack/kube-state-metrics-0" Nov 25 15:17:51 crc kubenswrapper[4890]: I1125 15:17:51.137228 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mtcq\" (UniqueName: \"kubernetes.io/projected/24000ee2-b285-4f1a-afe1-75e643c1e172-kube-api-access-6mtcq\") pod \"kube-state-metrics-0\" (UID: \"24000ee2-b285-4f1a-afe1-75e643c1e172\") " pod="openstack/kube-state-metrics-0" Nov 25 15:17:51 crc kubenswrapper[4890]: I1125 15:17:51.178896 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.416510 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.420075 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.423559 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.424082 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.424217 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-z4bgs" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.424394 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.424782 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.425194 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.568082 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/49e51424-feba-4038-88a6-1cd56637984c-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"49e51424-feba-4038-88a6-1cd56637984c\") " pod="openstack/ovsdbserver-nb-0" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.568524 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49e51424-feba-4038-88a6-1cd56637984c-config\") pod \"ovsdbserver-nb-0\" (UID: \"49e51424-feba-4038-88a6-1cd56637984c\") " pod="openstack/ovsdbserver-nb-0" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.568547 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4h2fr\" (UniqueName: \"kubernetes.io/projected/49e51424-feba-4038-88a6-1cd56637984c-kube-api-access-4h2fr\") pod \"ovsdbserver-nb-0\" (UID: \"49e51424-feba-4038-88a6-1cd56637984c\") " pod="openstack/ovsdbserver-nb-0" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.568597 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"49e51424-feba-4038-88a6-1cd56637984c\") " pod="openstack/ovsdbserver-nb-0" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.568638 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/49e51424-feba-4038-88a6-1cd56637984c-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"49e51424-feba-4038-88a6-1cd56637984c\") " pod="openstack/ovsdbserver-nb-0" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.568659 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/49e51424-feba-4038-88a6-1cd56637984c-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"49e51424-feba-4038-88a6-1cd56637984c\") " pod="openstack/ovsdbserver-nb-0" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.568680 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/49e51424-feba-4038-88a6-1cd56637984c-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"49e51424-feba-4038-88a6-1cd56637984c\") " pod="openstack/ovsdbserver-nb-0" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.568696 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49e51424-feba-4038-88a6-1cd56637984c-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"49e51424-feba-4038-88a6-1cd56637984c\") " pod="openstack/ovsdbserver-nb-0" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.586216 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-nrbtc"] Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.588911 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-nrbtc" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.590875 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-5tg6b"] Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.592402 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.592579 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-t2qq4" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.592728 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.593151 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-5tg6b" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.596861 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-nrbtc"] Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.603365 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-5tg6b"] Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.671274 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"49e51424-feba-4038-88a6-1cd56637984c\") " pod="openstack/ovsdbserver-nb-0" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.671705 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/524ff107-da15-4e2f-a0c8-acca0e91b27a-scripts\") pod \"ovn-controller-nrbtc\" (UID: \"524ff107-da15-4e2f-a0c8-acca0e91b27a\") " pod="openstack/ovn-controller-nrbtc" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.671740 4890 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"49e51424-feba-4038-88a6-1cd56637984c\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/ovsdbserver-nb-0" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.671793 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/524ff107-da15-4e2f-a0c8-acca0e91b27a-ovn-controller-tls-certs\") pod \"ovn-controller-nrbtc\" (UID: \"524ff107-da15-4e2f-a0c8-acca0e91b27a\") " pod="openstack/ovn-controller-nrbtc" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.672005 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/524ff107-da15-4e2f-a0c8-acca0e91b27a-combined-ca-bundle\") pod \"ovn-controller-nrbtc\" (UID: \"524ff107-da15-4e2f-a0c8-acca0e91b27a\") " pod="openstack/ovn-controller-nrbtc" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.672136 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/524ff107-da15-4e2f-a0c8-acca0e91b27a-var-run-ovn\") pod \"ovn-controller-nrbtc\" (UID: \"524ff107-da15-4e2f-a0c8-acca0e91b27a\") " pod="openstack/ovn-controller-nrbtc" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.672237 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/49e51424-feba-4038-88a6-1cd56637984c-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"49e51424-feba-4038-88a6-1cd56637984c\") " pod="openstack/ovsdbserver-nb-0" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.672317 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/49e51424-feba-4038-88a6-1cd56637984c-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"49e51424-feba-4038-88a6-1cd56637984c\") " pod="openstack/ovsdbserver-nb-0" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.672403 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/49e51424-feba-4038-88a6-1cd56637984c-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"49e51424-feba-4038-88a6-1cd56637984c\") " pod="openstack/ovsdbserver-nb-0" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.672499 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49e51424-feba-4038-88a6-1cd56637984c-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"49e51424-feba-4038-88a6-1cd56637984c\") " pod="openstack/ovsdbserver-nb-0" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.672578 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfmhn\" (UniqueName: \"kubernetes.io/projected/524ff107-da15-4e2f-a0c8-acca0e91b27a-kube-api-access-jfmhn\") pod \"ovn-controller-nrbtc\" (UID: \"524ff107-da15-4e2f-a0c8-acca0e91b27a\") " pod="openstack/ovn-controller-nrbtc" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.672655 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/524ff107-da15-4e2f-a0c8-acca0e91b27a-var-run\") pod \"ovn-controller-nrbtc\" (UID: \"524ff107-da15-4e2f-a0c8-acca0e91b27a\") " pod="openstack/ovn-controller-nrbtc" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.672756 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/49e51424-feba-4038-88a6-1cd56637984c-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"49e51424-feba-4038-88a6-1cd56637984c\") " pod="openstack/ovsdbserver-nb-0" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.672840 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49e51424-feba-4038-88a6-1cd56637984c-config\") pod \"ovsdbserver-nb-0\" (UID: \"49e51424-feba-4038-88a6-1cd56637984c\") " pod="openstack/ovsdbserver-nb-0" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.672912 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/524ff107-da15-4e2f-a0c8-acca0e91b27a-var-log-ovn\") pod \"ovn-controller-nrbtc\" (UID: \"524ff107-da15-4e2f-a0c8-acca0e91b27a\") " pod="openstack/ovn-controller-nrbtc" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.672992 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4h2fr\" (UniqueName: \"kubernetes.io/projected/49e51424-feba-4038-88a6-1cd56637984c-kube-api-access-4h2fr\") pod \"ovsdbserver-nb-0\" (UID: \"49e51424-feba-4038-88a6-1cd56637984c\") " pod="openstack/ovsdbserver-nb-0" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.673448 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/49e51424-feba-4038-88a6-1cd56637984c-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"49e51424-feba-4038-88a6-1cd56637984c\") " pod="openstack/ovsdbserver-nb-0" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.674956 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49e51424-feba-4038-88a6-1cd56637984c-config\") pod \"ovsdbserver-nb-0\" (UID: \"49e51424-feba-4038-88a6-1cd56637984c\") " pod="openstack/ovsdbserver-nb-0" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.675678 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/49e51424-feba-4038-88a6-1cd56637984c-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"49e51424-feba-4038-88a6-1cd56637984c\") " pod="openstack/ovsdbserver-nb-0" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.685859 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/49e51424-feba-4038-88a6-1cd56637984c-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"49e51424-feba-4038-88a6-1cd56637984c\") " pod="openstack/ovsdbserver-nb-0" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.686794 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/49e51424-feba-4038-88a6-1cd56637984c-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"49e51424-feba-4038-88a6-1cd56637984c\") " pod="openstack/ovsdbserver-nb-0" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.687535 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49e51424-feba-4038-88a6-1cd56637984c-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"49e51424-feba-4038-88a6-1cd56637984c\") " pod="openstack/ovsdbserver-nb-0" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.697432 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4h2fr\" (UniqueName: \"kubernetes.io/projected/49e51424-feba-4038-88a6-1cd56637984c-kube-api-access-4h2fr\") pod \"ovsdbserver-nb-0\" (UID: \"49e51424-feba-4038-88a6-1cd56637984c\") " pod="openstack/ovsdbserver-nb-0" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.701370 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"49e51424-feba-4038-88a6-1cd56637984c\") " pod="openstack/ovsdbserver-nb-0" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.754129 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.774098 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f649q\" (UniqueName: \"kubernetes.io/projected/289df7bb-0b8e-4411-ab2f-935629ed4019-kube-api-access-f649q\") pod \"ovn-controller-ovs-5tg6b\" (UID: \"289df7bb-0b8e-4411-ab2f-935629ed4019\") " pod="openstack/ovn-controller-ovs-5tg6b" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.774156 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfmhn\" (UniqueName: \"kubernetes.io/projected/524ff107-da15-4e2f-a0c8-acca0e91b27a-kube-api-access-jfmhn\") pod \"ovn-controller-nrbtc\" (UID: \"524ff107-da15-4e2f-a0c8-acca0e91b27a\") " pod="openstack/ovn-controller-nrbtc" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.774219 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/289df7bb-0b8e-4411-ab2f-935629ed4019-var-log\") pod \"ovn-controller-ovs-5tg6b\" (UID: \"289df7bb-0b8e-4411-ab2f-935629ed4019\") " pod="openstack/ovn-controller-ovs-5tg6b" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.774239 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/524ff107-da15-4e2f-a0c8-acca0e91b27a-var-run\") pod \"ovn-controller-nrbtc\" (UID: \"524ff107-da15-4e2f-a0c8-acca0e91b27a\") " pod="openstack/ovn-controller-nrbtc" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.774370 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/524ff107-da15-4e2f-a0c8-acca0e91b27a-var-log-ovn\") pod \"ovn-controller-nrbtc\" (UID: \"524ff107-da15-4e2f-a0c8-acca0e91b27a\") " pod="openstack/ovn-controller-nrbtc" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.774823 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/524ff107-da15-4e2f-a0c8-acca0e91b27a-var-run\") pod \"ovn-controller-nrbtc\" (UID: \"524ff107-da15-4e2f-a0c8-acca0e91b27a\") " pod="openstack/ovn-controller-nrbtc" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.774891 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/524ff107-da15-4e2f-a0c8-acca0e91b27a-var-log-ovn\") pod \"ovn-controller-nrbtc\" (UID: \"524ff107-da15-4e2f-a0c8-acca0e91b27a\") " pod="openstack/ovn-controller-nrbtc" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.774933 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/289df7bb-0b8e-4411-ab2f-935629ed4019-var-run\") pod \"ovn-controller-ovs-5tg6b\" (UID: \"289df7bb-0b8e-4411-ab2f-935629ed4019\") " pod="openstack/ovn-controller-ovs-5tg6b" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.774964 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/289df7bb-0b8e-4411-ab2f-935629ed4019-var-lib\") pod \"ovn-controller-ovs-5tg6b\" (UID: \"289df7bb-0b8e-4411-ab2f-935629ed4019\") " pod="openstack/ovn-controller-ovs-5tg6b" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.774997 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/289df7bb-0b8e-4411-ab2f-935629ed4019-etc-ovs\") pod \"ovn-controller-ovs-5tg6b\" (UID: \"289df7bb-0b8e-4411-ab2f-935629ed4019\") " pod="openstack/ovn-controller-ovs-5tg6b" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.775051 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/524ff107-da15-4e2f-a0c8-acca0e91b27a-scripts\") pod \"ovn-controller-nrbtc\" (UID: \"524ff107-da15-4e2f-a0c8-acca0e91b27a\") " pod="openstack/ovn-controller-nrbtc" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.775090 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/524ff107-da15-4e2f-a0c8-acca0e91b27a-ovn-controller-tls-certs\") pod \"ovn-controller-nrbtc\" (UID: \"524ff107-da15-4e2f-a0c8-acca0e91b27a\") " pod="openstack/ovn-controller-nrbtc" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.775110 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/524ff107-da15-4e2f-a0c8-acca0e91b27a-combined-ca-bundle\") pod \"ovn-controller-nrbtc\" (UID: \"524ff107-da15-4e2f-a0c8-acca0e91b27a\") " pod="openstack/ovn-controller-nrbtc" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.775239 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/524ff107-da15-4e2f-a0c8-acca0e91b27a-var-run-ovn\") pod \"ovn-controller-nrbtc\" (UID: \"524ff107-da15-4e2f-a0c8-acca0e91b27a\") " pod="openstack/ovn-controller-nrbtc" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.775574 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/289df7bb-0b8e-4411-ab2f-935629ed4019-scripts\") pod \"ovn-controller-ovs-5tg6b\" (UID: \"289df7bb-0b8e-4411-ab2f-935629ed4019\") " pod="openstack/ovn-controller-ovs-5tg6b" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.775705 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/524ff107-da15-4e2f-a0c8-acca0e91b27a-var-run-ovn\") pod \"ovn-controller-nrbtc\" (UID: \"524ff107-da15-4e2f-a0c8-acca0e91b27a\") " pod="openstack/ovn-controller-nrbtc" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.776848 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/524ff107-da15-4e2f-a0c8-acca0e91b27a-scripts\") pod \"ovn-controller-nrbtc\" (UID: \"524ff107-da15-4e2f-a0c8-acca0e91b27a\") " pod="openstack/ovn-controller-nrbtc" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.778544 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/524ff107-da15-4e2f-a0c8-acca0e91b27a-combined-ca-bundle\") pod \"ovn-controller-nrbtc\" (UID: \"524ff107-da15-4e2f-a0c8-acca0e91b27a\") " pod="openstack/ovn-controller-nrbtc" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.778545 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/524ff107-da15-4e2f-a0c8-acca0e91b27a-ovn-controller-tls-certs\") pod \"ovn-controller-nrbtc\" (UID: \"524ff107-da15-4e2f-a0c8-acca0e91b27a\") " pod="openstack/ovn-controller-nrbtc" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.793101 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfmhn\" (UniqueName: \"kubernetes.io/projected/524ff107-da15-4e2f-a0c8-acca0e91b27a-kube-api-access-jfmhn\") pod \"ovn-controller-nrbtc\" (UID: \"524ff107-da15-4e2f-a0c8-acca0e91b27a\") " pod="openstack/ovn-controller-nrbtc" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.876658 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/289df7bb-0b8e-4411-ab2f-935629ed4019-var-run\") pod \"ovn-controller-ovs-5tg6b\" (UID: \"289df7bb-0b8e-4411-ab2f-935629ed4019\") " pod="openstack/ovn-controller-ovs-5tg6b" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.876720 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/289df7bb-0b8e-4411-ab2f-935629ed4019-var-lib\") pod \"ovn-controller-ovs-5tg6b\" (UID: \"289df7bb-0b8e-4411-ab2f-935629ed4019\") " pod="openstack/ovn-controller-ovs-5tg6b" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.876760 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/289df7bb-0b8e-4411-ab2f-935629ed4019-etc-ovs\") pod \"ovn-controller-ovs-5tg6b\" (UID: \"289df7bb-0b8e-4411-ab2f-935629ed4019\") " pod="openstack/ovn-controller-ovs-5tg6b" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.876813 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/289df7bb-0b8e-4411-ab2f-935629ed4019-scripts\") pod \"ovn-controller-ovs-5tg6b\" (UID: \"289df7bb-0b8e-4411-ab2f-935629ed4019\") " pod="openstack/ovn-controller-ovs-5tg6b" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.876844 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f649q\" (UniqueName: \"kubernetes.io/projected/289df7bb-0b8e-4411-ab2f-935629ed4019-kube-api-access-f649q\") pod \"ovn-controller-ovs-5tg6b\" (UID: \"289df7bb-0b8e-4411-ab2f-935629ed4019\") " pod="openstack/ovn-controller-ovs-5tg6b" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.876883 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/289df7bb-0b8e-4411-ab2f-935629ed4019-var-log\") pod \"ovn-controller-ovs-5tg6b\" (UID: \"289df7bb-0b8e-4411-ab2f-935629ed4019\") " pod="openstack/ovn-controller-ovs-5tg6b" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.876961 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/289df7bb-0b8e-4411-ab2f-935629ed4019-var-run\") pod \"ovn-controller-ovs-5tg6b\" (UID: \"289df7bb-0b8e-4411-ab2f-935629ed4019\") " pod="openstack/ovn-controller-ovs-5tg6b" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.877194 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/289df7bb-0b8e-4411-ab2f-935629ed4019-etc-ovs\") pod \"ovn-controller-ovs-5tg6b\" (UID: \"289df7bb-0b8e-4411-ab2f-935629ed4019\") " pod="openstack/ovn-controller-ovs-5tg6b" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.877199 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/289df7bb-0b8e-4411-ab2f-935629ed4019-var-log\") pod \"ovn-controller-ovs-5tg6b\" (UID: \"289df7bb-0b8e-4411-ab2f-935629ed4019\") " pod="openstack/ovn-controller-ovs-5tg6b" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.877257 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/289df7bb-0b8e-4411-ab2f-935629ed4019-var-lib\") pod \"ovn-controller-ovs-5tg6b\" (UID: \"289df7bb-0b8e-4411-ab2f-935629ed4019\") " pod="openstack/ovn-controller-ovs-5tg6b" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.880169 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/289df7bb-0b8e-4411-ab2f-935629ed4019-scripts\") pod \"ovn-controller-ovs-5tg6b\" (UID: \"289df7bb-0b8e-4411-ab2f-935629ed4019\") " pod="openstack/ovn-controller-ovs-5tg6b" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.895616 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f649q\" (UniqueName: \"kubernetes.io/projected/289df7bb-0b8e-4411-ab2f-935629ed4019-kube-api-access-f649q\") pod \"ovn-controller-ovs-5tg6b\" (UID: \"289df7bb-0b8e-4411-ab2f-935629ed4019\") " pod="openstack/ovn-controller-ovs-5tg6b" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.911476 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-nrbtc" Nov 25 15:17:54 crc kubenswrapper[4890]: I1125 15:17:54.918912 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-5tg6b" Nov 25 15:17:56 crc kubenswrapper[4890]: I1125 15:17:56.448133 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 15:17:56 crc kubenswrapper[4890]: I1125 15:17:56.448750 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 15:17:56 crc kubenswrapper[4890]: I1125 15:17:56.448800 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" Nov 25 15:17:56 crc kubenswrapper[4890]: I1125 15:17:56.451336 4890 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b51c3e1b3270ec2e9b5e884e9f00a4274f242cf4d6c3240274a36d149cb2249e"} pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 15:17:56 crc kubenswrapper[4890]: I1125 15:17:56.451411 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" containerID="cri-o://b51c3e1b3270ec2e9b5e884e9f00a4274f242cf4d6c3240274a36d149cb2249e" gracePeriod=600 Nov 25 15:17:57 crc kubenswrapper[4890]: I1125 15:17:57.966231 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 25 15:17:57 crc kubenswrapper[4890]: I1125 15:17:57.968068 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 25 15:17:57 crc kubenswrapper[4890]: I1125 15:17:57.970243 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Nov 25 15:17:57 crc kubenswrapper[4890]: I1125 15:17:57.971153 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Nov 25 15:17:57 crc kubenswrapper[4890]: I1125 15:17:57.971578 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-798gb" Nov 25 15:17:57 crc kubenswrapper[4890]: I1125 15:17:57.972075 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Nov 25 15:17:57 crc kubenswrapper[4890]: I1125 15:17:57.985897 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 25 15:17:58 crc kubenswrapper[4890]: I1125 15:17:58.138584 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9667a27-2ed9-4b9a-a75a-4901a93cab5e-config\") pod \"ovsdbserver-sb-0\" (UID: \"c9667a27-2ed9-4b9a-a75a-4901a93cab5e\") " pod="openstack/ovsdbserver-sb-0" Nov 25 15:17:58 crc kubenswrapper[4890]: I1125 15:17:58.138731 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvvrk\" (UniqueName: \"kubernetes.io/projected/c9667a27-2ed9-4b9a-a75a-4901a93cab5e-kube-api-access-vvvrk\") pod \"ovsdbserver-sb-0\" (UID: \"c9667a27-2ed9-4b9a-a75a-4901a93cab5e\") " pod="openstack/ovsdbserver-sb-0" Nov 25 15:17:58 crc kubenswrapper[4890]: I1125 15:17:58.138810 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-sb-0\" (UID: \"c9667a27-2ed9-4b9a-a75a-4901a93cab5e\") " pod="openstack/ovsdbserver-sb-0" Nov 25 15:17:58 crc kubenswrapper[4890]: I1125 15:17:58.138846 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c9667a27-2ed9-4b9a-a75a-4901a93cab5e-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"c9667a27-2ed9-4b9a-a75a-4901a93cab5e\") " pod="openstack/ovsdbserver-sb-0" Nov 25 15:17:58 crc kubenswrapper[4890]: I1125 15:17:58.138895 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9667a27-2ed9-4b9a-a75a-4901a93cab5e-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"c9667a27-2ed9-4b9a-a75a-4901a93cab5e\") " pod="openstack/ovsdbserver-sb-0" Nov 25 15:17:58 crc kubenswrapper[4890]: I1125 15:17:58.138922 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9667a27-2ed9-4b9a-a75a-4901a93cab5e-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"c9667a27-2ed9-4b9a-a75a-4901a93cab5e\") " pod="openstack/ovsdbserver-sb-0" Nov 25 15:17:58 crc kubenswrapper[4890]: I1125 15:17:58.139089 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c9667a27-2ed9-4b9a-a75a-4901a93cab5e-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"c9667a27-2ed9-4b9a-a75a-4901a93cab5e\") " pod="openstack/ovsdbserver-sb-0" Nov 25 15:17:58 crc kubenswrapper[4890]: I1125 15:17:58.139248 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9667a27-2ed9-4b9a-a75a-4901a93cab5e-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"c9667a27-2ed9-4b9a-a75a-4901a93cab5e\") " pod="openstack/ovsdbserver-sb-0" Nov 25 15:17:58 crc kubenswrapper[4890]: I1125 15:17:58.240855 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9667a27-2ed9-4b9a-a75a-4901a93cab5e-config\") pod \"ovsdbserver-sb-0\" (UID: \"c9667a27-2ed9-4b9a-a75a-4901a93cab5e\") " pod="openstack/ovsdbserver-sb-0" Nov 25 15:17:58 crc kubenswrapper[4890]: I1125 15:17:58.241019 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvvrk\" (UniqueName: \"kubernetes.io/projected/c9667a27-2ed9-4b9a-a75a-4901a93cab5e-kube-api-access-vvvrk\") pod \"ovsdbserver-sb-0\" (UID: \"c9667a27-2ed9-4b9a-a75a-4901a93cab5e\") " pod="openstack/ovsdbserver-sb-0" Nov 25 15:17:58 crc kubenswrapper[4890]: I1125 15:17:58.241053 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-sb-0\" (UID: \"c9667a27-2ed9-4b9a-a75a-4901a93cab5e\") " pod="openstack/ovsdbserver-sb-0" Nov 25 15:17:58 crc kubenswrapper[4890]: I1125 15:17:58.241711 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c9667a27-2ed9-4b9a-a75a-4901a93cab5e-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"c9667a27-2ed9-4b9a-a75a-4901a93cab5e\") " pod="openstack/ovsdbserver-sb-0" Nov 25 15:17:58 crc kubenswrapper[4890]: I1125 15:17:58.241844 4890 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-sb-0\" (UID: \"c9667a27-2ed9-4b9a-a75a-4901a93cab5e\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/ovsdbserver-sb-0" Nov 25 15:17:58 crc kubenswrapper[4890]: I1125 15:17:58.242197 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9667a27-2ed9-4b9a-a75a-4901a93cab5e-config\") pod \"ovsdbserver-sb-0\" (UID: \"c9667a27-2ed9-4b9a-a75a-4901a93cab5e\") " pod="openstack/ovsdbserver-sb-0" Nov 25 15:17:58 crc kubenswrapper[4890]: I1125 15:17:58.242468 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c9667a27-2ed9-4b9a-a75a-4901a93cab5e-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"c9667a27-2ed9-4b9a-a75a-4901a93cab5e\") " pod="openstack/ovsdbserver-sb-0" Nov 25 15:17:58 crc kubenswrapper[4890]: I1125 15:17:58.242991 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9667a27-2ed9-4b9a-a75a-4901a93cab5e-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"c9667a27-2ed9-4b9a-a75a-4901a93cab5e\") " pod="openstack/ovsdbserver-sb-0" Nov 25 15:17:58 crc kubenswrapper[4890]: I1125 15:17:58.243051 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9667a27-2ed9-4b9a-a75a-4901a93cab5e-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"c9667a27-2ed9-4b9a-a75a-4901a93cab5e\") " pod="openstack/ovsdbserver-sb-0" Nov 25 15:17:58 crc kubenswrapper[4890]: I1125 15:17:58.243085 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c9667a27-2ed9-4b9a-a75a-4901a93cab5e-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"c9667a27-2ed9-4b9a-a75a-4901a93cab5e\") " pod="openstack/ovsdbserver-sb-0" Nov 25 15:17:58 crc kubenswrapper[4890]: I1125 15:17:58.243141 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9667a27-2ed9-4b9a-a75a-4901a93cab5e-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"c9667a27-2ed9-4b9a-a75a-4901a93cab5e\") " pod="openstack/ovsdbserver-sb-0" Nov 25 15:17:58 crc kubenswrapper[4890]: I1125 15:17:58.246250 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c9667a27-2ed9-4b9a-a75a-4901a93cab5e-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"c9667a27-2ed9-4b9a-a75a-4901a93cab5e\") " pod="openstack/ovsdbserver-sb-0" Nov 25 15:17:58 crc kubenswrapper[4890]: I1125 15:17:58.250523 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9667a27-2ed9-4b9a-a75a-4901a93cab5e-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"c9667a27-2ed9-4b9a-a75a-4901a93cab5e\") " pod="openstack/ovsdbserver-sb-0" Nov 25 15:17:58 crc kubenswrapper[4890]: I1125 15:17:58.252667 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9667a27-2ed9-4b9a-a75a-4901a93cab5e-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"c9667a27-2ed9-4b9a-a75a-4901a93cab5e\") " pod="openstack/ovsdbserver-sb-0" Nov 25 15:17:58 crc kubenswrapper[4890]: I1125 15:17:58.259074 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9667a27-2ed9-4b9a-a75a-4901a93cab5e-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"c9667a27-2ed9-4b9a-a75a-4901a93cab5e\") " pod="openstack/ovsdbserver-sb-0" Nov 25 15:17:58 crc kubenswrapper[4890]: I1125 15:17:58.272031 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvvrk\" (UniqueName: \"kubernetes.io/projected/c9667a27-2ed9-4b9a-a75a-4901a93cab5e-kube-api-access-vvvrk\") pod \"ovsdbserver-sb-0\" (UID: \"c9667a27-2ed9-4b9a-a75a-4901a93cab5e\") " pod="openstack/ovsdbserver-sb-0" Nov 25 15:17:58 crc kubenswrapper[4890]: I1125 15:17:58.274005 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-sb-0\" (UID: \"c9667a27-2ed9-4b9a-a75a-4901a93cab5e\") " pod="openstack/ovsdbserver-sb-0" Nov 25 15:17:58 crc kubenswrapper[4890]: I1125 15:17:58.287743 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 25 15:18:02 crc kubenswrapper[4890]: I1125 15:18:02.119081 4890 generic.go:334] "Generic (PLEG): container finished" podID="4e4f849d-f239-4727-a73e-18327856929a" containerID="b51c3e1b3270ec2e9b5e884e9f00a4274f242cf4d6c3240274a36d149cb2249e" exitCode=0 Nov 25 15:18:02 crc kubenswrapper[4890]: I1125 15:18:02.119143 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" event={"ID":"4e4f849d-f239-4727-a73e-18327856929a","Type":"ContainerDied","Data":"b51c3e1b3270ec2e9b5e884e9f00a4274f242cf4d6c3240274a36d149cb2249e"} Nov 25 15:18:02 crc kubenswrapper[4890]: I1125 15:18:02.119950 4890 scope.go:117] "RemoveContainer" containerID="361ef05d7f3bcf13e88a1632cd64dd7b6044add1c3fc4ac44bf729fc819bec71" Nov 25 15:18:03 crc kubenswrapper[4890]: E1125 15:18:03.562347 4890 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = reading blob sha256:b03381363fe3e1b0227bd9a8b2ddfddea6dbc49944b60d715969e94985f431da: Get \"https://quay.io/v2/podified-antelope-centos9/openstack-rabbitmq/blobs/sha256:b03381363fe3e1b0227bd9a8b2ddfddea6dbc49944b60d715969e94985f431da\": context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Nov 25 15:18:03 crc kubenswrapper[4890]: E1125 15:18:03.562955 4890 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wbr45,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(ae7f09bf-3fc2-4da7-9b41-9f4564dd7562): ErrImagePull: rpc error: code = Canceled desc = reading blob sha256:b03381363fe3e1b0227bd9a8b2ddfddea6dbc49944b60d715969e94985f431da: Get \"https://quay.io/v2/podified-antelope-centos9/openstack-rabbitmq/blobs/sha256:b03381363fe3e1b0227bd9a8b2ddfddea6dbc49944b60d715969e94985f431da\": context canceled" logger="UnhandledError" Nov 25 15:18:03 crc kubenswrapper[4890]: E1125 15:18:03.564244 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = reading blob sha256:b03381363fe3e1b0227bd9a8b2ddfddea6dbc49944b60d715969e94985f431da: Get \\\"https://quay.io/v2/podified-antelope-centos9/openstack-rabbitmq/blobs/sha256:b03381363fe3e1b0227bd9a8b2ddfddea6dbc49944b60d715969e94985f431da\\\": context canceled\"" pod="openstack/rabbitmq-server-0" podUID="ae7f09bf-3fc2-4da7-9b41-9f4564dd7562" Nov 25 15:18:04 crc kubenswrapper[4890]: I1125 15:18:04.136991 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"90881a25-bc18-4b43-a226-a44e3165cb3a","Type":"ContainerStarted","Data":"49a7abe12ca6ceb014cf36d5d3095e1520db2c8fcc61a1f084dd91c15c4e142f"} Nov 25 15:18:04 crc kubenswrapper[4890]: E1125 15:18:04.137905 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-server-0" podUID="ae7f09bf-3fc2-4da7-9b41-9f4564dd7562" Nov 25 15:18:07 crc kubenswrapper[4890]: I1125 15:18:07.013503 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-5tg6b"] Nov 25 15:18:07 crc kubenswrapper[4890]: E1125 15:18:07.478026 4890 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Nov 25 15:18:07 crc kubenswrapper[4890]: E1125 15:18:07.479313 4890 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tcxvg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(c14ce73e-1a95-4f41-b6b6-33478079806f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 15:18:07 crc kubenswrapper[4890]: E1125 15:18:07.480449 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="c14ce73e-1a95-4f41-b6b6-33478079806f" Nov 25 15:18:08 crc kubenswrapper[4890]: E1125 15:18:08.177706 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="c14ce73e-1a95-4f41-b6b6-33478079806f" Nov 25 15:18:14 crc kubenswrapper[4890]: I1125 15:18:14.213605 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-5tg6b" event={"ID":"289df7bb-0b8e-4411-ab2f-935629ed4019","Type":"ContainerStarted","Data":"e11cb0c5c20eb9ce11964763200981c22b4588cd7c1fadd30611ebd16f419c12"} Nov 25 15:18:15 crc kubenswrapper[4890]: E1125 15:18:15.259577 4890 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 25 15:18:15 crc kubenswrapper[4890]: E1125 15:18:15.259737 4890 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-t4w4l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-27vl6_openstack(f467b253-fb57-465a-8e67-5168b2dc0e55): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 15:18:15 crc kubenswrapper[4890]: E1125 15:18:15.261127 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-27vl6" podUID="f467b253-fb57-465a-8e67-5168b2dc0e55" Nov 25 15:18:16 crc kubenswrapper[4890]: E1125 15:18:16.255895 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-27vl6" podUID="f467b253-fb57-465a-8e67-5168b2dc0e55" Nov 25 15:18:16 crc kubenswrapper[4890]: E1125 15:18:16.595261 4890 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 25 15:18:16 crc kubenswrapper[4890]: E1125 15:18:16.595738 4890 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-x6qg9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-7nfql_openstack(536d18a7-2067-46d8-864c-8c84be2a87a8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 15:18:16 crc kubenswrapper[4890]: E1125 15:18:16.597101 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-7nfql" podUID="536d18a7-2067-46d8-864c-8c84be2a87a8" Nov 25 15:18:16 crc kubenswrapper[4890]: I1125 15:18:16.791388 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-nrbtc"] Nov 25 15:18:16 crc kubenswrapper[4890]: I1125 15:18:16.804654 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 25 15:18:16 crc kubenswrapper[4890]: W1125 15:18:16.813108 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9a500d8a_0929_4a91_b7bf_3f2ccb40e3ab.slice/crio-512376e21d1592193e631ad13fda6962ad0e65179058ca4e716bf23b1dc08288 WatchSource:0}: Error finding container 512376e21d1592193e631ad13fda6962ad0e65179058ca4e716bf23b1dc08288: Status 404 returned error can't find the container with id 512376e21d1592193e631ad13fda6962ad0e65179058ca4e716bf23b1dc08288 Nov 25 15:18:16 crc kubenswrapper[4890]: I1125 15:18:16.814052 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 15:18:16 crc kubenswrapper[4890]: I1125 15:18:16.819700 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 25 15:18:16 crc kubenswrapper[4890]: I1125 15:18:16.825461 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 25 15:18:16 crc kubenswrapper[4890]: W1125 15:18:16.828387 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod24000ee2_b285_4f1a_afe1_75e643c1e172.slice/crio-9a252e68275b5109f09d5c7b05b3bff493c11dbbe5426349770b48dfde18848e WatchSource:0}: Error finding container 9a252e68275b5109f09d5c7b05b3bff493c11dbbe5426349770b48dfde18848e: Status 404 returned error can't find the container with id 9a252e68275b5109f09d5c7b05b3bff493c11dbbe5426349770b48dfde18848e Nov 25 15:18:16 crc kubenswrapper[4890]: W1125 15:18:16.838150 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod49e51424_feba_4038_88a6_1cd56637984c.slice/crio-1ccc1fa0553e79818bf6f56b56e1ada9b880cd81bdb00166aa6e25818309d071 WatchSource:0}: Error finding container 1ccc1fa0553e79818bf6f56b56e1ada9b880cd81bdb00166aa6e25818309d071: Status 404 returned error can't find the container with id 1ccc1fa0553e79818bf6f56b56e1ada9b880cd81bdb00166aa6e25818309d071 Nov 25 15:18:16 crc kubenswrapper[4890]: W1125 15:18:16.838655 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc5630332_c335_4b9a_8486_e845c6baaa9c.slice/crio-85bf0b1e65a5489ce3f1deace8f6859244217982a7bbecac786c395728272a58 WatchSource:0}: Error finding container 85bf0b1e65a5489ce3f1deace8f6859244217982a7bbecac786c395728272a58: Status 404 returned error can't find the container with id 85bf0b1e65a5489ce3f1deace8f6859244217982a7bbecac786c395728272a58 Nov 25 15:18:16 crc kubenswrapper[4890]: I1125 15:18:16.895234 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 25 15:18:17 crc kubenswrapper[4890]: I1125 15:18:17.265714 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"c9667a27-2ed9-4b9a-a75a-4901a93cab5e","Type":"ContainerStarted","Data":"7eb67196dd92ea896f3a2e79a1b5e34f12ea04c8ba2e475d1f4a2139a727229d"} Nov 25 15:18:17 crc kubenswrapper[4890]: I1125 15:18:17.267232 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"24000ee2-b285-4f1a-afe1-75e643c1e172","Type":"ContainerStarted","Data":"9a252e68275b5109f09d5c7b05b3bff493c11dbbe5426349770b48dfde18848e"} Nov 25 15:18:17 crc kubenswrapper[4890]: I1125 15:18:17.268556 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"49e51424-feba-4038-88a6-1cd56637984c","Type":"ContainerStarted","Data":"1ccc1fa0553e79818bf6f56b56e1ada9b880cd81bdb00166aa6e25818309d071"} Nov 25 15:18:17 crc kubenswrapper[4890]: I1125 15:18:17.270268 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab","Type":"ContainerStarted","Data":"512376e21d1592193e631ad13fda6962ad0e65179058ca4e716bf23b1dc08288"} Nov 25 15:18:17 crc kubenswrapper[4890]: E1125 15:18:17.271291 4890 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 25 15:18:17 crc kubenswrapper[4890]: E1125 15:18:17.271495 4890 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-l26cv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-6fkdh_openstack(c2f7ef20-6da9-4d06-a733-4a529aecffb4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 15:18:17 crc kubenswrapper[4890]: I1125 15:18:17.273472 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"c5630332-c335-4b9a-8486-e845c6baaa9c","Type":"ContainerStarted","Data":"85bf0b1e65a5489ce3f1deace8f6859244217982a7bbecac786c395728272a58"} Nov 25 15:18:17 crc kubenswrapper[4890]: E1125 15:18:17.274835 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-6fkdh" podUID="c2f7ef20-6da9-4d06-a733-4a529aecffb4" Nov 25 15:18:17 crc kubenswrapper[4890]: I1125 15:18:17.283259 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" event={"ID":"4e4f849d-f239-4727-a73e-18327856929a","Type":"ContainerStarted","Data":"53708c37db2dded329d6f7de3f6f52a10c460a5fc2cf14a1b264e8539a50f494"} Nov 25 15:18:17 crc kubenswrapper[4890]: I1125 15:18:17.287812 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"90881a25-bc18-4b43-a226-a44e3165cb3a","Type":"ContainerStarted","Data":"56e24924252b3b1489495280daa9baae96278e9c3dc4a115bbfec5c75a18ae6b"} Nov 25 15:18:17 crc kubenswrapper[4890]: I1125 15:18:17.288976 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-nrbtc" event={"ID":"524ff107-da15-4e2f-a0c8-acca0e91b27a","Type":"ContainerStarted","Data":"dbcd5d588e06c89c9a91fb6c97436e9cdcacbff828e4e7453dc5915c06ac20d9"} Nov 25 15:18:17 crc kubenswrapper[4890]: E1125 15:18:17.290412 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-7nfql" podUID="536d18a7-2067-46d8-864c-8c84be2a87a8" Nov 25 15:18:17 crc kubenswrapper[4890]: E1125 15:18:17.503625 4890 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 25 15:18:17 crc kubenswrapper[4890]: E1125 15:18:17.504353 4890 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jtbps,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-7pn5f_openstack(a6c5bcbf-f90f-4d40-819d-c0bc8a5a47fa): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 15:18:17 crc kubenswrapper[4890]: E1125 15:18:17.505847 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-7pn5f" podUID="a6c5bcbf-f90f-4d40-819d-c0bc8a5a47fa" Nov 25 15:18:18 crc kubenswrapper[4890]: I1125 15:18:18.296939 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab","Type":"ContainerStarted","Data":"6086ce14b5aa9c9e444adc6228021efb6937378d91e687312558d437e10de523"} Nov 25 15:18:18 crc kubenswrapper[4890]: I1125 15:18:18.752229 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-6fkdh" Nov 25 15:18:18 crc kubenswrapper[4890]: I1125 15:18:18.850464 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l26cv\" (UniqueName: \"kubernetes.io/projected/c2f7ef20-6da9-4d06-a733-4a529aecffb4-kube-api-access-l26cv\") pod \"c2f7ef20-6da9-4d06-a733-4a529aecffb4\" (UID: \"c2f7ef20-6da9-4d06-a733-4a529aecffb4\") " Nov 25 15:18:18 crc kubenswrapper[4890]: I1125 15:18:18.850880 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c2f7ef20-6da9-4d06-a733-4a529aecffb4-dns-svc\") pod \"c2f7ef20-6da9-4d06-a733-4a529aecffb4\" (UID: \"c2f7ef20-6da9-4d06-a733-4a529aecffb4\") " Nov 25 15:18:18 crc kubenswrapper[4890]: I1125 15:18:18.851027 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2f7ef20-6da9-4d06-a733-4a529aecffb4-config\") pod \"c2f7ef20-6da9-4d06-a733-4a529aecffb4\" (UID: \"c2f7ef20-6da9-4d06-a733-4a529aecffb4\") " Nov 25 15:18:18 crc kubenswrapper[4890]: I1125 15:18:18.851687 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2f7ef20-6da9-4d06-a733-4a529aecffb4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c2f7ef20-6da9-4d06-a733-4a529aecffb4" (UID: "c2f7ef20-6da9-4d06-a733-4a529aecffb4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:18:18 crc kubenswrapper[4890]: I1125 15:18:18.851752 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2f7ef20-6da9-4d06-a733-4a529aecffb4-config" (OuterVolumeSpecName: "config") pod "c2f7ef20-6da9-4d06-a733-4a529aecffb4" (UID: "c2f7ef20-6da9-4d06-a733-4a529aecffb4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:18:18 crc kubenswrapper[4890]: I1125 15:18:18.852211 4890 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2f7ef20-6da9-4d06-a733-4a529aecffb4-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:18:18 crc kubenswrapper[4890]: I1125 15:18:18.852231 4890 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c2f7ef20-6da9-4d06-a733-4a529aecffb4-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 15:18:18 crc kubenswrapper[4890]: I1125 15:18:18.853934 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2f7ef20-6da9-4d06-a733-4a529aecffb4-kube-api-access-l26cv" (OuterVolumeSpecName: "kube-api-access-l26cv") pod "c2f7ef20-6da9-4d06-a733-4a529aecffb4" (UID: "c2f7ef20-6da9-4d06-a733-4a529aecffb4"). InnerVolumeSpecName "kube-api-access-l26cv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:18:18 crc kubenswrapper[4890]: I1125 15:18:18.953521 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l26cv\" (UniqueName: \"kubernetes.io/projected/c2f7ef20-6da9-4d06-a733-4a529aecffb4-kube-api-access-l26cv\") on node \"crc\" DevicePath \"\"" Nov 25 15:18:19 crc kubenswrapper[4890]: I1125 15:18:19.015413 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-7pn5f" Nov 25 15:18:19 crc kubenswrapper[4890]: I1125 15:18:19.157103 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jtbps\" (UniqueName: \"kubernetes.io/projected/a6c5bcbf-f90f-4d40-819d-c0bc8a5a47fa-kube-api-access-jtbps\") pod \"a6c5bcbf-f90f-4d40-819d-c0bc8a5a47fa\" (UID: \"a6c5bcbf-f90f-4d40-819d-c0bc8a5a47fa\") " Nov 25 15:18:19 crc kubenswrapper[4890]: I1125 15:18:19.157316 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6c5bcbf-f90f-4d40-819d-c0bc8a5a47fa-config\") pod \"a6c5bcbf-f90f-4d40-819d-c0bc8a5a47fa\" (UID: \"a6c5bcbf-f90f-4d40-819d-c0bc8a5a47fa\") " Nov 25 15:18:19 crc kubenswrapper[4890]: I1125 15:18:19.157702 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6c5bcbf-f90f-4d40-819d-c0bc8a5a47fa-config" (OuterVolumeSpecName: "config") pod "a6c5bcbf-f90f-4d40-819d-c0bc8a5a47fa" (UID: "a6c5bcbf-f90f-4d40-819d-c0bc8a5a47fa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:18:19 crc kubenswrapper[4890]: I1125 15:18:19.158138 4890 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6c5bcbf-f90f-4d40-819d-c0bc8a5a47fa-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:18:19 crc kubenswrapper[4890]: I1125 15:18:19.167368 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6c5bcbf-f90f-4d40-819d-c0bc8a5a47fa-kube-api-access-jtbps" (OuterVolumeSpecName: "kube-api-access-jtbps") pod "a6c5bcbf-f90f-4d40-819d-c0bc8a5a47fa" (UID: "a6c5bcbf-f90f-4d40-819d-c0bc8a5a47fa"). InnerVolumeSpecName "kube-api-access-jtbps". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:18:19 crc kubenswrapper[4890]: I1125 15:18:19.259487 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jtbps\" (UniqueName: \"kubernetes.io/projected/a6c5bcbf-f90f-4d40-819d-c0bc8a5a47fa-kube-api-access-jtbps\") on node \"crc\" DevicePath \"\"" Nov 25 15:18:19 crc kubenswrapper[4890]: I1125 15:18:19.308074 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-7pn5f" event={"ID":"a6c5bcbf-f90f-4d40-819d-c0bc8a5a47fa","Type":"ContainerDied","Data":"391ae6f417a13745e5a1e5e36836a2f2c902909d118b7a39eef9a309b7f9466a"} Nov 25 15:18:19 crc kubenswrapper[4890]: I1125 15:18:19.308151 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-7pn5f" Nov 25 15:18:19 crc kubenswrapper[4890]: I1125 15:18:19.316478 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-6fkdh" event={"ID":"c2f7ef20-6da9-4d06-a733-4a529aecffb4","Type":"ContainerDied","Data":"dfa7ec5ff51ebc8667cd597ac8e2559da9365067699eabf6c10f258657d85c42"} Nov 25 15:18:19 crc kubenswrapper[4890]: I1125 15:18:19.316550 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-6fkdh" Nov 25 15:18:19 crc kubenswrapper[4890]: I1125 15:18:19.361747 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-7pn5f"] Nov 25 15:18:19 crc kubenswrapper[4890]: I1125 15:18:19.366562 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-7pn5f"] Nov 25 15:18:19 crc kubenswrapper[4890]: I1125 15:18:19.387279 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-6fkdh"] Nov 25 15:18:19 crc kubenswrapper[4890]: I1125 15:18:19.387326 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-6fkdh"] Nov 25 15:18:20 crc kubenswrapper[4890]: I1125 15:18:20.190698 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6c5bcbf-f90f-4d40-819d-c0bc8a5a47fa" path="/var/lib/kubelet/pods/a6c5bcbf-f90f-4d40-819d-c0bc8a5a47fa/volumes" Nov 25 15:18:20 crc kubenswrapper[4890]: I1125 15:18:20.191586 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2f7ef20-6da9-4d06-a733-4a529aecffb4" path="/var/lib/kubelet/pods/c2f7ef20-6da9-4d06-a733-4a529aecffb4/volumes" Nov 25 15:18:20 crc kubenswrapper[4890]: I1125 15:18:20.325443 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562","Type":"ContainerStarted","Data":"3c3e2348928c66cad8da81580a88530958cd3237ca6204733fd5e8d2343fdb71"} Nov 25 15:18:27 crc kubenswrapper[4890]: I1125 15:18:27.378811 4890 generic.go:334] "Generic (PLEG): container finished" podID="90881a25-bc18-4b43-a226-a44e3165cb3a" containerID="56e24924252b3b1489495280daa9baae96278e9c3dc4a115bbfec5c75a18ae6b" exitCode=0 Nov 25 15:18:27 crc kubenswrapper[4890]: I1125 15:18:27.378907 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"90881a25-bc18-4b43-a226-a44e3165cb3a","Type":"ContainerDied","Data":"56e24924252b3b1489495280daa9baae96278e9c3dc4a115bbfec5c75a18ae6b"} Nov 25 15:18:27 crc kubenswrapper[4890]: I1125 15:18:27.382973 4890 generic.go:334] "Generic (PLEG): container finished" podID="9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab" containerID="6086ce14b5aa9c9e444adc6228021efb6937378d91e687312558d437e10de523" exitCode=0 Nov 25 15:18:27 crc kubenswrapper[4890]: I1125 15:18:27.383023 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab","Type":"ContainerDied","Data":"6086ce14b5aa9c9e444adc6228021efb6937378d91e687312558d437e10de523"} Nov 25 15:18:42 crc kubenswrapper[4890]: I1125 15:18:42.490476 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"90881a25-bc18-4b43-a226-a44e3165cb3a","Type":"ContainerStarted","Data":"dc1b5df96364c36f54a79ed6af15fcb270cccf41451cc8774685e340c393c768"} Nov 25 15:18:42 crc kubenswrapper[4890]: I1125 15:18:42.511688 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=43.359995761 podStartE2EDuration="56.511668049s" podCreationTimestamp="2025-11-25 15:17:46 +0000 UTC" firstStartedPulling="2025-11-25 15:18:03.233775297 +0000 UTC m=+941.676237937" lastFinishedPulling="2025-11-25 15:18:16.385447615 +0000 UTC m=+954.827910225" observedRunningTime="2025-11-25 15:18:42.507664189 +0000 UTC m=+980.950126809" watchObservedRunningTime="2025-11-25 15:18:42.511668049 +0000 UTC m=+980.954130659" Nov 25 15:18:44 crc kubenswrapper[4890]: I1125 15:18:44.510522 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab","Type":"ContainerStarted","Data":"77de70fee73f8e922081913636d36a52ebbcd52950f0148fcf9005b7b3bef880"} Nov 25 15:18:44 crc kubenswrapper[4890]: I1125 15:18:44.538659 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=57.538630459 podStartE2EDuration="57.538630459s" podCreationTimestamp="2025-11-25 15:17:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:18:44.533547612 +0000 UTC m=+982.976010232" watchObservedRunningTime="2025-11-25 15:18:44.538630459 +0000 UTC m=+982.981093109" Nov 25 15:18:47 crc kubenswrapper[4890]: I1125 15:18:47.504624 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Nov 25 15:18:47 crc kubenswrapper[4890]: I1125 15:18:47.505194 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Nov 25 15:18:48 crc kubenswrapper[4890]: I1125 15:18:48.545990 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"c9667a27-2ed9-4b9a-a75a-4901a93cab5e","Type":"ContainerStarted","Data":"be76d61c0764250fcac81ff3ec00390b8dd5b707108cdd4c5c1e08e5af528281"} Nov 25 15:18:48 crc kubenswrapper[4890]: I1125 15:18:48.548929 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"49e51424-feba-4038-88a6-1cd56637984c","Type":"ContainerStarted","Data":"e7b6ace834747de4ed36a988f952dfbda147386ae2f021e8677224fdc711d81b"} Nov 25 15:18:48 crc kubenswrapper[4890]: I1125 15:18:48.550814 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"c5630332-c335-4b9a-8486-e845c6baaa9c","Type":"ContainerStarted","Data":"3c4d34a82f04a5a77473c5badbcab7ce39db56716fd2a1bb00ba9df473114588"} Nov 25 15:18:48 crc kubenswrapper[4890]: I1125 15:18:48.553593 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-5tg6b" event={"ID":"289df7bb-0b8e-4411-ab2f-935629ed4019","Type":"ContainerStarted","Data":"bd05dc4ee7d77f59f62d548e230ec4d63b578c319419abf7e99401d650517180"} Nov 25 15:18:48 crc kubenswrapper[4890]: I1125 15:18:48.554906 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-nrbtc" event={"ID":"524ff107-da15-4e2f-a0c8-acca0e91b27a","Type":"ContainerStarted","Data":"de9eb3671b86226f60bcd4a2ab4ea16fd9254b0776e1268124d61d60eff68b2e"} Nov 25 15:18:48 crc kubenswrapper[4890]: E1125 15:18:48.761134 4890 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0" Nov 25 15:18:48 crc kubenswrapper[4890]: E1125 15:18:48.761592 4890 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0" Nov 25 15:18:48 crc kubenswrapper[4890]: E1125 15:18:48.761920 4890 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-state-metrics,Image:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,Command:[],Args:[--resources=pods --namespaces=openstack],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:http-metrics,HostPort:0,ContainerPort:8080,Protocol:TCP,HostIP:,},ContainerPort{Name:telemetry,HostPort:0,ContainerPort:8081,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6mtcq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/livez,Port:{0 8080 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod kube-state-metrics-0_openstack(24000ee2-b285-4f1a-afe1-75e643c1e172): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 15:18:48 crc kubenswrapper[4890]: E1125 15:18:48.763377 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/kube-state-metrics-0" podUID="24000ee2-b285-4f1a-afe1-75e643c1e172" Nov 25 15:18:48 crc kubenswrapper[4890]: I1125 15:18:48.847009 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Nov 25 15:18:48 crc kubenswrapper[4890]: I1125 15:18:48.847066 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Nov 25 15:18:49 crc kubenswrapper[4890]: I1125 15:18:49.565342 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c14ce73e-1a95-4f41-b6b6-33478079806f","Type":"ContainerStarted","Data":"7b0b47083ae2bc3b6521792dbfc2606a74d4a262bfbeca4040d61b4873018ae8"} Nov 25 15:18:49 crc kubenswrapper[4890]: I1125 15:18:49.565890 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Nov 25 15:18:49 crc kubenswrapper[4890]: I1125 15:18:49.565995 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-nrbtc" Nov 25 15:18:49 crc kubenswrapper[4890]: E1125 15:18:49.567924 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0\\\"\"" pod="openstack/kube-state-metrics-0" podUID="24000ee2-b285-4f1a-afe1-75e643c1e172" Nov 25 15:18:49 crc kubenswrapper[4890]: I1125 15:18:49.590300 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=41.881392381 podStartE2EDuration="1m1.590278112s" podCreationTimestamp="2025-11-25 15:17:48 +0000 UTC" firstStartedPulling="2025-11-25 15:18:16.846151979 +0000 UTC m=+955.288614589" lastFinishedPulling="2025-11-25 15:18:36.5550377 +0000 UTC m=+974.997500320" observedRunningTime="2025-11-25 15:18:49.58419071 +0000 UTC m=+988.026653330" watchObservedRunningTime="2025-11-25 15:18:49.590278112 +0000 UTC m=+988.032740722" Nov 25 15:18:49 crc kubenswrapper[4890]: I1125 15:18:49.602989 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-nrbtc" podStartSLOduration=31.898305086 podStartE2EDuration="55.602973199s" podCreationTimestamp="2025-11-25 15:17:54 +0000 UTC" firstStartedPulling="2025-11-25 15:18:16.804897178 +0000 UTC m=+955.247359788" lastFinishedPulling="2025-11-25 15:18:40.509565281 +0000 UTC m=+978.952027901" observedRunningTime="2025-11-25 15:18:49.599200485 +0000 UTC m=+988.041663105" watchObservedRunningTime="2025-11-25 15:18:49.602973199 +0000 UTC m=+988.045435809" Nov 25 15:18:52 crc kubenswrapper[4890]: I1125 15:18:52.598302 4890 generic.go:334] "Generic (PLEG): container finished" podID="289df7bb-0b8e-4411-ab2f-935629ed4019" containerID="bd05dc4ee7d77f59f62d548e230ec4d63b578c319419abf7e99401d650517180" exitCode=0 Nov 25 15:18:52 crc kubenswrapper[4890]: I1125 15:18:52.598368 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-5tg6b" event={"ID":"289df7bb-0b8e-4411-ab2f-935629ed4019","Type":"ContainerDied","Data":"bd05dc4ee7d77f59f62d548e230ec4d63b578c319419abf7e99401d650517180"} Nov 25 15:18:53 crc kubenswrapper[4890]: I1125 15:18:53.612023 4890 generic.go:334] "Generic (PLEG): container finished" podID="ae7f09bf-3fc2-4da7-9b41-9f4564dd7562" containerID="3c3e2348928c66cad8da81580a88530958cd3237ca6204733fd5e8d2343fdb71" exitCode=0 Nov 25 15:18:53 crc kubenswrapper[4890]: I1125 15:18:53.612108 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562","Type":"ContainerDied","Data":"3c3e2348928c66cad8da81580a88530958cd3237ca6204733fd5e8d2343fdb71"} Nov 25 15:18:54 crc kubenswrapper[4890]: I1125 15:18:54.043826 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Nov 25 15:18:55 crc kubenswrapper[4890]: I1125 15:18:55.634792 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562","Type":"ContainerStarted","Data":"4d7577d6f8303eed0d77102268bf3e3dd24a0be05f90274612e2dc2f3373322d"} Nov 25 15:18:55 crc kubenswrapper[4890]: I1125 15:18:55.635896 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 25 15:18:55 crc kubenswrapper[4890]: I1125 15:18:55.669318 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=38.881709112 podStartE2EDuration="1m11.669295464s" podCreationTimestamp="2025-11-25 15:17:44 +0000 UTC" firstStartedPulling="2025-11-25 15:17:45.855705653 +0000 UTC m=+924.298168263" lastFinishedPulling="2025-11-25 15:18:18.643292005 +0000 UTC m=+957.085754615" observedRunningTime="2025-11-25 15:18:55.664214817 +0000 UTC m=+994.106677447" watchObservedRunningTime="2025-11-25 15:18:55.669295464 +0000 UTC m=+994.111758074" Nov 25 15:19:01 crc kubenswrapper[4890]: I1125 15:19:01.215485 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-7nfql"] Nov 25 15:19:01 crc kubenswrapper[4890]: I1125 15:19:01.250047 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-d8lf2"] Nov 25 15:19:01 crc kubenswrapper[4890]: I1125 15:19:01.251655 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-d8lf2" Nov 25 15:19:01 crc kubenswrapper[4890]: I1125 15:19:01.256625 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a74f4c1-2950-4305-b91f-50105ed2be79-config\") pod \"dnsmasq-dns-7cb5889db5-d8lf2\" (UID: \"2a74f4c1-2950-4305-b91f-50105ed2be79\") " pod="openstack/dnsmasq-dns-7cb5889db5-d8lf2" Nov 25 15:19:01 crc kubenswrapper[4890]: I1125 15:19:01.256784 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rs2h9\" (UniqueName: \"kubernetes.io/projected/2a74f4c1-2950-4305-b91f-50105ed2be79-kube-api-access-rs2h9\") pod \"dnsmasq-dns-7cb5889db5-d8lf2\" (UID: \"2a74f4c1-2950-4305-b91f-50105ed2be79\") " pod="openstack/dnsmasq-dns-7cb5889db5-d8lf2" Nov 25 15:19:01 crc kubenswrapper[4890]: I1125 15:19:01.256826 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2a74f4c1-2950-4305-b91f-50105ed2be79-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-d8lf2\" (UID: \"2a74f4c1-2950-4305-b91f-50105ed2be79\") " pod="openstack/dnsmasq-dns-7cb5889db5-d8lf2" Nov 25 15:19:01 crc kubenswrapper[4890]: I1125 15:19:01.311854 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-d8lf2"] Nov 25 15:19:01 crc kubenswrapper[4890]: I1125 15:19:01.358320 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rs2h9\" (UniqueName: \"kubernetes.io/projected/2a74f4c1-2950-4305-b91f-50105ed2be79-kube-api-access-rs2h9\") pod \"dnsmasq-dns-7cb5889db5-d8lf2\" (UID: \"2a74f4c1-2950-4305-b91f-50105ed2be79\") " pod="openstack/dnsmasq-dns-7cb5889db5-d8lf2" Nov 25 15:19:01 crc kubenswrapper[4890]: I1125 15:19:01.358677 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2a74f4c1-2950-4305-b91f-50105ed2be79-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-d8lf2\" (UID: \"2a74f4c1-2950-4305-b91f-50105ed2be79\") " pod="openstack/dnsmasq-dns-7cb5889db5-d8lf2" Nov 25 15:19:01 crc kubenswrapper[4890]: I1125 15:19:01.358876 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a74f4c1-2950-4305-b91f-50105ed2be79-config\") pod \"dnsmasq-dns-7cb5889db5-d8lf2\" (UID: \"2a74f4c1-2950-4305-b91f-50105ed2be79\") " pod="openstack/dnsmasq-dns-7cb5889db5-d8lf2" Nov 25 15:19:01 crc kubenswrapper[4890]: I1125 15:19:01.367903 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2a74f4c1-2950-4305-b91f-50105ed2be79-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-d8lf2\" (UID: \"2a74f4c1-2950-4305-b91f-50105ed2be79\") " pod="openstack/dnsmasq-dns-7cb5889db5-d8lf2" Nov 25 15:19:01 crc kubenswrapper[4890]: I1125 15:19:01.368260 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a74f4c1-2950-4305-b91f-50105ed2be79-config\") pod \"dnsmasq-dns-7cb5889db5-d8lf2\" (UID: \"2a74f4c1-2950-4305-b91f-50105ed2be79\") " pod="openstack/dnsmasq-dns-7cb5889db5-d8lf2" Nov 25 15:19:01 crc kubenswrapper[4890]: I1125 15:19:01.378847 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rs2h9\" (UniqueName: \"kubernetes.io/projected/2a74f4c1-2950-4305-b91f-50105ed2be79-kube-api-access-rs2h9\") pod \"dnsmasq-dns-7cb5889db5-d8lf2\" (UID: \"2a74f4c1-2950-4305-b91f-50105ed2be79\") " pod="openstack/dnsmasq-dns-7cb5889db5-d8lf2" Nov 25 15:19:01 crc kubenswrapper[4890]: I1125 15:19:01.578873 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-d8lf2" Nov 25 15:19:02 crc kubenswrapper[4890]: I1125 15:19:02.407507 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Nov 25 15:19:02 crc kubenswrapper[4890]: I1125 15:19:02.413470 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 25 15:19:02 crc kubenswrapper[4890]: I1125 15:19:02.425598 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Nov 25 15:19:02 crc kubenswrapper[4890]: I1125 15:19:02.425629 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Nov 25 15:19:02 crc kubenswrapper[4890]: I1125 15:19:02.425666 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Nov 25 15:19:02 crc kubenswrapper[4890]: I1125 15:19:02.425858 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-8zbpv" Nov 25 15:19:02 crc kubenswrapper[4890]: I1125 15:19:02.430572 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 25 15:19:02 crc kubenswrapper[4890]: I1125 15:19:02.481878 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lp76\" (UniqueName: \"kubernetes.io/projected/6517688d-0723-4809-ba51-bcaf48524157-kube-api-access-8lp76\") pod \"swift-storage-0\" (UID: \"6517688d-0723-4809-ba51-bcaf48524157\") " pod="openstack/swift-storage-0" Nov 25 15:19:02 crc kubenswrapper[4890]: I1125 15:19:02.481920 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/6517688d-0723-4809-ba51-bcaf48524157-lock\") pod \"swift-storage-0\" (UID: \"6517688d-0723-4809-ba51-bcaf48524157\") " pod="openstack/swift-storage-0" Nov 25 15:19:02 crc kubenswrapper[4890]: I1125 15:19:02.481945 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"swift-storage-0\" (UID: \"6517688d-0723-4809-ba51-bcaf48524157\") " pod="openstack/swift-storage-0" Nov 25 15:19:02 crc kubenswrapper[4890]: I1125 15:19:02.482055 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6517688d-0723-4809-ba51-bcaf48524157-etc-swift\") pod \"swift-storage-0\" (UID: \"6517688d-0723-4809-ba51-bcaf48524157\") " pod="openstack/swift-storage-0" Nov 25 15:19:02 crc kubenswrapper[4890]: I1125 15:19:02.482074 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/6517688d-0723-4809-ba51-bcaf48524157-cache\") pod \"swift-storage-0\" (UID: \"6517688d-0723-4809-ba51-bcaf48524157\") " pod="openstack/swift-storage-0" Nov 25 15:19:02 crc kubenswrapper[4890]: I1125 15:19:02.583443 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lp76\" (UniqueName: \"kubernetes.io/projected/6517688d-0723-4809-ba51-bcaf48524157-kube-api-access-8lp76\") pod \"swift-storage-0\" (UID: \"6517688d-0723-4809-ba51-bcaf48524157\") " pod="openstack/swift-storage-0" Nov 25 15:19:02 crc kubenswrapper[4890]: I1125 15:19:02.583501 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/6517688d-0723-4809-ba51-bcaf48524157-lock\") pod \"swift-storage-0\" (UID: \"6517688d-0723-4809-ba51-bcaf48524157\") " pod="openstack/swift-storage-0" Nov 25 15:19:02 crc kubenswrapper[4890]: I1125 15:19:02.583531 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"swift-storage-0\" (UID: \"6517688d-0723-4809-ba51-bcaf48524157\") " pod="openstack/swift-storage-0" Nov 25 15:19:02 crc kubenswrapper[4890]: I1125 15:19:02.583579 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6517688d-0723-4809-ba51-bcaf48524157-etc-swift\") pod \"swift-storage-0\" (UID: \"6517688d-0723-4809-ba51-bcaf48524157\") " pod="openstack/swift-storage-0" Nov 25 15:19:02 crc kubenswrapper[4890]: I1125 15:19:02.583596 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/6517688d-0723-4809-ba51-bcaf48524157-cache\") pod \"swift-storage-0\" (UID: \"6517688d-0723-4809-ba51-bcaf48524157\") " pod="openstack/swift-storage-0" Nov 25 15:19:02 crc kubenswrapper[4890]: I1125 15:19:02.583950 4890 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"swift-storage-0\" (UID: \"6517688d-0723-4809-ba51-bcaf48524157\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/swift-storage-0" Nov 25 15:19:02 crc kubenswrapper[4890]: I1125 15:19:02.584290 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/6517688d-0723-4809-ba51-bcaf48524157-cache\") pod \"swift-storage-0\" (UID: \"6517688d-0723-4809-ba51-bcaf48524157\") " pod="openstack/swift-storage-0" Nov 25 15:19:02 crc kubenswrapper[4890]: I1125 15:19:02.584354 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/6517688d-0723-4809-ba51-bcaf48524157-lock\") pod \"swift-storage-0\" (UID: \"6517688d-0723-4809-ba51-bcaf48524157\") " pod="openstack/swift-storage-0" Nov 25 15:19:02 crc kubenswrapper[4890]: E1125 15:19:02.584475 4890 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 25 15:19:02 crc kubenswrapper[4890]: E1125 15:19:02.584497 4890 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 25 15:19:02 crc kubenswrapper[4890]: E1125 15:19:02.584550 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/6517688d-0723-4809-ba51-bcaf48524157-etc-swift podName:6517688d-0723-4809-ba51-bcaf48524157 nodeName:}" failed. No retries permitted until 2025-11-25 15:19:03.084529565 +0000 UTC m=+1001.526992175 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/6517688d-0723-4809-ba51-bcaf48524157-etc-swift") pod "swift-storage-0" (UID: "6517688d-0723-4809-ba51-bcaf48524157") : configmap "swift-ring-files" not found Nov 25 15:19:02 crc kubenswrapper[4890]: I1125 15:19:02.607392 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lp76\" (UniqueName: \"kubernetes.io/projected/6517688d-0723-4809-ba51-bcaf48524157-kube-api-access-8lp76\") pod \"swift-storage-0\" (UID: \"6517688d-0723-4809-ba51-bcaf48524157\") " pod="openstack/swift-storage-0" Nov 25 15:19:02 crc kubenswrapper[4890]: I1125 15:19:02.631572 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"swift-storage-0\" (UID: \"6517688d-0723-4809-ba51-bcaf48524157\") " pod="openstack/swift-storage-0" Nov 25 15:19:03 crc kubenswrapper[4890]: I1125 15:19:03.045199 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-vs9dw"] Nov 25 15:19:03 crc kubenswrapper[4890]: I1125 15:19:03.046488 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-vs9dw" Nov 25 15:19:03 crc kubenswrapper[4890]: I1125 15:19:03.048120 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 25 15:19:03 crc kubenswrapper[4890]: I1125 15:19:03.048376 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Nov 25 15:19:03 crc kubenswrapper[4890]: I1125 15:19:03.051040 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Nov 25 15:19:03 crc kubenswrapper[4890]: I1125 15:19:03.057200 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-vs9dw"] Nov 25 15:19:03 crc kubenswrapper[4890]: I1125 15:19:03.091930 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/cf34ad11-7dd6-4572-b3d2-4c14127a1d76-dispersionconf\") pod \"swift-ring-rebalance-vs9dw\" (UID: \"cf34ad11-7dd6-4572-b3d2-4c14127a1d76\") " pod="openstack/swift-ring-rebalance-vs9dw" Nov 25 15:19:03 crc kubenswrapper[4890]: I1125 15:19:03.092003 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/cf34ad11-7dd6-4572-b3d2-4c14127a1d76-ring-data-devices\") pod \"swift-ring-rebalance-vs9dw\" (UID: \"cf34ad11-7dd6-4572-b3d2-4c14127a1d76\") " pod="openstack/swift-ring-rebalance-vs9dw" Nov 25 15:19:03 crc kubenswrapper[4890]: I1125 15:19:03.092123 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6517688d-0723-4809-ba51-bcaf48524157-etc-swift\") pod \"swift-storage-0\" (UID: \"6517688d-0723-4809-ba51-bcaf48524157\") " pod="openstack/swift-storage-0" Nov 25 15:19:03 crc kubenswrapper[4890]: I1125 15:19:03.092222 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnhvr\" (UniqueName: \"kubernetes.io/projected/cf34ad11-7dd6-4572-b3d2-4c14127a1d76-kube-api-access-nnhvr\") pod \"swift-ring-rebalance-vs9dw\" (UID: \"cf34ad11-7dd6-4572-b3d2-4c14127a1d76\") " pod="openstack/swift-ring-rebalance-vs9dw" Nov 25 15:19:03 crc kubenswrapper[4890]: I1125 15:19:03.092276 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cf34ad11-7dd6-4572-b3d2-4c14127a1d76-scripts\") pod \"swift-ring-rebalance-vs9dw\" (UID: \"cf34ad11-7dd6-4572-b3d2-4c14127a1d76\") " pod="openstack/swift-ring-rebalance-vs9dw" Nov 25 15:19:03 crc kubenswrapper[4890]: I1125 15:19:03.092305 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/cf34ad11-7dd6-4572-b3d2-4c14127a1d76-swiftconf\") pod \"swift-ring-rebalance-vs9dw\" (UID: \"cf34ad11-7dd6-4572-b3d2-4c14127a1d76\") " pod="openstack/swift-ring-rebalance-vs9dw" Nov 25 15:19:03 crc kubenswrapper[4890]: I1125 15:19:03.092357 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/cf34ad11-7dd6-4572-b3d2-4c14127a1d76-etc-swift\") pod \"swift-ring-rebalance-vs9dw\" (UID: \"cf34ad11-7dd6-4572-b3d2-4c14127a1d76\") " pod="openstack/swift-ring-rebalance-vs9dw" Nov 25 15:19:03 crc kubenswrapper[4890]: I1125 15:19:03.092459 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf34ad11-7dd6-4572-b3d2-4c14127a1d76-combined-ca-bundle\") pod \"swift-ring-rebalance-vs9dw\" (UID: \"cf34ad11-7dd6-4572-b3d2-4c14127a1d76\") " pod="openstack/swift-ring-rebalance-vs9dw" Nov 25 15:19:03 crc kubenswrapper[4890]: E1125 15:19:03.092625 4890 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 25 15:19:03 crc kubenswrapper[4890]: E1125 15:19:03.092644 4890 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 25 15:19:03 crc kubenswrapper[4890]: E1125 15:19:03.092682 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/6517688d-0723-4809-ba51-bcaf48524157-etc-swift podName:6517688d-0723-4809-ba51-bcaf48524157 nodeName:}" failed. No retries permitted until 2025-11-25 15:19:04.092668575 +0000 UTC m=+1002.535131185 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/6517688d-0723-4809-ba51-bcaf48524157-etc-swift") pod "swift-storage-0" (UID: "6517688d-0723-4809-ba51-bcaf48524157") : configmap "swift-ring-files" not found Nov 25 15:19:03 crc kubenswrapper[4890]: I1125 15:19:03.193390 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnhvr\" (UniqueName: \"kubernetes.io/projected/cf34ad11-7dd6-4572-b3d2-4c14127a1d76-kube-api-access-nnhvr\") pod \"swift-ring-rebalance-vs9dw\" (UID: \"cf34ad11-7dd6-4572-b3d2-4c14127a1d76\") " pod="openstack/swift-ring-rebalance-vs9dw" Nov 25 15:19:03 crc kubenswrapper[4890]: I1125 15:19:03.193446 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cf34ad11-7dd6-4572-b3d2-4c14127a1d76-scripts\") pod \"swift-ring-rebalance-vs9dw\" (UID: \"cf34ad11-7dd6-4572-b3d2-4c14127a1d76\") " pod="openstack/swift-ring-rebalance-vs9dw" Nov 25 15:19:03 crc kubenswrapper[4890]: I1125 15:19:03.193467 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/cf34ad11-7dd6-4572-b3d2-4c14127a1d76-swiftconf\") pod \"swift-ring-rebalance-vs9dw\" (UID: \"cf34ad11-7dd6-4572-b3d2-4c14127a1d76\") " pod="openstack/swift-ring-rebalance-vs9dw" Nov 25 15:19:03 crc kubenswrapper[4890]: I1125 15:19:03.193491 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/cf34ad11-7dd6-4572-b3d2-4c14127a1d76-etc-swift\") pod \"swift-ring-rebalance-vs9dw\" (UID: \"cf34ad11-7dd6-4572-b3d2-4c14127a1d76\") " pod="openstack/swift-ring-rebalance-vs9dw" Nov 25 15:19:03 crc kubenswrapper[4890]: I1125 15:19:03.193533 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf34ad11-7dd6-4572-b3d2-4c14127a1d76-combined-ca-bundle\") pod \"swift-ring-rebalance-vs9dw\" (UID: \"cf34ad11-7dd6-4572-b3d2-4c14127a1d76\") " pod="openstack/swift-ring-rebalance-vs9dw" Nov 25 15:19:03 crc kubenswrapper[4890]: I1125 15:19:03.193561 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/cf34ad11-7dd6-4572-b3d2-4c14127a1d76-dispersionconf\") pod \"swift-ring-rebalance-vs9dw\" (UID: \"cf34ad11-7dd6-4572-b3d2-4c14127a1d76\") " pod="openstack/swift-ring-rebalance-vs9dw" Nov 25 15:19:03 crc kubenswrapper[4890]: I1125 15:19:03.193626 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/cf34ad11-7dd6-4572-b3d2-4c14127a1d76-ring-data-devices\") pod \"swift-ring-rebalance-vs9dw\" (UID: \"cf34ad11-7dd6-4572-b3d2-4c14127a1d76\") " pod="openstack/swift-ring-rebalance-vs9dw" Nov 25 15:19:03 crc kubenswrapper[4890]: I1125 15:19:03.193945 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/cf34ad11-7dd6-4572-b3d2-4c14127a1d76-etc-swift\") pod \"swift-ring-rebalance-vs9dw\" (UID: \"cf34ad11-7dd6-4572-b3d2-4c14127a1d76\") " pod="openstack/swift-ring-rebalance-vs9dw" Nov 25 15:19:03 crc kubenswrapper[4890]: I1125 15:19:03.194392 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/cf34ad11-7dd6-4572-b3d2-4c14127a1d76-ring-data-devices\") pod \"swift-ring-rebalance-vs9dw\" (UID: \"cf34ad11-7dd6-4572-b3d2-4c14127a1d76\") " pod="openstack/swift-ring-rebalance-vs9dw" Nov 25 15:19:03 crc kubenswrapper[4890]: I1125 15:19:03.194630 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cf34ad11-7dd6-4572-b3d2-4c14127a1d76-scripts\") pod \"swift-ring-rebalance-vs9dw\" (UID: \"cf34ad11-7dd6-4572-b3d2-4c14127a1d76\") " pod="openstack/swift-ring-rebalance-vs9dw" Nov 25 15:19:03 crc kubenswrapper[4890]: I1125 15:19:03.197268 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/cf34ad11-7dd6-4572-b3d2-4c14127a1d76-swiftconf\") pod \"swift-ring-rebalance-vs9dw\" (UID: \"cf34ad11-7dd6-4572-b3d2-4c14127a1d76\") " pod="openstack/swift-ring-rebalance-vs9dw" Nov 25 15:19:03 crc kubenswrapper[4890]: I1125 15:19:03.201842 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf34ad11-7dd6-4572-b3d2-4c14127a1d76-combined-ca-bundle\") pod \"swift-ring-rebalance-vs9dw\" (UID: \"cf34ad11-7dd6-4572-b3d2-4c14127a1d76\") " pod="openstack/swift-ring-rebalance-vs9dw" Nov 25 15:19:03 crc kubenswrapper[4890]: I1125 15:19:03.209391 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/cf34ad11-7dd6-4572-b3d2-4c14127a1d76-dispersionconf\") pod \"swift-ring-rebalance-vs9dw\" (UID: \"cf34ad11-7dd6-4572-b3d2-4c14127a1d76\") " pod="openstack/swift-ring-rebalance-vs9dw" Nov 25 15:19:03 crc kubenswrapper[4890]: I1125 15:19:03.210075 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnhvr\" (UniqueName: \"kubernetes.io/projected/cf34ad11-7dd6-4572-b3d2-4c14127a1d76-kube-api-access-nnhvr\") pod \"swift-ring-rebalance-vs9dw\" (UID: \"cf34ad11-7dd6-4572-b3d2-4c14127a1d76\") " pod="openstack/swift-ring-rebalance-vs9dw" Nov 25 15:19:03 crc kubenswrapper[4890]: I1125 15:19:03.425796 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-vs9dw" Nov 25 15:19:04 crc kubenswrapper[4890]: I1125 15:19:04.106441 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6517688d-0723-4809-ba51-bcaf48524157-etc-swift\") pod \"swift-storage-0\" (UID: \"6517688d-0723-4809-ba51-bcaf48524157\") " pod="openstack/swift-storage-0" Nov 25 15:19:04 crc kubenswrapper[4890]: E1125 15:19:04.106581 4890 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 25 15:19:04 crc kubenswrapper[4890]: E1125 15:19:04.106880 4890 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 25 15:19:04 crc kubenswrapper[4890]: E1125 15:19:04.106949 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/6517688d-0723-4809-ba51-bcaf48524157-etc-swift podName:6517688d-0723-4809-ba51-bcaf48524157 nodeName:}" failed. No retries permitted until 2025-11-25 15:19:06.106927565 +0000 UTC m=+1004.549390175 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/6517688d-0723-4809-ba51-bcaf48524157-etc-swift") pod "swift-storage-0" (UID: "6517688d-0723-4809-ba51-bcaf48524157") : configmap "swift-ring-files" not found Nov 25 15:19:05 crc kubenswrapper[4890]: I1125 15:19:05.470670 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Nov 25 15:19:05 crc kubenswrapper[4890]: I1125 15:19:05.494209 4890 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="ae7f09bf-3fc2-4da7-9b41-9f4564dd7562" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.101:5671: connect: connection refused" Nov 25 15:19:05 crc kubenswrapper[4890]: I1125 15:19:05.624438 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Nov 25 15:19:06 crc kubenswrapper[4890]: I1125 15:19:06.016400 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-d8lf2"] Nov 25 15:19:06 crc kubenswrapper[4890]: W1125 15:19:06.020730 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2a74f4c1_2950_4305_b91f_50105ed2be79.slice/crio-62c131099a60fd2a495b6556c9fc55f1320ef57b762369e3c67267d65109e896 WatchSource:0}: Error finding container 62c131099a60fd2a495b6556c9fc55f1320ef57b762369e3c67267d65109e896: Status 404 returned error can't find the container with id 62c131099a60fd2a495b6556c9fc55f1320ef57b762369e3c67267d65109e896 Nov 25 15:19:06 crc kubenswrapper[4890]: I1125 15:19:06.109222 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-vs9dw"] Nov 25 15:19:06 crc kubenswrapper[4890]: W1125 15:19:06.123126 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcf34ad11_7dd6_4572_b3d2_4c14127a1d76.slice/crio-ee321cbe36fab6fbdc08de675467b8b88bd16887233dd57e6eb443ed15a9617b WatchSource:0}: Error finding container ee321cbe36fab6fbdc08de675467b8b88bd16887233dd57e6eb443ed15a9617b: Status 404 returned error can't find the container with id ee321cbe36fab6fbdc08de675467b8b88bd16887233dd57e6eb443ed15a9617b Nov 25 15:19:06 crc kubenswrapper[4890]: I1125 15:19:06.147447 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6517688d-0723-4809-ba51-bcaf48524157-etc-swift\") pod \"swift-storage-0\" (UID: \"6517688d-0723-4809-ba51-bcaf48524157\") " pod="openstack/swift-storage-0" Nov 25 15:19:06 crc kubenswrapper[4890]: E1125 15:19:06.147644 4890 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 25 15:19:06 crc kubenswrapper[4890]: E1125 15:19:06.147675 4890 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 25 15:19:06 crc kubenswrapper[4890]: E1125 15:19:06.147770 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/6517688d-0723-4809-ba51-bcaf48524157-etc-swift podName:6517688d-0723-4809-ba51-bcaf48524157 nodeName:}" failed. No retries permitted until 2025-11-25 15:19:10.14775315 +0000 UTC m=+1008.590215760 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/6517688d-0723-4809-ba51-bcaf48524157-etc-swift") pod "swift-storage-0" (UID: "6517688d-0723-4809-ba51-bcaf48524157") : configmap "swift-ring-files" not found Nov 25 15:19:06 crc kubenswrapper[4890]: I1125 15:19:06.742390 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"49e51424-feba-4038-88a6-1cd56637984c","Type":"ContainerStarted","Data":"a238d3d35d85d5a07ae383cacf77f2367a1426612d402e81dc164b329465edb3"} Nov 25 15:19:06 crc kubenswrapper[4890]: I1125 15:19:06.744857 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-vs9dw" event={"ID":"cf34ad11-7dd6-4572-b3d2-4c14127a1d76","Type":"ContainerStarted","Data":"ee321cbe36fab6fbdc08de675467b8b88bd16887233dd57e6eb443ed15a9617b"} Nov 25 15:19:06 crc kubenswrapper[4890]: I1125 15:19:06.747920 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-5tg6b" event={"ID":"289df7bb-0b8e-4411-ab2f-935629ed4019","Type":"ContainerStarted","Data":"d1209f3e952fc2625a1d457af46a46a7ee06773dcbd6906a830a233c9defd980"} Nov 25 15:19:06 crc kubenswrapper[4890]: I1125 15:19:06.747963 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-5tg6b" event={"ID":"289df7bb-0b8e-4411-ab2f-935629ed4019","Type":"ContainerStarted","Data":"19bd1bdd5385e43549d093a09e23c67eaa774980ae9909f3b08b10e86c520697"} Nov 25 15:19:06 crc kubenswrapper[4890]: I1125 15:19:06.749290 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-5tg6b" Nov 25 15:19:06 crc kubenswrapper[4890]: I1125 15:19:06.749348 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-5tg6b" Nov 25 15:19:06 crc kubenswrapper[4890]: I1125 15:19:06.752095 4890 generic.go:334] "Generic (PLEG): container finished" podID="f467b253-fb57-465a-8e67-5168b2dc0e55" containerID="4a63da169190aff17a808d433eed7bda43db2ec03573447985fd5d5264f50970" exitCode=0 Nov 25 15:19:06 crc kubenswrapper[4890]: I1125 15:19:06.752209 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-27vl6" event={"ID":"f467b253-fb57-465a-8e67-5168b2dc0e55","Type":"ContainerDied","Data":"4a63da169190aff17a808d433eed7bda43db2ec03573447985fd5d5264f50970"} Nov 25 15:19:06 crc kubenswrapper[4890]: I1125 15:19:06.756120 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Nov 25 15:19:06 crc kubenswrapper[4890]: I1125 15:19:06.763557 4890 generic.go:334] "Generic (PLEG): container finished" podID="536d18a7-2067-46d8-864c-8c84be2a87a8" containerID="a424f3c1d2d053db30b546c3021a04ea92673b7b8dc2f977885e308e4790efa0" exitCode=0 Nov 25 15:19:06 crc kubenswrapper[4890]: I1125 15:19:06.763720 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-7nfql" event={"ID":"536d18a7-2067-46d8-864c-8c84be2a87a8","Type":"ContainerDied","Data":"a424f3c1d2d053db30b546c3021a04ea92673b7b8dc2f977885e308e4790efa0"} Nov 25 15:19:06 crc kubenswrapper[4890]: I1125 15:19:06.787304 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=24.599911358 podStartE2EDuration="1m13.787280424s" podCreationTimestamp="2025-11-25 15:17:53 +0000 UTC" firstStartedPulling="2025-11-25 15:18:16.842330084 +0000 UTC m=+955.284792694" lastFinishedPulling="2025-11-25 15:19:06.02969915 +0000 UTC m=+1004.472161760" observedRunningTime="2025-11-25 15:19:06.786415532 +0000 UTC m=+1005.228878142" watchObservedRunningTime="2025-11-25 15:19:06.787280424 +0000 UTC m=+1005.229743034" Nov 25 15:19:06 crc kubenswrapper[4890]: I1125 15:19:06.801060 4890 generic.go:334] "Generic (PLEG): container finished" podID="2a74f4c1-2950-4305-b91f-50105ed2be79" containerID="a5043d70aace123925effb17a26960def0d5d81be666305ad6a0aa08abeba7b4" exitCode=0 Nov 25 15:19:06 crc kubenswrapper[4890]: I1125 15:19:06.801674 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-d8lf2" event={"ID":"2a74f4c1-2950-4305-b91f-50105ed2be79","Type":"ContainerDied","Data":"a5043d70aace123925effb17a26960def0d5d81be666305ad6a0aa08abeba7b4"} Nov 25 15:19:06 crc kubenswrapper[4890]: I1125 15:19:06.801812 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-d8lf2" event={"ID":"2a74f4c1-2950-4305-b91f-50105ed2be79","Type":"ContainerStarted","Data":"62c131099a60fd2a495b6556c9fc55f1320ef57b762369e3c67267d65109e896"} Nov 25 15:19:06 crc kubenswrapper[4890]: I1125 15:19:06.810566 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"c9667a27-2ed9-4b9a-a75a-4901a93cab5e","Type":"ContainerStarted","Data":"8dbd71e7e4bf017660d2725424549a55ffa3e8f42369cfbcf26e954f4bb700ff"} Nov 25 15:19:06 crc kubenswrapper[4890]: I1125 15:19:06.840405 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"24000ee2-b285-4f1a-afe1-75e643c1e172","Type":"ContainerStarted","Data":"49320dd334704d9959f25dc7d2d46346328fbd1b58da0ba378c9421bba59f6ee"} Nov 25 15:19:06 crc kubenswrapper[4890]: I1125 15:19:06.841322 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 25 15:19:06 crc kubenswrapper[4890]: I1125 15:19:06.966090 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-5tg6b" podStartSLOduration=57.02941037 podStartE2EDuration="1m12.966066722s" podCreationTimestamp="2025-11-25 15:17:54 +0000 UTC" firstStartedPulling="2025-11-25 15:18:14.002330444 +0000 UTC m=+952.444793074" lastFinishedPulling="2025-11-25 15:18:29.938986826 +0000 UTC m=+968.381449426" observedRunningTime="2025-11-25 15:19:06.920305229 +0000 UTC m=+1005.362767839" watchObservedRunningTime="2025-11-25 15:19:06.966066722 +0000 UTC m=+1005.408529332" Nov 25 15:19:06 crc kubenswrapper[4890]: I1125 15:19:06.986099 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=22.38578729 podStartE2EDuration="1m10.986074513s" podCreationTimestamp="2025-11-25 15:17:56 +0000 UTC" firstStartedPulling="2025-11-25 15:18:16.904835526 +0000 UTC m=+955.347298136" lastFinishedPulling="2025-11-25 15:19:05.505122749 +0000 UTC m=+1003.947585359" observedRunningTime="2025-11-25 15:19:06.944911404 +0000 UTC m=+1005.387374024" watchObservedRunningTime="2025-11-25 15:19:06.986074513 +0000 UTC m=+1005.428537123" Nov 25 15:19:06 crc kubenswrapper[4890]: I1125 15:19:06.990571 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=27.779821435 podStartE2EDuration="1m16.990559165s" podCreationTimestamp="2025-11-25 15:17:50 +0000 UTC" firstStartedPulling="2025-11-25 15:18:16.837673807 +0000 UTC m=+955.280136417" lastFinishedPulling="2025-11-25 15:19:06.048411537 +0000 UTC m=+1004.490874147" observedRunningTime="2025-11-25 15:19:06.972075053 +0000 UTC m=+1005.414537683" watchObservedRunningTime="2025-11-25 15:19:06.990559165 +0000 UTC m=+1005.433021775" Nov 25 15:19:07 crc kubenswrapper[4890]: I1125 15:19:07.001938 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Nov 25 15:19:07 crc kubenswrapper[4890]: I1125 15:19:07.030743 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Nov 25 15:19:07 crc kubenswrapper[4890]: I1125 15:19:07.197622 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Nov 25 15:19:07 crc kubenswrapper[4890]: E1125 15:19:07.255417 4890 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Nov 25 15:19:07 crc kubenswrapper[4890]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/f467b253-fb57-465a-8e67-5168b2dc0e55/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Nov 25 15:19:07 crc kubenswrapper[4890]: > podSandboxID="987bab745511b1ff1afe4fb20a103a10631d862c6389768fc840303ba7fca9a2" Nov 25 15:19:07 crc kubenswrapper[4890]: E1125 15:19:07.256067 4890 kuberuntime_manager.go:1274] "Unhandled Error" err=< Nov 25 15:19:07 crc kubenswrapper[4890]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-t4w4l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-27vl6_openstack(f467b253-fb57-465a-8e67-5168b2dc0e55): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/f467b253-fb57-465a-8e67-5168b2dc0e55/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Nov 25 15:19:07 crc kubenswrapper[4890]: > logger="UnhandledError" Nov 25 15:19:07 crc kubenswrapper[4890]: E1125 15:19:07.257244 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/f467b253-fb57-465a-8e67-5168b2dc0e55/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-57d769cc4f-27vl6" podUID="f467b253-fb57-465a-8e67-5168b2dc0e55" Nov 25 15:19:07 crc kubenswrapper[4890]: I1125 15:19:07.289313 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Nov 25 15:19:07 crc kubenswrapper[4890]: I1125 15:19:07.291633 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-7nfql" Nov 25 15:19:07 crc kubenswrapper[4890]: I1125 15:19:07.344589 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Nov 25 15:19:07 crc kubenswrapper[4890]: I1125 15:19:07.381738 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/536d18a7-2067-46d8-864c-8c84be2a87a8-dns-svc\") pod \"536d18a7-2067-46d8-864c-8c84be2a87a8\" (UID: \"536d18a7-2067-46d8-864c-8c84be2a87a8\") " Nov 25 15:19:07 crc kubenswrapper[4890]: I1125 15:19:07.381860 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/536d18a7-2067-46d8-864c-8c84be2a87a8-config\") pod \"536d18a7-2067-46d8-864c-8c84be2a87a8\" (UID: \"536d18a7-2067-46d8-864c-8c84be2a87a8\") " Nov 25 15:19:07 crc kubenswrapper[4890]: I1125 15:19:07.381890 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6qg9\" (UniqueName: \"kubernetes.io/projected/536d18a7-2067-46d8-864c-8c84be2a87a8-kube-api-access-x6qg9\") pod \"536d18a7-2067-46d8-864c-8c84be2a87a8\" (UID: \"536d18a7-2067-46d8-864c-8c84be2a87a8\") " Nov 25 15:19:07 crc kubenswrapper[4890]: I1125 15:19:07.386423 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/536d18a7-2067-46d8-864c-8c84be2a87a8-kube-api-access-x6qg9" (OuterVolumeSpecName: "kube-api-access-x6qg9") pod "536d18a7-2067-46d8-864c-8c84be2a87a8" (UID: "536d18a7-2067-46d8-864c-8c84be2a87a8"). InnerVolumeSpecName "kube-api-access-x6qg9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:19:07 crc kubenswrapper[4890]: I1125 15:19:07.400875 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/536d18a7-2067-46d8-864c-8c84be2a87a8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "536d18a7-2067-46d8-864c-8c84be2a87a8" (UID: "536d18a7-2067-46d8-864c-8c84be2a87a8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:19:07 crc kubenswrapper[4890]: I1125 15:19:07.402545 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/536d18a7-2067-46d8-864c-8c84be2a87a8-config" (OuterVolumeSpecName: "config") pod "536d18a7-2067-46d8-864c-8c84be2a87a8" (UID: "536d18a7-2067-46d8-864c-8c84be2a87a8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:19:07 crc kubenswrapper[4890]: I1125 15:19:07.484074 4890 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/536d18a7-2067-46d8-864c-8c84be2a87a8-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:07 crc kubenswrapper[4890]: I1125 15:19:07.484113 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6qg9\" (UniqueName: \"kubernetes.io/projected/536d18a7-2067-46d8-864c-8c84be2a87a8-kube-api-access-x6qg9\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:07 crc kubenswrapper[4890]: I1125 15:19:07.484123 4890 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/536d18a7-2067-46d8-864c-8c84be2a87a8-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:07 crc kubenswrapper[4890]: I1125 15:19:07.864657 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-7nfql" event={"ID":"536d18a7-2067-46d8-864c-8c84be2a87a8","Type":"ContainerDied","Data":"cd2f43755d4af48a478a8b9adda57cb42a52662faf40d5d6938059b090f834ff"} Nov 25 15:19:07 crc kubenswrapper[4890]: I1125 15:19:07.864696 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-7nfql" Nov 25 15:19:07 crc kubenswrapper[4890]: I1125 15:19:07.865188 4890 scope.go:117] "RemoveContainer" containerID="a424f3c1d2d053db30b546c3021a04ea92673b7b8dc2f977885e308e4790efa0" Nov 25 15:19:07 crc kubenswrapper[4890]: I1125 15:19:07.870525 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-d8lf2" event={"ID":"2a74f4c1-2950-4305-b91f-50105ed2be79","Type":"ContainerStarted","Data":"adaa66f7d9144b5a2f72bd271f56ce0f13c973c504470b1f68a70ada3009cad0"} Nov 25 15:19:07 crc kubenswrapper[4890]: I1125 15:19:07.871497 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Nov 25 15:19:07 crc kubenswrapper[4890]: I1125 15:19:07.871539 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7cb5889db5-d8lf2" Nov 25 15:19:07 crc kubenswrapper[4890]: I1125 15:19:07.871556 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Nov 25 15:19:07 crc kubenswrapper[4890]: I1125 15:19:07.893735 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7cb5889db5-d8lf2" podStartSLOduration=6.893718977 podStartE2EDuration="6.893718977s" podCreationTimestamp="2025-11-25 15:19:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:19:07.888260571 +0000 UTC m=+1006.330723181" watchObservedRunningTime="2025-11-25 15:19:07.893718977 +0000 UTC m=+1006.336181587" Nov 25 15:19:07 crc kubenswrapper[4890]: I1125 15:19:07.921642 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-7nfql"] Nov 25 15:19:07 crc kubenswrapper[4890]: I1125 15:19:07.926126 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-7nfql"] Nov 25 15:19:07 crc kubenswrapper[4890]: I1125 15:19:07.948686 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Nov 25 15:19:07 crc kubenswrapper[4890]: I1125 15:19:07.957301 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.187811 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="536d18a7-2067-46d8-864c-8c84be2a87a8" path="/var/lib/kubelet/pods/536d18a7-2067-46d8-864c-8c84be2a87a8/volumes" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.191012 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-27vl6"] Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.226697 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74f6f696b9-c9ctc"] Nov 25 15:19:08 crc kubenswrapper[4890]: E1125 15:19:08.227127 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="536d18a7-2067-46d8-864c-8c84be2a87a8" containerName="init" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.227142 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="536d18a7-2067-46d8-864c-8c84be2a87a8" containerName="init" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.227336 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="536d18a7-2067-46d8-864c-8c84be2a87a8" containerName="init" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.228225 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6f696b9-c9ctc" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.231353 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.302458 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c1e0704-e54b-48b3-bcc3-6af856dfd5ff-config\") pod \"dnsmasq-dns-74f6f696b9-c9ctc\" (UID: \"1c1e0704-e54b-48b3-bcc3-6af856dfd5ff\") " pod="openstack/dnsmasq-dns-74f6f696b9-c9ctc" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.302616 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ld8g8\" (UniqueName: \"kubernetes.io/projected/1c1e0704-e54b-48b3-bcc3-6af856dfd5ff-kube-api-access-ld8g8\") pod \"dnsmasq-dns-74f6f696b9-c9ctc\" (UID: \"1c1e0704-e54b-48b3-bcc3-6af856dfd5ff\") " pod="openstack/dnsmasq-dns-74f6f696b9-c9ctc" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.302663 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1c1e0704-e54b-48b3-bcc3-6af856dfd5ff-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6f696b9-c9ctc\" (UID: \"1c1e0704-e54b-48b3-bcc3-6af856dfd5ff\") " pod="openstack/dnsmasq-dns-74f6f696b9-c9ctc" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.302766 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1c1e0704-e54b-48b3-bcc3-6af856dfd5ff-dns-svc\") pod \"dnsmasq-dns-74f6f696b9-c9ctc\" (UID: \"1c1e0704-e54b-48b3-bcc3-6af856dfd5ff\") " pod="openstack/dnsmasq-dns-74f6f696b9-c9ctc" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.308983 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-v59z7"] Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.312191 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-v59z7" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.316917 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.341406 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-v59z7"] Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.356826 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6f696b9-c9ctc"] Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.404652 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/f282d61d-d8c2-4933-99c1-d95a9d4f40e2-ovn-rundir\") pod \"ovn-controller-metrics-v59z7\" (UID: \"f282d61d-d8c2-4933-99c1-d95a9d4f40e2\") " pod="openstack/ovn-controller-metrics-v59z7" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.404732 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c1e0704-e54b-48b3-bcc3-6af856dfd5ff-config\") pod \"dnsmasq-dns-74f6f696b9-c9ctc\" (UID: \"1c1e0704-e54b-48b3-bcc3-6af856dfd5ff\") " pod="openstack/dnsmasq-dns-74f6f696b9-c9ctc" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.404821 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/f282d61d-d8c2-4933-99c1-d95a9d4f40e2-ovs-rundir\") pod \"ovn-controller-metrics-v59z7\" (UID: \"f282d61d-d8c2-4933-99c1-d95a9d4f40e2\") " pod="openstack/ovn-controller-metrics-v59z7" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.404886 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbsbv\" (UniqueName: \"kubernetes.io/projected/f282d61d-d8c2-4933-99c1-d95a9d4f40e2-kube-api-access-cbsbv\") pod \"ovn-controller-metrics-v59z7\" (UID: \"f282d61d-d8c2-4933-99c1-d95a9d4f40e2\") " pod="openstack/ovn-controller-metrics-v59z7" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.405135 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ld8g8\" (UniqueName: \"kubernetes.io/projected/1c1e0704-e54b-48b3-bcc3-6af856dfd5ff-kube-api-access-ld8g8\") pod \"dnsmasq-dns-74f6f696b9-c9ctc\" (UID: \"1c1e0704-e54b-48b3-bcc3-6af856dfd5ff\") " pod="openstack/dnsmasq-dns-74f6f696b9-c9ctc" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.405240 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f282d61d-d8c2-4933-99c1-d95a9d4f40e2-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-v59z7\" (UID: \"f282d61d-d8c2-4933-99c1-d95a9d4f40e2\") " pod="openstack/ovn-controller-metrics-v59z7" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.405285 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1c1e0704-e54b-48b3-bcc3-6af856dfd5ff-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6f696b9-c9ctc\" (UID: \"1c1e0704-e54b-48b3-bcc3-6af856dfd5ff\") " pod="openstack/dnsmasq-dns-74f6f696b9-c9ctc" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.405422 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1c1e0704-e54b-48b3-bcc3-6af856dfd5ff-dns-svc\") pod \"dnsmasq-dns-74f6f696b9-c9ctc\" (UID: \"1c1e0704-e54b-48b3-bcc3-6af856dfd5ff\") " pod="openstack/dnsmasq-dns-74f6f696b9-c9ctc" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.405473 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f282d61d-d8c2-4933-99c1-d95a9d4f40e2-combined-ca-bundle\") pod \"ovn-controller-metrics-v59z7\" (UID: \"f282d61d-d8c2-4933-99c1-d95a9d4f40e2\") " pod="openstack/ovn-controller-metrics-v59z7" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.405511 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f282d61d-d8c2-4933-99c1-d95a9d4f40e2-config\") pod \"ovn-controller-metrics-v59z7\" (UID: \"f282d61d-d8c2-4933-99c1-d95a9d4f40e2\") " pod="openstack/ovn-controller-metrics-v59z7" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.405770 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c1e0704-e54b-48b3-bcc3-6af856dfd5ff-config\") pod \"dnsmasq-dns-74f6f696b9-c9ctc\" (UID: \"1c1e0704-e54b-48b3-bcc3-6af856dfd5ff\") " pod="openstack/dnsmasq-dns-74f6f696b9-c9ctc" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.406338 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1c1e0704-e54b-48b3-bcc3-6af856dfd5ff-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6f696b9-c9ctc\" (UID: \"1c1e0704-e54b-48b3-bcc3-6af856dfd5ff\") " pod="openstack/dnsmasq-dns-74f6f696b9-c9ctc" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.406785 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1c1e0704-e54b-48b3-bcc3-6af856dfd5ff-dns-svc\") pod \"dnsmasq-dns-74f6f696b9-c9ctc\" (UID: \"1c1e0704-e54b-48b3-bcc3-6af856dfd5ff\") " pod="openstack/dnsmasq-dns-74f6f696b9-c9ctc" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.428124 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ld8g8\" (UniqueName: \"kubernetes.io/projected/1c1e0704-e54b-48b3-bcc3-6af856dfd5ff-kube-api-access-ld8g8\") pod \"dnsmasq-dns-74f6f696b9-c9ctc\" (UID: \"1c1e0704-e54b-48b3-bcc3-6af856dfd5ff\") " pod="openstack/dnsmasq-dns-74f6f696b9-c9ctc" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.506917 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f282d61d-d8c2-4933-99c1-d95a9d4f40e2-combined-ca-bundle\") pod \"ovn-controller-metrics-v59z7\" (UID: \"f282d61d-d8c2-4933-99c1-d95a9d4f40e2\") " pod="openstack/ovn-controller-metrics-v59z7" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.507076 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f282d61d-d8c2-4933-99c1-d95a9d4f40e2-config\") pod \"ovn-controller-metrics-v59z7\" (UID: \"f282d61d-d8c2-4933-99c1-d95a9d4f40e2\") " pod="openstack/ovn-controller-metrics-v59z7" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.507125 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/f282d61d-d8c2-4933-99c1-d95a9d4f40e2-ovn-rundir\") pod \"ovn-controller-metrics-v59z7\" (UID: \"f282d61d-d8c2-4933-99c1-d95a9d4f40e2\") " pod="openstack/ovn-controller-metrics-v59z7" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.507331 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/f282d61d-d8c2-4933-99c1-d95a9d4f40e2-ovs-rundir\") pod \"ovn-controller-metrics-v59z7\" (UID: \"f282d61d-d8c2-4933-99c1-d95a9d4f40e2\") " pod="openstack/ovn-controller-metrics-v59z7" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.507421 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbsbv\" (UniqueName: \"kubernetes.io/projected/f282d61d-d8c2-4933-99c1-d95a9d4f40e2-kube-api-access-cbsbv\") pod \"ovn-controller-metrics-v59z7\" (UID: \"f282d61d-d8c2-4933-99c1-d95a9d4f40e2\") " pod="openstack/ovn-controller-metrics-v59z7" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.507459 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f282d61d-d8c2-4933-99c1-d95a9d4f40e2-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-v59z7\" (UID: \"f282d61d-d8c2-4933-99c1-d95a9d4f40e2\") " pod="openstack/ovn-controller-metrics-v59z7" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.507548 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/f282d61d-d8c2-4933-99c1-d95a9d4f40e2-ovn-rundir\") pod \"ovn-controller-metrics-v59z7\" (UID: \"f282d61d-d8c2-4933-99c1-d95a9d4f40e2\") " pod="openstack/ovn-controller-metrics-v59z7" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.507679 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/f282d61d-d8c2-4933-99c1-d95a9d4f40e2-ovs-rundir\") pod \"ovn-controller-metrics-v59z7\" (UID: \"f282d61d-d8c2-4933-99c1-d95a9d4f40e2\") " pod="openstack/ovn-controller-metrics-v59z7" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.507931 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f282d61d-d8c2-4933-99c1-d95a9d4f40e2-config\") pod \"ovn-controller-metrics-v59z7\" (UID: \"f282d61d-d8c2-4933-99c1-d95a9d4f40e2\") " pod="openstack/ovn-controller-metrics-v59z7" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.516007 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f282d61d-d8c2-4933-99c1-d95a9d4f40e2-combined-ca-bundle\") pod \"ovn-controller-metrics-v59z7\" (UID: \"f282d61d-d8c2-4933-99c1-d95a9d4f40e2\") " pod="openstack/ovn-controller-metrics-v59z7" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.519641 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f282d61d-d8c2-4933-99c1-d95a9d4f40e2-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-v59z7\" (UID: \"f282d61d-d8c2-4933-99c1-d95a9d4f40e2\") " pod="openstack/ovn-controller-metrics-v59z7" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.533450 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbsbv\" (UniqueName: \"kubernetes.io/projected/f282d61d-d8c2-4933-99c1-d95a9d4f40e2-kube-api-access-cbsbv\") pod \"ovn-controller-metrics-v59z7\" (UID: \"f282d61d-d8c2-4933-99c1-d95a9d4f40e2\") " pod="openstack/ovn-controller-metrics-v59z7" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.556103 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6f696b9-c9ctc" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.637326 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-v59z7" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.688934 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-d8lf2"] Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.725888 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-lgp8l"] Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.740359 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-lgp8l" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.745223 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-lgp8l"] Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.747452 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.764454 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.767386 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.772373 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.772614 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.772743 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.772881 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-f28kv" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.821893 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/effbba39-45db-4d3b-a883-03f38d0b49f8-scripts\") pod \"ovn-northd-0\" (UID: \"effbba39-45db-4d3b-a883-03f38d0b49f8\") " pod="openstack/ovn-northd-0" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.822304 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/df163dc8-8261-4fdf-b446-bba9e8e1d8eb-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-lgp8l\" (UID: \"df163dc8-8261-4fdf-b446-bba9e8e1d8eb\") " pod="openstack/dnsmasq-dns-698758b865-lgp8l" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.822403 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/effbba39-45db-4d3b-a883-03f38d0b49f8-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"effbba39-45db-4d3b-a883-03f38d0b49f8\") " pod="openstack/ovn-northd-0" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.822498 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v28mf\" (UniqueName: \"kubernetes.io/projected/df163dc8-8261-4fdf-b446-bba9e8e1d8eb-kube-api-access-v28mf\") pod \"dnsmasq-dns-698758b865-lgp8l\" (UID: \"df163dc8-8261-4fdf-b446-bba9e8e1d8eb\") " pod="openstack/dnsmasq-dns-698758b865-lgp8l" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.822572 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/df163dc8-8261-4fdf-b446-bba9e8e1d8eb-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-lgp8l\" (UID: \"df163dc8-8261-4fdf-b446-bba9e8e1d8eb\") " pod="openstack/dnsmasq-dns-698758b865-lgp8l" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.822697 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/effbba39-45db-4d3b-a883-03f38d0b49f8-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"effbba39-45db-4d3b-a883-03f38d0b49f8\") " pod="openstack/ovn-northd-0" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.822792 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/effbba39-45db-4d3b-a883-03f38d0b49f8-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"effbba39-45db-4d3b-a883-03f38d0b49f8\") " pod="openstack/ovn-northd-0" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.822890 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/effbba39-45db-4d3b-a883-03f38d0b49f8-config\") pod \"ovn-northd-0\" (UID: \"effbba39-45db-4d3b-a883-03f38d0b49f8\") " pod="openstack/ovn-northd-0" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.822963 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df163dc8-8261-4fdf-b446-bba9e8e1d8eb-config\") pod \"dnsmasq-dns-698758b865-lgp8l\" (UID: \"df163dc8-8261-4fdf-b446-bba9e8e1d8eb\") " pod="openstack/dnsmasq-dns-698758b865-lgp8l" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.823045 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tvmr\" (UniqueName: \"kubernetes.io/projected/effbba39-45db-4d3b-a883-03f38d0b49f8-kube-api-access-8tvmr\") pod \"ovn-northd-0\" (UID: \"effbba39-45db-4d3b-a883-03f38d0b49f8\") " pod="openstack/ovn-northd-0" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.823131 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/effbba39-45db-4d3b-a883-03f38d0b49f8-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"effbba39-45db-4d3b-a883-03f38d0b49f8\") " pod="openstack/ovn-northd-0" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.823239 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/df163dc8-8261-4fdf-b446-bba9e8e1d8eb-dns-svc\") pod \"dnsmasq-dns-698758b865-lgp8l\" (UID: \"df163dc8-8261-4fdf-b446-bba9e8e1d8eb\") " pod="openstack/dnsmasq-dns-698758b865-lgp8l" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.823517 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-4ppjx"] Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.830646 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-4ppjx" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.842129 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-0806-account-create-xr7r5"] Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.844874 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-0806-account-create-xr7r5" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.848201 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.882442 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.894421 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-4ppjx"] Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.903758 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-0806-account-create-xr7r5"] Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.924785 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df163dc8-8261-4fdf-b446-bba9e8e1d8eb-config\") pod \"dnsmasq-dns-698758b865-lgp8l\" (UID: \"df163dc8-8261-4fdf-b446-bba9e8e1d8eb\") " pod="openstack/dnsmasq-dns-698758b865-lgp8l" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.924827 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tvmr\" (UniqueName: \"kubernetes.io/projected/effbba39-45db-4d3b-a883-03f38d0b49f8-kube-api-access-8tvmr\") pod \"ovn-northd-0\" (UID: \"effbba39-45db-4d3b-a883-03f38d0b49f8\") " pod="openstack/ovn-northd-0" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.924850 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/effbba39-45db-4d3b-a883-03f38d0b49f8-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"effbba39-45db-4d3b-a883-03f38d0b49f8\") " pod="openstack/ovn-northd-0" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.924873 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/df163dc8-8261-4fdf-b446-bba9e8e1d8eb-dns-svc\") pod \"dnsmasq-dns-698758b865-lgp8l\" (UID: \"df163dc8-8261-4fdf-b446-bba9e8e1d8eb\") " pod="openstack/dnsmasq-dns-698758b865-lgp8l" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.924898 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd3ad34d-4081-4341-b103-20d6ba8da9c5-operator-scripts\") pod \"keystone-0806-account-create-xr7r5\" (UID: \"dd3ad34d-4081-4341-b103-20d6ba8da9c5\") " pod="openstack/keystone-0806-account-create-xr7r5" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.924917 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9076492c-2832-41b8-966c-f0ca12a43792-operator-scripts\") pod \"keystone-db-create-4ppjx\" (UID: \"9076492c-2832-41b8-966c-f0ca12a43792\") " pod="openstack/keystone-db-create-4ppjx" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.924988 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whgj5\" (UniqueName: \"kubernetes.io/projected/dd3ad34d-4081-4341-b103-20d6ba8da9c5-kube-api-access-whgj5\") pod \"keystone-0806-account-create-xr7r5\" (UID: \"dd3ad34d-4081-4341-b103-20d6ba8da9c5\") " pod="openstack/keystone-0806-account-create-xr7r5" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.925038 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/effbba39-45db-4d3b-a883-03f38d0b49f8-scripts\") pod \"ovn-northd-0\" (UID: \"effbba39-45db-4d3b-a883-03f38d0b49f8\") " pod="openstack/ovn-northd-0" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.925072 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/df163dc8-8261-4fdf-b446-bba9e8e1d8eb-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-lgp8l\" (UID: \"df163dc8-8261-4fdf-b446-bba9e8e1d8eb\") " pod="openstack/dnsmasq-dns-698758b865-lgp8l" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.925098 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/effbba39-45db-4d3b-a883-03f38d0b49f8-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"effbba39-45db-4d3b-a883-03f38d0b49f8\") " pod="openstack/ovn-northd-0" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.925118 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmp9s\" (UniqueName: \"kubernetes.io/projected/9076492c-2832-41b8-966c-f0ca12a43792-kube-api-access-cmp9s\") pod \"keystone-db-create-4ppjx\" (UID: \"9076492c-2832-41b8-966c-f0ca12a43792\") " pod="openstack/keystone-db-create-4ppjx" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.925154 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v28mf\" (UniqueName: \"kubernetes.io/projected/df163dc8-8261-4fdf-b446-bba9e8e1d8eb-kube-api-access-v28mf\") pod \"dnsmasq-dns-698758b865-lgp8l\" (UID: \"df163dc8-8261-4fdf-b446-bba9e8e1d8eb\") " pod="openstack/dnsmasq-dns-698758b865-lgp8l" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.925190 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/df163dc8-8261-4fdf-b446-bba9e8e1d8eb-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-lgp8l\" (UID: \"df163dc8-8261-4fdf-b446-bba9e8e1d8eb\") " pod="openstack/dnsmasq-dns-698758b865-lgp8l" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.925212 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/effbba39-45db-4d3b-a883-03f38d0b49f8-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"effbba39-45db-4d3b-a883-03f38d0b49f8\") " pod="openstack/ovn-northd-0" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.925266 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/effbba39-45db-4d3b-a883-03f38d0b49f8-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"effbba39-45db-4d3b-a883-03f38d0b49f8\") " pod="openstack/ovn-northd-0" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.925330 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/effbba39-45db-4d3b-a883-03f38d0b49f8-config\") pod \"ovn-northd-0\" (UID: \"effbba39-45db-4d3b-a883-03f38d0b49f8\") " pod="openstack/ovn-northd-0" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.926319 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/df163dc8-8261-4fdf-b446-bba9e8e1d8eb-dns-svc\") pod \"dnsmasq-dns-698758b865-lgp8l\" (UID: \"df163dc8-8261-4fdf-b446-bba9e8e1d8eb\") " pod="openstack/dnsmasq-dns-698758b865-lgp8l" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.927401 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/effbba39-45db-4d3b-a883-03f38d0b49f8-config\") pod \"ovn-northd-0\" (UID: \"effbba39-45db-4d3b-a883-03f38d0b49f8\") " pod="openstack/ovn-northd-0" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.927146 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/df163dc8-8261-4fdf-b446-bba9e8e1d8eb-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-lgp8l\" (UID: \"df163dc8-8261-4fdf-b446-bba9e8e1d8eb\") " pod="openstack/dnsmasq-dns-698758b865-lgp8l" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.928475 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/df163dc8-8261-4fdf-b446-bba9e8e1d8eb-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-lgp8l\" (UID: \"df163dc8-8261-4fdf-b446-bba9e8e1d8eb\") " pod="openstack/dnsmasq-dns-698758b865-lgp8l" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.928825 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/effbba39-45db-4d3b-a883-03f38d0b49f8-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"effbba39-45db-4d3b-a883-03f38d0b49f8\") " pod="openstack/ovn-northd-0" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.928862 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/effbba39-45db-4d3b-a883-03f38d0b49f8-scripts\") pod \"ovn-northd-0\" (UID: \"effbba39-45db-4d3b-a883-03f38d0b49f8\") " pod="openstack/ovn-northd-0" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.929004 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df163dc8-8261-4fdf-b446-bba9e8e1d8eb-config\") pod \"dnsmasq-dns-698758b865-lgp8l\" (UID: \"df163dc8-8261-4fdf-b446-bba9e8e1d8eb\") " pod="openstack/dnsmasq-dns-698758b865-lgp8l" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.931028 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/effbba39-45db-4d3b-a883-03f38d0b49f8-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"effbba39-45db-4d3b-a883-03f38d0b49f8\") " pod="openstack/ovn-northd-0" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.931619 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/effbba39-45db-4d3b-a883-03f38d0b49f8-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"effbba39-45db-4d3b-a883-03f38d0b49f8\") " pod="openstack/ovn-northd-0" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.933488 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/effbba39-45db-4d3b-a883-03f38d0b49f8-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"effbba39-45db-4d3b-a883-03f38d0b49f8\") " pod="openstack/ovn-northd-0" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.956518 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tvmr\" (UniqueName: \"kubernetes.io/projected/effbba39-45db-4d3b-a883-03f38d0b49f8-kube-api-access-8tvmr\") pod \"ovn-northd-0\" (UID: \"effbba39-45db-4d3b-a883-03f38d0b49f8\") " pod="openstack/ovn-northd-0" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.957413 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v28mf\" (UniqueName: \"kubernetes.io/projected/df163dc8-8261-4fdf-b446-bba9e8e1d8eb-kube-api-access-v28mf\") pod \"dnsmasq-dns-698758b865-lgp8l\" (UID: \"df163dc8-8261-4fdf-b446-bba9e8e1d8eb\") " pod="openstack/dnsmasq-dns-698758b865-lgp8l" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.967368 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-w9c2z"] Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.968710 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-w9c2z" Nov 25 15:19:08 crc kubenswrapper[4890]: I1125 15:19:08.987118 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-w9c2z"] Nov 25 15:19:09 crc kubenswrapper[4890]: I1125 15:19:09.030059 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b69893e0-19f6-425b-b36f-b39fd2fd80d5-operator-scripts\") pod \"placement-db-create-w9c2z\" (UID: \"b69893e0-19f6-425b-b36f-b39fd2fd80d5\") " pod="openstack/placement-db-create-w9c2z" Nov 25 15:19:09 crc kubenswrapper[4890]: I1125 15:19:09.030140 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvsph\" (UniqueName: \"kubernetes.io/projected/b69893e0-19f6-425b-b36f-b39fd2fd80d5-kube-api-access-vvsph\") pod \"placement-db-create-w9c2z\" (UID: \"b69893e0-19f6-425b-b36f-b39fd2fd80d5\") " pod="openstack/placement-db-create-w9c2z" Nov 25 15:19:09 crc kubenswrapper[4890]: I1125 15:19:09.030285 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd3ad34d-4081-4341-b103-20d6ba8da9c5-operator-scripts\") pod \"keystone-0806-account-create-xr7r5\" (UID: \"dd3ad34d-4081-4341-b103-20d6ba8da9c5\") " pod="openstack/keystone-0806-account-create-xr7r5" Nov 25 15:19:09 crc kubenswrapper[4890]: I1125 15:19:09.030328 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9076492c-2832-41b8-966c-f0ca12a43792-operator-scripts\") pod \"keystone-db-create-4ppjx\" (UID: \"9076492c-2832-41b8-966c-f0ca12a43792\") " pod="openstack/keystone-db-create-4ppjx" Nov 25 15:19:09 crc kubenswrapper[4890]: I1125 15:19:09.030389 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whgj5\" (UniqueName: \"kubernetes.io/projected/dd3ad34d-4081-4341-b103-20d6ba8da9c5-kube-api-access-whgj5\") pod \"keystone-0806-account-create-xr7r5\" (UID: \"dd3ad34d-4081-4341-b103-20d6ba8da9c5\") " pod="openstack/keystone-0806-account-create-xr7r5" Nov 25 15:19:09 crc kubenswrapper[4890]: I1125 15:19:09.030449 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmp9s\" (UniqueName: \"kubernetes.io/projected/9076492c-2832-41b8-966c-f0ca12a43792-kube-api-access-cmp9s\") pod \"keystone-db-create-4ppjx\" (UID: \"9076492c-2832-41b8-966c-f0ca12a43792\") " pod="openstack/keystone-db-create-4ppjx" Nov 25 15:19:09 crc kubenswrapper[4890]: I1125 15:19:09.031272 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd3ad34d-4081-4341-b103-20d6ba8da9c5-operator-scripts\") pod \"keystone-0806-account-create-xr7r5\" (UID: \"dd3ad34d-4081-4341-b103-20d6ba8da9c5\") " pod="openstack/keystone-0806-account-create-xr7r5" Nov 25 15:19:09 crc kubenswrapper[4890]: I1125 15:19:09.031601 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9076492c-2832-41b8-966c-f0ca12a43792-operator-scripts\") pod \"keystone-db-create-4ppjx\" (UID: \"9076492c-2832-41b8-966c-f0ca12a43792\") " pod="openstack/keystone-db-create-4ppjx" Nov 25 15:19:09 crc kubenswrapper[4890]: I1125 15:19:09.052642 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whgj5\" (UniqueName: \"kubernetes.io/projected/dd3ad34d-4081-4341-b103-20d6ba8da9c5-kube-api-access-whgj5\") pod \"keystone-0806-account-create-xr7r5\" (UID: \"dd3ad34d-4081-4341-b103-20d6ba8da9c5\") " pod="openstack/keystone-0806-account-create-xr7r5" Nov 25 15:19:09 crc kubenswrapper[4890]: I1125 15:19:09.068743 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmp9s\" (UniqueName: \"kubernetes.io/projected/9076492c-2832-41b8-966c-f0ca12a43792-kube-api-access-cmp9s\") pod \"keystone-db-create-4ppjx\" (UID: \"9076492c-2832-41b8-966c-f0ca12a43792\") " pod="openstack/keystone-db-create-4ppjx" Nov 25 15:19:09 crc kubenswrapper[4890]: I1125 15:19:09.075452 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-lgp8l" Nov 25 15:19:09 crc kubenswrapper[4890]: I1125 15:19:09.096231 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-aa80-account-create-hcvwx"] Nov 25 15:19:09 crc kubenswrapper[4890]: I1125 15:19:09.098372 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-aa80-account-create-hcvwx" Nov 25 15:19:09 crc kubenswrapper[4890]: I1125 15:19:09.102988 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Nov 25 15:19:09 crc kubenswrapper[4890]: I1125 15:19:09.106140 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-aa80-account-create-hcvwx"] Nov 25 15:19:09 crc kubenswrapper[4890]: I1125 15:19:09.108555 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 25 15:19:09 crc kubenswrapper[4890]: I1125 15:19:09.132937 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4gf7\" (UniqueName: \"kubernetes.io/projected/3af7c86b-336f-48ce-b5bf-294c0b8bc5ee-kube-api-access-k4gf7\") pod \"placement-aa80-account-create-hcvwx\" (UID: \"3af7c86b-336f-48ce-b5bf-294c0b8bc5ee\") " pod="openstack/placement-aa80-account-create-hcvwx" Nov 25 15:19:09 crc kubenswrapper[4890]: I1125 15:19:09.133049 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b69893e0-19f6-425b-b36f-b39fd2fd80d5-operator-scripts\") pod \"placement-db-create-w9c2z\" (UID: \"b69893e0-19f6-425b-b36f-b39fd2fd80d5\") " pod="openstack/placement-db-create-w9c2z" Nov 25 15:19:09 crc kubenswrapper[4890]: I1125 15:19:09.133333 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvsph\" (UniqueName: \"kubernetes.io/projected/b69893e0-19f6-425b-b36f-b39fd2fd80d5-kube-api-access-vvsph\") pod \"placement-db-create-w9c2z\" (UID: \"b69893e0-19f6-425b-b36f-b39fd2fd80d5\") " pod="openstack/placement-db-create-w9c2z" Nov 25 15:19:09 crc kubenswrapper[4890]: I1125 15:19:09.133815 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3af7c86b-336f-48ce-b5bf-294c0b8bc5ee-operator-scripts\") pod \"placement-aa80-account-create-hcvwx\" (UID: \"3af7c86b-336f-48ce-b5bf-294c0b8bc5ee\") " pod="openstack/placement-aa80-account-create-hcvwx" Nov 25 15:19:09 crc kubenswrapper[4890]: I1125 15:19:09.133840 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b69893e0-19f6-425b-b36f-b39fd2fd80d5-operator-scripts\") pod \"placement-db-create-w9c2z\" (UID: \"b69893e0-19f6-425b-b36f-b39fd2fd80d5\") " pod="openstack/placement-db-create-w9c2z" Nov 25 15:19:09 crc kubenswrapper[4890]: I1125 15:19:09.153835 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvsph\" (UniqueName: \"kubernetes.io/projected/b69893e0-19f6-425b-b36f-b39fd2fd80d5-kube-api-access-vvsph\") pod \"placement-db-create-w9c2z\" (UID: \"b69893e0-19f6-425b-b36f-b39fd2fd80d5\") " pod="openstack/placement-db-create-w9c2z" Nov 25 15:19:09 crc kubenswrapper[4890]: I1125 15:19:09.172196 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-4ppjx" Nov 25 15:19:09 crc kubenswrapper[4890]: I1125 15:19:09.185906 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-0806-account-create-xr7r5" Nov 25 15:19:09 crc kubenswrapper[4890]: I1125 15:19:09.235416 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3af7c86b-336f-48ce-b5bf-294c0b8bc5ee-operator-scripts\") pod \"placement-aa80-account-create-hcvwx\" (UID: \"3af7c86b-336f-48ce-b5bf-294c0b8bc5ee\") " pod="openstack/placement-aa80-account-create-hcvwx" Nov 25 15:19:09 crc kubenswrapper[4890]: I1125 15:19:09.235509 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4gf7\" (UniqueName: \"kubernetes.io/projected/3af7c86b-336f-48ce-b5bf-294c0b8bc5ee-kube-api-access-k4gf7\") pod \"placement-aa80-account-create-hcvwx\" (UID: \"3af7c86b-336f-48ce-b5bf-294c0b8bc5ee\") " pod="openstack/placement-aa80-account-create-hcvwx" Nov 25 15:19:09 crc kubenswrapper[4890]: I1125 15:19:09.236243 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3af7c86b-336f-48ce-b5bf-294c0b8bc5ee-operator-scripts\") pod \"placement-aa80-account-create-hcvwx\" (UID: \"3af7c86b-336f-48ce-b5bf-294c0b8bc5ee\") " pod="openstack/placement-aa80-account-create-hcvwx" Nov 25 15:19:09 crc kubenswrapper[4890]: I1125 15:19:09.257701 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4gf7\" (UniqueName: \"kubernetes.io/projected/3af7c86b-336f-48ce-b5bf-294c0b8bc5ee-kube-api-access-k4gf7\") pod \"placement-aa80-account-create-hcvwx\" (UID: \"3af7c86b-336f-48ce-b5bf-294c0b8bc5ee\") " pod="openstack/placement-aa80-account-create-hcvwx" Nov 25 15:19:09 crc kubenswrapper[4890]: I1125 15:19:09.329752 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-w9c2z" Nov 25 15:19:09 crc kubenswrapper[4890]: I1125 15:19:09.421936 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-aa80-account-create-hcvwx" Nov 25 15:19:09 crc kubenswrapper[4890]: I1125 15:19:09.893609 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7cb5889db5-d8lf2" podUID="2a74f4c1-2950-4305-b91f-50105ed2be79" containerName="dnsmasq-dns" containerID="cri-o://adaa66f7d9144b5a2f72bd271f56ce0f13c973c504470b1f68a70ada3009cad0" gracePeriod=10 Nov 25 15:19:10 crc kubenswrapper[4890]: I1125 15:19:10.151175 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6517688d-0723-4809-ba51-bcaf48524157-etc-swift\") pod \"swift-storage-0\" (UID: \"6517688d-0723-4809-ba51-bcaf48524157\") " pod="openstack/swift-storage-0" Nov 25 15:19:10 crc kubenswrapper[4890]: E1125 15:19:10.151376 4890 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 25 15:19:10 crc kubenswrapper[4890]: E1125 15:19:10.151392 4890 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 25 15:19:10 crc kubenswrapper[4890]: E1125 15:19:10.151435 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/6517688d-0723-4809-ba51-bcaf48524157-etc-swift podName:6517688d-0723-4809-ba51-bcaf48524157 nodeName:}" failed. No retries permitted until 2025-11-25 15:19:18.151421924 +0000 UTC m=+1016.593884534 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/6517688d-0723-4809-ba51-bcaf48524157-etc-swift") pod "swift-storage-0" (UID: "6517688d-0723-4809-ba51-bcaf48524157") : configmap "swift-ring-files" not found Nov 25 15:19:10 crc kubenswrapper[4890]: I1125 15:19:10.664305 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-27vl6" Nov 25 15:19:10 crc kubenswrapper[4890]: I1125 15:19:10.764986 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f467b253-fb57-465a-8e67-5168b2dc0e55-dns-svc\") pod \"f467b253-fb57-465a-8e67-5168b2dc0e55\" (UID: \"f467b253-fb57-465a-8e67-5168b2dc0e55\") " Nov 25 15:19:10 crc kubenswrapper[4890]: I1125 15:19:10.765076 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t4w4l\" (UniqueName: \"kubernetes.io/projected/f467b253-fb57-465a-8e67-5168b2dc0e55-kube-api-access-t4w4l\") pod \"f467b253-fb57-465a-8e67-5168b2dc0e55\" (UID: \"f467b253-fb57-465a-8e67-5168b2dc0e55\") " Nov 25 15:19:10 crc kubenswrapper[4890]: I1125 15:19:10.765157 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f467b253-fb57-465a-8e67-5168b2dc0e55-config\") pod \"f467b253-fb57-465a-8e67-5168b2dc0e55\" (UID: \"f467b253-fb57-465a-8e67-5168b2dc0e55\") " Nov 25 15:19:10 crc kubenswrapper[4890]: I1125 15:19:10.772277 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f467b253-fb57-465a-8e67-5168b2dc0e55-kube-api-access-t4w4l" (OuterVolumeSpecName: "kube-api-access-t4w4l") pod "f467b253-fb57-465a-8e67-5168b2dc0e55" (UID: "f467b253-fb57-465a-8e67-5168b2dc0e55"). InnerVolumeSpecName "kube-api-access-t4w4l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:19:10 crc kubenswrapper[4890]: I1125 15:19:10.868329 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t4w4l\" (UniqueName: \"kubernetes.io/projected/f467b253-fb57-465a-8e67-5168b2dc0e55-kube-api-access-t4w4l\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:10 crc kubenswrapper[4890]: I1125 15:19:10.871415 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f467b253-fb57-465a-8e67-5168b2dc0e55-config" (OuterVolumeSpecName: "config") pod "f467b253-fb57-465a-8e67-5168b2dc0e55" (UID: "f467b253-fb57-465a-8e67-5168b2dc0e55"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:19:10 crc kubenswrapper[4890]: I1125 15:19:10.903775 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f467b253-fb57-465a-8e67-5168b2dc0e55-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f467b253-fb57-465a-8e67-5168b2dc0e55" (UID: "f467b253-fb57-465a-8e67-5168b2dc0e55"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:19:10 crc kubenswrapper[4890]: I1125 15:19:10.920801 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-d8lf2" Nov 25 15:19:10 crc kubenswrapper[4890]: I1125 15:19:10.935925 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-27vl6" event={"ID":"f467b253-fb57-465a-8e67-5168b2dc0e55","Type":"ContainerDied","Data":"987bab745511b1ff1afe4fb20a103a10631d862c6389768fc840303ba7fca9a2"} Nov 25 15:19:10 crc kubenswrapper[4890]: I1125 15:19:10.935987 4890 scope.go:117] "RemoveContainer" containerID="4a63da169190aff17a808d433eed7bda43db2ec03573447985fd5d5264f50970" Nov 25 15:19:10 crc kubenswrapper[4890]: I1125 15:19:10.936094 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-27vl6" Nov 25 15:19:10 crc kubenswrapper[4890]: I1125 15:19:10.963686 4890 generic.go:334] "Generic (PLEG): container finished" podID="2a74f4c1-2950-4305-b91f-50105ed2be79" containerID="adaa66f7d9144b5a2f72bd271f56ce0f13c973c504470b1f68a70ada3009cad0" exitCode=0 Nov 25 15:19:10 crc kubenswrapper[4890]: I1125 15:19:10.963731 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-d8lf2" event={"ID":"2a74f4c1-2950-4305-b91f-50105ed2be79","Type":"ContainerDied","Data":"adaa66f7d9144b5a2f72bd271f56ce0f13c973c504470b1f68a70ada3009cad0"} Nov 25 15:19:10 crc kubenswrapper[4890]: I1125 15:19:10.963758 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-d8lf2" event={"ID":"2a74f4c1-2950-4305-b91f-50105ed2be79","Type":"ContainerDied","Data":"62c131099a60fd2a495b6556c9fc55f1320ef57b762369e3c67267d65109e896"} Nov 25 15:19:10 crc kubenswrapper[4890]: I1125 15:19:10.963813 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-d8lf2" Nov 25 15:19:10 crc kubenswrapper[4890]: I1125 15:19:10.969634 4890 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f467b253-fb57-465a-8e67-5168b2dc0e55-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:10 crc kubenswrapper[4890]: I1125 15:19:10.969665 4890 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f467b253-fb57-465a-8e67-5168b2dc0e55-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:11 crc kubenswrapper[4890]: I1125 15:19:11.024528 4890 scope.go:117] "RemoveContainer" containerID="adaa66f7d9144b5a2f72bd271f56ce0f13c973c504470b1f68a70ada3009cad0" Nov 25 15:19:11 crc kubenswrapper[4890]: I1125 15:19:11.064310 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-27vl6"] Nov 25 15:19:11 crc kubenswrapper[4890]: I1125 15:19:11.071006 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-27vl6"] Nov 25 15:19:11 crc kubenswrapper[4890]: I1125 15:19:11.071649 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rs2h9\" (UniqueName: \"kubernetes.io/projected/2a74f4c1-2950-4305-b91f-50105ed2be79-kube-api-access-rs2h9\") pod \"2a74f4c1-2950-4305-b91f-50105ed2be79\" (UID: \"2a74f4c1-2950-4305-b91f-50105ed2be79\") " Nov 25 15:19:11 crc kubenswrapper[4890]: I1125 15:19:11.071739 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a74f4c1-2950-4305-b91f-50105ed2be79-config\") pod \"2a74f4c1-2950-4305-b91f-50105ed2be79\" (UID: \"2a74f4c1-2950-4305-b91f-50105ed2be79\") " Nov 25 15:19:11 crc kubenswrapper[4890]: I1125 15:19:11.071882 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2a74f4c1-2950-4305-b91f-50105ed2be79-dns-svc\") pod \"2a74f4c1-2950-4305-b91f-50105ed2be79\" (UID: \"2a74f4c1-2950-4305-b91f-50105ed2be79\") " Nov 25 15:19:11 crc kubenswrapper[4890]: I1125 15:19:11.075846 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a74f4c1-2950-4305-b91f-50105ed2be79-kube-api-access-rs2h9" (OuterVolumeSpecName: "kube-api-access-rs2h9") pod "2a74f4c1-2950-4305-b91f-50105ed2be79" (UID: "2a74f4c1-2950-4305-b91f-50105ed2be79"). InnerVolumeSpecName "kube-api-access-rs2h9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:19:11 crc kubenswrapper[4890]: I1125 15:19:11.104285 4890 scope.go:117] "RemoveContainer" containerID="a5043d70aace123925effb17a26960def0d5d81be666305ad6a0aa08abeba7b4" Nov 25 15:19:11 crc kubenswrapper[4890]: I1125 15:19:11.124403 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a74f4c1-2950-4305-b91f-50105ed2be79-config" (OuterVolumeSpecName: "config") pod "2a74f4c1-2950-4305-b91f-50105ed2be79" (UID: "2a74f4c1-2950-4305-b91f-50105ed2be79"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:19:11 crc kubenswrapper[4890]: I1125 15:19:11.128024 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a74f4c1-2950-4305-b91f-50105ed2be79-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2a74f4c1-2950-4305-b91f-50105ed2be79" (UID: "2a74f4c1-2950-4305-b91f-50105ed2be79"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:19:11 crc kubenswrapper[4890]: I1125 15:19:11.131631 4890 scope.go:117] "RemoveContainer" containerID="adaa66f7d9144b5a2f72bd271f56ce0f13c973c504470b1f68a70ada3009cad0" Nov 25 15:19:11 crc kubenswrapper[4890]: E1125 15:19:11.132055 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"adaa66f7d9144b5a2f72bd271f56ce0f13c973c504470b1f68a70ada3009cad0\": container with ID starting with adaa66f7d9144b5a2f72bd271f56ce0f13c973c504470b1f68a70ada3009cad0 not found: ID does not exist" containerID="adaa66f7d9144b5a2f72bd271f56ce0f13c973c504470b1f68a70ada3009cad0" Nov 25 15:19:11 crc kubenswrapper[4890]: I1125 15:19:11.132100 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"adaa66f7d9144b5a2f72bd271f56ce0f13c973c504470b1f68a70ada3009cad0"} err="failed to get container status \"adaa66f7d9144b5a2f72bd271f56ce0f13c973c504470b1f68a70ada3009cad0\": rpc error: code = NotFound desc = could not find container \"adaa66f7d9144b5a2f72bd271f56ce0f13c973c504470b1f68a70ada3009cad0\": container with ID starting with adaa66f7d9144b5a2f72bd271f56ce0f13c973c504470b1f68a70ada3009cad0 not found: ID does not exist" Nov 25 15:19:11 crc kubenswrapper[4890]: I1125 15:19:11.132125 4890 scope.go:117] "RemoveContainer" containerID="a5043d70aace123925effb17a26960def0d5d81be666305ad6a0aa08abeba7b4" Nov 25 15:19:11 crc kubenswrapper[4890]: E1125 15:19:11.132617 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5043d70aace123925effb17a26960def0d5d81be666305ad6a0aa08abeba7b4\": container with ID starting with a5043d70aace123925effb17a26960def0d5d81be666305ad6a0aa08abeba7b4 not found: ID does not exist" containerID="a5043d70aace123925effb17a26960def0d5d81be666305ad6a0aa08abeba7b4" Nov 25 15:19:11 crc kubenswrapper[4890]: I1125 15:19:11.132666 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5043d70aace123925effb17a26960def0d5d81be666305ad6a0aa08abeba7b4"} err="failed to get container status \"a5043d70aace123925effb17a26960def0d5d81be666305ad6a0aa08abeba7b4\": rpc error: code = NotFound desc = could not find container \"a5043d70aace123925effb17a26960def0d5d81be666305ad6a0aa08abeba7b4\": container with ID starting with a5043d70aace123925effb17a26960def0d5d81be666305ad6a0aa08abeba7b4 not found: ID does not exist" Nov 25 15:19:11 crc kubenswrapper[4890]: I1125 15:19:11.177631 4890 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2a74f4c1-2950-4305-b91f-50105ed2be79-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:11 crc kubenswrapper[4890]: I1125 15:19:11.177665 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rs2h9\" (UniqueName: \"kubernetes.io/projected/2a74f4c1-2950-4305-b91f-50105ed2be79-kube-api-access-rs2h9\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:11 crc kubenswrapper[4890]: I1125 15:19:11.177677 4890 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a74f4c1-2950-4305-b91f-50105ed2be79-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:11 crc kubenswrapper[4890]: I1125 15:19:11.189664 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 25 15:19:11 crc kubenswrapper[4890]: I1125 15:19:11.302999 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-d8lf2"] Nov 25 15:19:11 crc kubenswrapper[4890]: I1125 15:19:11.309918 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-d8lf2"] Nov 25 15:19:11 crc kubenswrapper[4890]: I1125 15:19:11.509704 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-0806-account-create-xr7r5"] Nov 25 15:19:11 crc kubenswrapper[4890]: W1125 15:19:11.514813 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeffbba39_45db_4d3b_a883_03f38d0b49f8.slice/crio-29afb1e723d124d86cfda2680a825900b80a6aa5029589e32a1814cdd2c047a9 WatchSource:0}: Error finding container 29afb1e723d124d86cfda2680a825900b80a6aa5029589e32a1814cdd2c047a9: Status 404 returned error can't find the container with id 29afb1e723d124d86cfda2680a825900b80a6aa5029589e32a1814cdd2c047a9 Nov 25 15:19:11 crc kubenswrapper[4890]: W1125 15:19:11.518014 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddd3ad34d_4081_4341_b103_20d6ba8da9c5.slice/crio-d878105c2a5f6bc14a01c801636d4e8549ca10a71212ef0a03b561e0e9dca092 WatchSource:0}: Error finding container d878105c2a5f6bc14a01c801636d4e8549ca10a71212ef0a03b561e0e9dca092: Status 404 returned error can't find the container with id d878105c2a5f6bc14a01c801636d4e8549ca10a71212ef0a03b561e0e9dca092 Nov 25 15:19:11 crc kubenswrapper[4890]: I1125 15:19:11.519507 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 25 15:19:11 crc kubenswrapper[4890]: I1125 15:19:11.702445 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-aa80-account-create-hcvwx"] Nov 25 15:19:11 crc kubenswrapper[4890]: I1125 15:19:11.733668 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-4ppjx"] Nov 25 15:19:11 crc kubenswrapper[4890]: W1125 15:19:11.741432 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9076492c_2832_41b8_966c_f0ca12a43792.slice/crio-4680f8eff52d8a99647cb8795321015c5237f6ba2f4f910d4f337535d92dbba8 WatchSource:0}: Error finding container 4680f8eff52d8a99647cb8795321015c5237f6ba2f4f910d4f337535d92dbba8: Status 404 returned error can't find the container with id 4680f8eff52d8a99647cb8795321015c5237f6ba2f4f910d4f337535d92dbba8 Nov 25 15:19:11 crc kubenswrapper[4890]: I1125 15:19:11.753290 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-lgp8l"] Nov 25 15:19:11 crc kubenswrapper[4890]: I1125 15:19:11.776815 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-w9c2z"] Nov 25 15:19:11 crc kubenswrapper[4890]: I1125 15:19:11.793035 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6f696b9-c9ctc"] Nov 25 15:19:11 crc kubenswrapper[4890]: W1125 15:19:11.796999 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddf163dc8_8261_4fdf_b446_bba9e8e1d8eb.slice/crio-d5aeafc5e22654c2b144daa5cac397a3e4424c28b802dcf6c6bfa7f95aa6ad8f WatchSource:0}: Error finding container d5aeafc5e22654c2b144daa5cac397a3e4424c28b802dcf6c6bfa7f95aa6ad8f: Status 404 returned error can't find the container with id d5aeafc5e22654c2b144daa5cac397a3e4424c28b802dcf6c6bfa7f95aa6ad8f Nov 25 15:19:11 crc kubenswrapper[4890]: I1125 15:19:11.799539 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-v59z7"] Nov 25 15:19:11 crc kubenswrapper[4890]: W1125 15:19:11.804762 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb69893e0_19f6_425b_b36f_b39fd2fd80d5.slice/crio-a64227a45fcc4b277fec4094c85c648a3b8ee686f08e679fbc3e2cad36d62ef3 WatchSource:0}: Error finding container a64227a45fcc4b277fec4094c85c648a3b8ee686f08e679fbc3e2cad36d62ef3: Status 404 returned error can't find the container with id a64227a45fcc4b277fec4094c85c648a3b8ee686f08e679fbc3e2cad36d62ef3 Nov 25 15:19:11 crc kubenswrapper[4890]: I1125 15:19:11.974439 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-v59z7" event={"ID":"f282d61d-d8c2-4933-99c1-d95a9d4f40e2","Type":"ContainerStarted","Data":"b8ccaccec0fb53ba298946f790df29f21ec130e115a114ce61832050e4e13279"} Nov 25 15:19:11 crc kubenswrapper[4890]: I1125 15:19:11.977485 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6f696b9-c9ctc" event={"ID":"1c1e0704-e54b-48b3-bcc3-6af856dfd5ff","Type":"ContainerStarted","Data":"b1d3c82893a0e6bf30870137a704670554e1533d91fd8eff9d571c3b4bb1f740"} Nov 25 15:19:11 crc kubenswrapper[4890]: I1125 15:19:11.979560 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-lgp8l" event={"ID":"df163dc8-8261-4fdf-b446-bba9e8e1d8eb","Type":"ContainerStarted","Data":"d5aeafc5e22654c2b144daa5cac397a3e4424c28b802dcf6c6bfa7f95aa6ad8f"} Nov 25 15:19:11 crc kubenswrapper[4890]: I1125 15:19:11.982821 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"effbba39-45db-4d3b-a883-03f38d0b49f8","Type":"ContainerStarted","Data":"29afb1e723d124d86cfda2680a825900b80a6aa5029589e32a1814cdd2c047a9"} Nov 25 15:19:11 crc kubenswrapper[4890]: I1125 15:19:11.985240 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-4ppjx" event={"ID":"9076492c-2832-41b8-966c-f0ca12a43792","Type":"ContainerStarted","Data":"4680f8eff52d8a99647cb8795321015c5237f6ba2f4f910d4f337535d92dbba8"} Nov 25 15:19:11 crc kubenswrapper[4890]: I1125 15:19:11.986838 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-0806-account-create-xr7r5" event={"ID":"dd3ad34d-4081-4341-b103-20d6ba8da9c5","Type":"ContainerStarted","Data":"5a21a62b5c7af01f6ed9932f831944f2ebd39425c36497b2710c3ec98c04de47"} Nov 25 15:19:11 crc kubenswrapper[4890]: I1125 15:19:11.987008 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-0806-account-create-xr7r5" event={"ID":"dd3ad34d-4081-4341-b103-20d6ba8da9c5","Type":"ContainerStarted","Data":"d878105c2a5f6bc14a01c801636d4e8549ca10a71212ef0a03b561e0e9dca092"} Nov 25 15:19:11 crc kubenswrapper[4890]: I1125 15:19:11.988992 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-vs9dw" event={"ID":"cf34ad11-7dd6-4572-b3d2-4c14127a1d76","Type":"ContainerStarted","Data":"4d2e53b036518a10a8ae51858d830607e13fedd4e8578b62ec853a3e1b0a89c2"} Nov 25 15:19:11 crc kubenswrapper[4890]: I1125 15:19:11.991587 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-w9c2z" event={"ID":"b69893e0-19f6-425b-b36f-b39fd2fd80d5","Type":"ContainerStarted","Data":"a64227a45fcc4b277fec4094c85c648a3b8ee686f08e679fbc3e2cad36d62ef3"} Nov 25 15:19:11 crc kubenswrapper[4890]: I1125 15:19:11.996526 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-aa80-account-create-hcvwx" event={"ID":"3af7c86b-336f-48ce-b5bf-294c0b8bc5ee","Type":"ContainerStarted","Data":"c1ab617d15586118e94112f9a4067c76d9be87979909b2caa010755de5ea09c2"} Nov 25 15:19:12 crc kubenswrapper[4890]: I1125 15:19:12.003204 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-0806-account-create-xr7r5" podStartSLOduration=4.003150334 podStartE2EDuration="4.003150334s" podCreationTimestamp="2025-11-25 15:19:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:19:12.001715398 +0000 UTC m=+1010.444178008" watchObservedRunningTime="2025-11-25 15:19:12.003150334 +0000 UTC m=+1010.445612944" Nov 25 15:19:12 crc kubenswrapper[4890]: I1125 15:19:12.029283 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-vs9dw" podStartSLOduration=4.333978738 podStartE2EDuration="9.029260707s" podCreationTimestamp="2025-11-25 15:19:03 +0000 UTC" firstStartedPulling="2025-11-25 15:19:06.141779121 +0000 UTC m=+1004.584241731" lastFinishedPulling="2025-11-25 15:19:10.83706109 +0000 UTC m=+1009.279523700" observedRunningTime="2025-11-25 15:19:12.024785125 +0000 UTC m=+1010.467247735" watchObservedRunningTime="2025-11-25 15:19:12.029260707 +0000 UTC m=+1010.471723317" Nov 25 15:19:12 crc kubenswrapper[4890]: I1125 15:19:12.184837 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a74f4c1-2950-4305-b91f-50105ed2be79" path="/var/lib/kubelet/pods/2a74f4c1-2950-4305-b91f-50105ed2be79/volumes" Nov 25 15:19:12 crc kubenswrapper[4890]: I1125 15:19:12.185636 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f467b253-fb57-465a-8e67-5168b2dc0e55" path="/var/lib/kubelet/pods/f467b253-fb57-465a-8e67-5168b2dc0e55/volumes" Nov 25 15:19:13 crc kubenswrapper[4890]: I1125 15:19:13.009542 4890 generic.go:334] "Generic (PLEG): container finished" podID="9076492c-2832-41b8-966c-f0ca12a43792" containerID="b0ef80b34e3b82d11514b90037a50d0155f4666970674afded415b4192f4e7bb" exitCode=0 Nov 25 15:19:13 crc kubenswrapper[4890]: I1125 15:19:13.009614 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-4ppjx" event={"ID":"9076492c-2832-41b8-966c-f0ca12a43792","Type":"ContainerDied","Data":"b0ef80b34e3b82d11514b90037a50d0155f4666970674afded415b4192f4e7bb"} Nov 25 15:19:13 crc kubenswrapper[4890]: I1125 15:19:13.011721 4890 generic.go:334] "Generic (PLEG): container finished" podID="dd3ad34d-4081-4341-b103-20d6ba8da9c5" containerID="5a21a62b5c7af01f6ed9932f831944f2ebd39425c36497b2710c3ec98c04de47" exitCode=0 Nov 25 15:19:13 crc kubenswrapper[4890]: I1125 15:19:13.011752 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-0806-account-create-xr7r5" event={"ID":"dd3ad34d-4081-4341-b103-20d6ba8da9c5","Type":"ContainerDied","Data":"5a21a62b5c7af01f6ed9932f831944f2ebd39425c36497b2710c3ec98c04de47"} Nov 25 15:19:13 crc kubenswrapper[4890]: I1125 15:19:13.013453 4890 generic.go:334] "Generic (PLEG): container finished" podID="df163dc8-8261-4fdf-b446-bba9e8e1d8eb" containerID="f8776797ac98bfd2201c355adc063e4b947dcb3f65103a6110c016067ba28acd" exitCode=0 Nov 25 15:19:13 crc kubenswrapper[4890]: I1125 15:19:13.013484 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-lgp8l" event={"ID":"df163dc8-8261-4fdf-b446-bba9e8e1d8eb","Type":"ContainerDied","Data":"f8776797ac98bfd2201c355adc063e4b947dcb3f65103a6110c016067ba28acd"} Nov 25 15:19:13 crc kubenswrapper[4890]: I1125 15:19:13.015493 4890 generic.go:334] "Generic (PLEG): container finished" podID="b69893e0-19f6-425b-b36f-b39fd2fd80d5" containerID="708d473f40456f828711ffe0bd476d3b270c68b158df27ec57944565e5c99d1f" exitCode=0 Nov 25 15:19:13 crc kubenswrapper[4890]: I1125 15:19:13.015547 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-w9c2z" event={"ID":"b69893e0-19f6-425b-b36f-b39fd2fd80d5","Type":"ContainerDied","Data":"708d473f40456f828711ffe0bd476d3b270c68b158df27ec57944565e5c99d1f"} Nov 25 15:19:13 crc kubenswrapper[4890]: I1125 15:19:13.017091 4890 generic.go:334] "Generic (PLEG): container finished" podID="3af7c86b-336f-48ce-b5bf-294c0b8bc5ee" containerID="0e455866be4285b7152b5e4083233f26e58160d23d57cf58741003d8dd3e68ac" exitCode=0 Nov 25 15:19:13 crc kubenswrapper[4890]: I1125 15:19:13.017113 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-aa80-account-create-hcvwx" event={"ID":"3af7c86b-336f-48ce-b5bf-294c0b8bc5ee","Type":"ContainerDied","Data":"0e455866be4285b7152b5e4083233f26e58160d23d57cf58741003d8dd3e68ac"} Nov 25 15:19:13 crc kubenswrapper[4890]: I1125 15:19:13.018211 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-v59z7" event={"ID":"f282d61d-d8c2-4933-99c1-d95a9d4f40e2","Type":"ContainerStarted","Data":"d2b72858e567dc52d3213ed89aa5ab94a2d7395a01608294fb080d5887a3270f"} Nov 25 15:19:13 crc kubenswrapper[4890]: I1125 15:19:13.019462 4890 generic.go:334] "Generic (PLEG): container finished" podID="1c1e0704-e54b-48b3-bcc3-6af856dfd5ff" containerID="d3d32a8939d3efcf680582a6c7f4ad33c857779698cbf795d61a9301d73439c2" exitCode=0 Nov 25 15:19:13 crc kubenswrapper[4890]: I1125 15:19:13.020408 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6f696b9-c9ctc" event={"ID":"1c1e0704-e54b-48b3-bcc3-6af856dfd5ff","Type":"ContainerDied","Data":"d3d32a8939d3efcf680582a6c7f4ad33c857779698cbf795d61a9301d73439c2"} Nov 25 15:19:13 crc kubenswrapper[4890]: I1125 15:19:13.060065 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-v59z7" podStartSLOduration=5.060043428 podStartE2EDuration="5.060043428s" podCreationTimestamp="2025-11-25 15:19:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:19:13.050376036 +0000 UTC m=+1011.492838666" watchObservedRunningTime="2025-11-25 15:19:13.060043428 +0000 UTC m=+1011.502506038" Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.047297 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6f696b9-c9ctc" event={"ID":"1c1e0704-e54b-48b3-bcc3-6af856dfd5ff","Type":"ContainerStarted","Data":"6a98dd24abec302a59e34c75f45a68d7b45738b22aaa348d61ca26a6fdb076a5"} Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.048734 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74f6f696b9-c9ctc" Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.052612 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-lgp8l" event={"ID":"df163dc8-8261-4fdf-b446-bba9e8e1d8eb","Type":"ContainerStarted","Data":"6a49096a7d81634a7d99776fad685186d59a98d07acb82c3cd930e0645292325"} Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.052736 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-lgp8l" Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.056636 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"effbba39-45db-4d3b-a883-03f38d0b49f8","Type":"ContainerStarted","Data":"425c762b3ae786cbfe41f099f99dda8db37c0a757357438792f31e5a1e4e5cb0"} Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.068737 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-74f6f696b9-c9ctc" podStartSLOduration=6.068710727 podStartE2EDuration="6.068710727s" podCreationTimestamp="2025-11-25 15:19:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:19:14.066143612 +0000 UTC m=+1012.508606242" watchObservedRunningTime="2025-11-25 15:19:14.068710727 +0000 UTC m=+1012.511173357" Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.104904 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-lgp8l" podStartSLOduration=6.104881711 podStartE2EDuration="6.104881711s" podCreationTimestamp="2025-11-25 15:19:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:19:14.098400249 +0000 UTC m=+1012.540862879" watchObservedRunningTime="2025-11-25 15:19:14.104881711 +0000 UTC m=+1012.547344331" Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.227112 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-sppf8"] Nov 25 15:19:14 crc kubenswrapper[4890]: E1125 15:19:14.227585 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a74f4c1-2950-4305-b91f-50105ed2be79" containerName="init" Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.227609 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a74f4c1-2950-4305-b91f-50105ed2be79" containerName="init" Nov 25 15:19:14 crc kubenswrapper[4890]: E1125 15:19:14.227627 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f467b253-fb57-465a-8e67-5168b2dc0e55" containerName="init" Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.227633 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="f467b253-fb57-465a-8e67-5168b2dc0e55" containerName="init" Nov 25 15:19:14 crc kubenswrapper[4890]: E1125 15:19:14.227663 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a74f4c1-2950-4305-b91f-50105ed2be79" containerName="dnsmasq-dns" Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.227672 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a74f4c1-2950-4305-b91f-50105ed2be79" containerName="dnsmasq-dns" Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.227839 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a74f4c1-2950-4305-b91f-50105ed2be79" containerName="dnsmasq-dns" Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.227860 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="f467b253-fb57-465a-8e67-5168b2dc0e55" containerName="init" Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.228629 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-sppf8" Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.243623 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/290df6b3-1433-4137-87e7-5caeba2bbb59-operator-scripts\") pod \"glance-db-create-sppf8\" (UID: \"290df6b3-1433-4137-87e7-5caeba2bbb59\") " pod="openstack/glance-db-create-sppf8" Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.243704 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ms6qp\" (UniqueName: \"kubernetes.io/projected/290df6b3-1433-4137-87e7-5caeba2bbb59-kube-api-access-ms6qp\") pod \"glance-db-create-sppf8\" (UID: \"290df6b3-1433-4137-87e7-5caeba2bbb59\") " pod="openstack/glance-db-create-sppf8" Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.248005 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-sppf8"] Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.299546 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-cf3e-account-create-4p2kd"] Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.301046 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-cf3e-account-create-4p2kd" Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.305068 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.311836 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-cf3e-account-create-4p2kd"] Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.344916 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/290df6b3-1433-4137-87e7-5caeba2bbb59-operator-scripts\") pod \"glance-db-create-sppf8\" (UID: \"290df6b3-1433-4137-87e7-5caeba2bbb59\") " pod="openstack/glance-db-create-sppf8" Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.345388 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ms6qp\" (UniqueName: \"kubernetes.io/projected/290df6b3-1433-4137-87e7-5caeba2bbb59-kube-api-access-ms6qp\") pod \"glance-db-create-sppf8\" (UID: \"290df6b3-1433-4137-87e7-5caeba2bbb59\") " pod="openstack/glance-db-create-sppf8" Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.346676 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/290df6b3-1433-4137-87e7-5caeba2bbb59-operator-scripts\") pod \"glance-db-create-sppf8\" (UID: \"290df6b3-1433-4137-87e7-5caeba2bbb59\") " pod="openstack/glance-db-create-sppf8" Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.367693 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ms6qp\" (UniqueName: \"kubernetes.io/projected/290df6b3-1433-4137-87e7-5caeba2bbb59-kube-api-access-ms6qp\") pod \"glance-db-create-sppf8\" (UID: \"290df6b3-1433-4137-87e7-5caeba2bbb59\") " pod="openstack/glance-db-create-sppf8" Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.452862 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b59f5530-3d09-4573-b6f0-beb315e71d79-operator-scripts\") pod \"glance-cf3e-account-create-4p2kd\" (UID: \"b59f5530-3d09-4573-b6f0-beb315e71d79\") " pod="openstack/glance-cf3e-account-create-4p2kd" Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.452999 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xbmk\" (UniqueName: \"kubernetes.io/projected/b59f5530-3d09-4573-b6f0-beb315e71d79-kube-api-access-7xbmk\") pod \"glance-cf3e-account-create-4p2kd\" (UID: \"b59f5530-3d09-4573-b6f0-beb315e71d79\") " pod="openstack/glance-cf3e-account-create-4p2kd" Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.476222 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-aa80-account-create-hcvwx" Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.550581 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-sppf8" Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.566852 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k4gf7\" (UniqueName: \"kubernetes.io/projected/3af7c86b-336f-48ce-b5bf-294c0b8bc5ee-kube-api-access-k4gf7\") pod \"3af7c86b-336f-48ce-b5bf-294c0b8bc5ee\" (UID: \"3af7c86b-336f-48ce-b5bf-294c0b8bc5ee\") " Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.566969 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3af7c86b-336f-48ce-b5bf-294c0b8bc5ee-operator-scripts\") pod \"3af7c86b-336f-48ce-b5bf-294c0b8bc5ee\" (UID: \"3af7c86b-336f-48ce-b5bf-294c0b8bc5ee\") " Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.567287 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xbmk\" (UniqueName: \"kubernetes.io/projected/b59f5530-3d09-4573-b6f0-beb315e71d79-kube-api-access-7xbmk\") pod \"glance-cf3e-account-create-4p2kd\" (UID: \"b59f5530-3d09-4573-b6f0-beb315e71d79\") " pod="openstack/glance-cf3e-account-create-4p2kd" Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.567400 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b59f5530-3d09-4573-b6f0-beb315e71d79-operator-scripts\") pod \"glance-cf3e-account-create-4p2kd\" (UID: \"b59f5530-3d09-4573-b6f0-beb315e71d79\") " pod="openstack/glance-cf3e-account-create-4p2kd" Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.568229 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3af7c86b-336f-48ce-b5bf-294c0b8bc5ee-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3af7c86b-336f-48ce-b5bf-294c0b8bc5ee" (UID: "3af7c86b-336f-48ce-b5bf-294c0b8bc5ee"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.568316 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b59f5530-3d09-4573-b6f0-beb315e71d79-operator-scripts\") pod \"glance-cf3e-account-create-4p2kd\" (UID: \"b59f5530-3d09-4573-b6f0-beb315e71d79\") " pod="openstack/glance-cf3e-account-create-4p2kd" Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.574432 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3af7c86b-336f-48ce-b5bf-294c0b8bc5ee-kube-api-access-k4gf7" (OuterVolumeSpecName: "kube-api-access-k4gf7") pod "3af7c86b-336f-48ce-b5bf-294c0b8bc5ee" (UID: "3af7c86b-336f-48ce-b5bf-294c0b8bc5ee"). InnerVolumeSpecName "kube-api-access-k4gf7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.590231 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xbmk\" (UniqueName: \"kubernetes.io/projected/b59f5530-3d09-4573-b6f0-beb315e71d79-kube-api-access-7xbmk\") pod \"glance-cf3e-account-create-4p2kd\" (UID: \"b59f5530-3d09-4573-b6f0-beb315e71d79\") " pod="openstack/glance-cf3e-account-create-4p2kd" Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.629099 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-cf3e-account-create-4p2kd" Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.670511 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k4gf7\" (UniqueName: \"kubernetes.io/projected/3af7c86b-336f-48ce-b5bf-294c0b8bc5ee-kube-api-access-k4gf7\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.670567 4890 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3af7c86b-336f-48ce-b5bf-294c0b8bc5ee-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.795194 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-w9c2z" Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.821733 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-0806-account-create-xr7r5" Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.832225 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-4ppjx" Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.976763 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cmp9s\" (UniqueName: \"kubernetes.io/projected/9076492c-2832-41b8-966c-f0ca12a43792-kube-api-access-cmp9s\") pod \"9076492c-2832-41b8-966c-f0ca12a43792\" (UID: \"9076492c-2832-41b8-966c-f0ca12a43792\") " Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.976918 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b69893e0-19f6-425b-b36f-b39fd2fd80d5-operator-scripts\") pod \"b69893e0-19f6-425b-b36f-b39fd2fd80d5\" (UID: \"b69893e0-19f6-425b-b36f-b39fd2fd80d5\") " Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.976957 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9076492c-2832-41b8-966c-f0ca12a43792-operator-scripts\") pod \"9076492c-2832-41b8-966c-f0ca12a43792\" (UID: \"9076492c-2832-41b8-966c-f0ca12a43792\") " Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.977015 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-whgj5\" (UniqueName: \"kubernetes.io/projected/dd3ad34d-4081-4341-b103-20d6ba8da9c5-kube-api-access-whgj5\") pod \"dd3ad34d-4081-4341-b103-20d6ba8da9c5\" (UID: \"dd3ad34d-4081-4341-b103-20d6ba8da9c5\") " Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.977040 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd3ad34d-4081-4341-b103-20d6ba8da9c5-operator-scripts\") pod \"dd3ad34d-4081-4341-b103-20d6ba8da9c5\" (UID: \"dd3ad34d-4081-4341-b103-20d6ba8da9c5\") " Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.977059 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvsph\" (UniqueName: \"kubernetes.io/projected/b69893e0-19f6-425b-b36f-b39fd2fd80d5-kube-api-access-vvsph\") pod \"b69893e0-19f6-425b-b36f-b39fd2fd80d5\" (UID: \"b69893e0-19f6-425b-b36f-b39fd2fd80d5\") " Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.977637 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9076492c-2832-41b8-966c-f0ca12a43792-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9076492c-2832-41b8-966c-f0ca12a43792" (UID: "9076492c-2832-41b8-966c-f0ca12a43792"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.977737 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b69893e0-19f6-425b-b36f-b39fd2fd80d5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b69893e0-19f6-425b-b36f-b39fd2fd80d5" (UID: "b69893e0-19f6-425b-b36f-b39fd2fd80d5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.977889 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd3ad34d-4081-4341-b103-20d6ba8da9c5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dd3ad34d-4081-4341-b103-20d6ba8da9c5" (UID: "dd3ad34d-4081-4341-b103-20d6ba8da9c5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.982950 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd3ad34d-4081-4341-b103-20d6ba8da9c5-kube-api-access-whgj5" (OuterVolumeSpecName: "kube-api-access-whgj5") pod "dd3ad34d-4081-4341-b103-20d6ba8da9c5" (UID: "dd3ad34d-4081-4341-b103-20d6ba8da9c5"). InnerVolumeSpecName "kube-api-access-whgj5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:19:14 crc kubenswrapper[4890]: I1125 15:19:14.983649 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b69893e0-19f6-425b-b36f-b39fd2fd80d5-kube-api-access-vvsph" (OuterVolumeSpecName: "kube-api-access-vvsph") pod "b69893e0-19f6-425b-b36f-b39fd2fd80d5" (UID: "b69893e0-19f6-425b-b36f-b39fd2fd80d5"). InnerVolumeSpecName "kube-api-access-vvsph". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:19:15 crc kubenswrapper[4890]: I1125 15:19:15.001543 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9076492c-2832-41b8-966c-f0ca12a43792-kube-api-access-cmp9s" (OuterVolumeSpecName: "kube-api-access-cmp9s") pod "9076492c-2832-41b8-966c-f0ca12a43792" (UID: "9076492c-2832-41b8-966c-f0ca12a43792"). InnerVolumeSpecName "kube-api-access-cmp9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:19:15 crc kubenswrapper[4890]: I1125 15:19:15.012284 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-sppf8"] Nov 25 15:19:15 crc kubenswrapper[4890]: I1125 15:19:15.068739 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-aa80-account-create-hcvwx" Nov 25 15:19:15 crc kubenswrapper[4890]: I1125 15:19:15.068729 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-aa80-account-create-hcvwx" event={"ID":"3af7c86b-336f-48ce-b5bf-294c0b8bc5ee","Type":"ContainerDied","Data":"c1ab617d15586118e94112f9a4067c76d9be87979909b2caa010755de5ea09c2"} Nov 25 15:19:15 crc kubenswrapper[4890]: I1125 15:19:15.068860 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c1ab617d15586118e94112f9a4067c76d9be87979909b2caa010755de5ea09c2" Nov 25 15:19:15 crc kubenswrapper[4890]: I1125 15:19:15.070775 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"effbba39-45db-4d3b-a883-03f38d0b49f8","Type":"ContainerStarted","Data":"b10f9ccb71d8e4256be4e48a7ee8fed9ad8092a3b9243738ca8877923860817b"} Nov 25 15:19:15 crc kubenswrapper[4890]: I1125 15:19:15.070902 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Nov 25 15:19:15 crc kubenswrapper[4890]: I1125 15:19:15.072364 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-4ppjx" event={"ID":"9076492c-2832-41b8-966c-f0ca12a43792","Type":"ContainerDied","Data":"4680f8eff52d8a99647cb8795321015c5237f6ba2f4f910d4f337535d92dbba8"} Nov 25 15:19:15 crc kubenswrapper[4890]: I1125 15:19:15.072386 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4680f8eff52d8a99647cb8795321015c5237f6ba2f4f910d4f337535d92dbba8" Nov 25 15:19:15 crc kubenswrapper[4890]: I1125 15:19:15.072421 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-4ppjx" Nov 25 15:19:15 crc kubenswrapper[4890]: I1125 15:19:15.075028 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-0806-account-create-xr7r5" Nov 25 15:19:15 crc kubenswrapper[4890]: I1125 15:19:15.075007 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-0806-account-create-xr7r5" event={"ID":"dd3ad34d-4081-4341-b103-20d6ba8da9c5","Type":"ContainerDied","Data":"d878105c2a5f6bc14a01c801636d4e8549ca10a71212ef0a03b561e0e9dca092"} Nov 25 15:19:15 crc kubenswrapper[4890]: I1125 15:19:15.075069 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d878105c2a5f6bc14a01c801636d4e8549ca10a71212ef0a03b561e0e9dca092" Nov 25 15:19:15 crc kubenswrapper[4890]: I1125 15:19:15.076365 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-sppf8" event={"ID":"290df6b3-1433-4137-87e7-5caeba2bbb59","Type":"ContainerStarted","Data":"75420806b2db3463f887f51a05a015c28e2614ddb1a2792fc1ee29e7f380f299"} Nov 25 15:19:15 crc kubenswrapper[4890]: I1125 15:19:15.078402 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cmp9s\" (UniqueName: \"kubernetes.io/projected/9076492c-2832-41b8-966c-f0ca12a43792-kube-api-access-cmp9s\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:15 crc kubenswrapper[4890]: I1125 15:19:15.078419 4890 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b69893e0-19f6-425b-b36f-b39fd2fd80d5-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:15 crc kubenswrapper[4890]: I1125 15:19:15.078428 4890 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9076492c-2832-41b8-966c-f0ca12a43792-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:15 crc kubenswrapper[4890]: I1125 15:19:15.078437 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-whgj5\" (UniqueName: \"kubernetes.io/projected/dd3ad34d-4081-4341-b103-20d6ba8da9c5-kube-api-access-whgj5\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:15 crc kubenswrapper[4890]: I1125 15:19:15.078445 4890 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd3ad34d-4081-4341-b103-20d6ba8da9c5-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:15 crc kubenswrapper[4890]: I1125 15:19:15.078454 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vvsph\" (UniqueName: \"kubernetes.io/projected/b69893e0-19f6-425b-b36f-b39fd2fd80d5-kube-api-access-vvsph\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:15 crc kubenswrapper[4890]: I1125 15:19:15.079010 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-w9c2z" Nov 25 15:19:15 crc kubenswrapper[4890]: I1125 15:19:15.082525 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-w9c2z" event={"ID":"b69893e0-19f6-425b-b36f-b39fd2fd80d5","Type":"ContainerDied","Data":"a64227a45fcc4b277fec4094c85c648a3b8ee686f08e679fbc3e2cad36d62ef3"} Nov 25 15:19:15 crc kubenswrapper[4890]: I1125 15:19:15.082639 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a64227a45fcc4b277fec4094c85c648a3b8ee686f08e679fbc3e2cad36d62ef3" Nov 25 15:19:15 crc kubenswrapper[4890]: I1125 15:19:15.093646 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=4.953084385 podStartE2EDuration="7.093625432s" podCreationTimestamp="2025-11-25 15:19:08 +0000 UTC" firstStartedPulling="2025-11-25 15:19:11.518347017 +0000 UTC m=+1009.960809627" lastFinishedPulling="2025-11-25 15:19:13.658888064 +0000 UTC m=+1012.101350674" observedRunningTime="2025-11-25 15:19:15.087241123 +0000 UTC m=+1013.529703743" watchObservedRunningTime="2025-11-25 15:19:15.093625432 +0000 UTC m=+1013.536088042" Nov 25 15:19:15 crc kubenswrapper[4890]: I1125 15:19:15.174576 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-cf3e-account-create-4p2kd"] Nov 25 15:19:15 crc kubenswrapper[4890]: W1125 15:19:15.175924 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb59f5530_3d09_4573_b6f0_beb315e71d79.slice/crio-3e1e9c172609ba3e3cd4d962326761f5f3ec1a38b3601981ef63824853d9fc66 WatchSource:0}: Error finding container 3e1e9c172609ba3e3cd4d962326761f5f3ec1a38b3601981ef63824853d9fc66: Status 404 returned error can't find the container with id 3e1e9c172609ba3e3cd4d962326761f5f3ec1a38b3601981ef63824853d9fc66 Nov 25 15:19:15 crc kubenswrapper[4890]: I1125 15:19:15.493384 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 25 15:19:15 crc kubenswrapper[4890]: I1125 15:19:15.828006 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-xtxvs"] Nov 25 15:19:15 crc kubenswrapper[4890]: E1125 15:19:15.828738 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9076492c-2832-41b8-966c-f0ca12a43792" containerName="mariadb-database-create" Nov 25 15:19:15 crc kubenswrapper[4890]: I1125 15:19:15.828758 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="9076492c-2832-41b8-966c-f0ca12a43792" containerName="mariadb-database-create" Nov 25 15:19:15 crc kubenswrapper[4890]: E1125 15:19:15.828771 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd3ad34d-4081-4341-b103-20d6ba8da9c5" containerName="mariadb-account-create" Nov 25 15:19:15 crc kubenswrapper[4890]: I1125 15:19:15.828777 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd3ad34d-4081-4341-b103-20d6ba8da9c5" containerName="mariadb-account-create" Nov 25 15:19:15 crc kubenswrapper[4890]: E1125 15:19:15.828789 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b69893e0-19f6-425b-b36f-b39fd2fd80d5" containerName="mariadb-database-create" Nov 25 15:19:15 crc kubenswrapper[4890]: I1125 15:19:15.828796 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="b69893e0-19f6-425b-b36f-b39fd2fd80d5" containerName="mariadb-database-create" Nov 25 15:19:15 crc kubenswrapper[4890]: E1125 15:19:15.828803 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3af7c86b-336f-48ce-b5bf-294c0b8bc5ee" containerName="mariadb-account-create" Nov 25 15:19:15 crc kubenswrapper[4890]: I1125 15:19:15.828809 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="3af7c86b-336f-48ce-b5bf-294c0b8bc5ee" containerName="mariadb-account-create" Nov 25 15:19:15 crc kubenswrapper[4890]: I1125 15:19:15.828981 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd3ad34d-4081-4341-b103-20d6ba8da9c5" containerName="mariadb-account-create" Nov 25 15:19:15 crc kubenswrapper[4890]: I1125 15:19:15.829015 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="b69893e0-19f6-425b-b36f-b39fd2fd80d5" containerName="mariadb-database-create" Nov 25 15:19:15 crc kubenswrapper[4890]: I1125 15:19:15.829033 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="3af7c86b-336f-48ce-b5bf-294c0b8bc5ee" containerName="mariadb-account-create" Nov 25 15:19:15 crc kubenswrapper[4890]: I1125 15:19:15.829044 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="9076492c-2832-41b8-966c-f0ca12a43792" containerName="mariadb-database-create" Nov 25 15:19:15 crc kubenswrapper[4890]: I1125 15:19:15.829721 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-xtxvs" Nov 25 15:19:15 crc kubenswrapper[4890]: I1125 15:19:15.854686 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-xtxvs"] Nov 25 15:19:15 crc kubenswrapper[4890]: I1125 15:19:15.865075 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-9354-account-create-5fdrf"] Nov 25 15:19:15 crc kubenswrapper[4890]: I1125 15:19:15.866516 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-9354-account-create-5fdrf" Nov 25 15:19:15 crc kubenswrapper[4890]: I1125 15:19:15.868741 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Nov 25 15:19:15 crc kubenswrapper[4890]: I1125 15:19:15.893692 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-9354-account-create-5fdrf"] Nov 25 15:19:15 crc kubenswrapper[4890]: I1125 15:19:15.937322 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-hc7pc"] Nov 25 15:19:15 crc kubenswrapper[4890]: I1125 15:19:15.938743 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-hc7pc" Nov 25 15:19:15 crc kubenswrapper[4890]: I1125 15:19:15.951243 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-hc7pc"] Nov 25 15:19:15 crc kubenswrapper[4890]: I1125 15:19:15.999432 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j49nl\" (UniqueName: \"kubernetes.io/projected/acfb093b-f3af-4856-94b8-64a43c9c0cc1-kube-api-access-j49nl\") pod \"cinder-db-create-xtxvs\" (UID: \"acfb093b-f3af-4856-94b8-64a43c9c0cc1\") " pod="openstack/cinder-db-create-xtxvs" Nov 25 15:19:15 crc kubenswrapper[4890]: I1125 15:19:15.999604 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/acfb093b-f3af-4856-94b8-64a43c9c0cc1-operator-scripts\") pod \"cinder-db-create-xtxvs\" (UID: \"acfb093b-f3af-4856-94b8-64a43c9c0cc1\") " pod="openstack/cinder-db-create-xtxvs" Nov 25 15:19:15 crc kubenswrapper[4890]: I1125 15:19:15.999701 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6z4w\" (UniqueName: \"kubernetes.io/projected/a8cd74cd-df01-4e7b-a8a9-391d5afbe2f9-kube-api-access-j6z4w\") pod \"cinder-9354-account-create-5fdrf\" (UID: \"a8cd74cd-df01-4e7b-a8a9-391d5afbe2f9\") " pod="openstack/cinder-9354-account-create-5fdrf" Nov 25 15:19:15 crc kubenswrapper[4890]: I1125 15:19:15.999792 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8cd74cd-df01-4e7b-a8a9-391d5afbe2f9-operator-scripts\") pod \"cinder-9354-account-create-5fdrf\" (UID: \"a8cd74cd-df01-4e7b-a8a9-391d5afbe2f9\") " pod="openstack/cinder-9354-account-create-5fdrf" Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.087404 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-cf3e-account-create-4p2kd" event={"ID":"b59f5530-3d09-4573-b6f0-beb315e71d79","Type":"ContainerStarted","Data":"e893380f6bc195a09503a2f355f95510799db8ab623f49a44eedfa4e6b040962"} Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.087507 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-cf3e-account-create-4p2kd" event={"ID":"b59f5530-3d09-4573-b6f0-beb315e71d79","Type":"ContainerStarted","Data":"3e1e9c172609ba3e3cd4d962326761f5f3ec1a38b3601981ef63824853d9fc66"} Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.089599 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-sppf8" event={"ID":"290df6b3-1433-4137-87e7-5caeba2bbb59","Type":"ContainerStarted","Data":"57d487aa9b7d58ac42a98c9c84936083cef357fa3420580c62c2769103dcd5b7"} Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.101706 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j49nl\" (UniqueName: \"kubernetes.io/projected/acfb093b-f3af-4856-94b8-64a43c9c0cc1-kube-api-access-j49nl\") pod \"cinder-db-create-xtxvs\" (UID: \"acfb093b-f3af-4856-94b8-64a43c9c0cc1\") " pod="openstack/cinder-db-create-xtxvs" Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.101809 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/acfb093b-f3af-4856-94b8-64a43c9c0cc1-operator-scripts\") pod \"cinder-db-create-xtxvs\" (UID: \"acfb093b-f3af-4856-94b8-64a43c9c0cc1\") " pod="openstack/cinder-db-create-xtxvs" Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.101852 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6z4w\" (UniqueName: \"kubernetes.io/projected/a8cd74cd-df01-4e7b-a8a9-391d5afbe2f9-kube-api-access-j6z4w\") pod \"cinder-9354-account-create-5fdrf\" (UID: \"a8cd74cd-df01-4e7b-a8a9-391d5afbe2f9\") " pod="openstack/cinder-9354-account-create-5fdrf" Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.101903 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hm8hb\" (UniqueName: \"kubernetes.io/projected/2c14ab22-4e5a-43e6-be42-956608ab877b-kube-api-access-hm8hb\") pod \"barbican-db-create-hc7pc\" (UID: \"2c14ab22-4e5a-43e6-be42-956608ab877b\") " pod="openstack/barbican-db-create-hc7pc" Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.101923 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8cd74cd-df01-4e7b-a8a9-391d5afbe2f9-operator-scripts\") pod \"cinder-9354-account-create-5fdrf\" (UID: \"a8cd74cd-df01-4e7b-a8a9-391d5afbe2f9\") " pod="openstack/cinder-9354-account-create-5fdrf" Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.101946 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2c14ab22-4e5a-43e6-be42-956608ab877b-operator-scripts\") pod \"barbican-db-create-hc7pc\" (UID: \"2c14ab22-4e5a-43e6-be42-956608ab877b\") " pod="openstack/barbican-db-create-hc7pc" Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.102588 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8cd74cd-df01-4e7b-a8a9-391d5afbe2f9-operator-scripts\") pod \"cinder-9354-account-create-5fdrf\" (UID: \"a8cd74cd-df01-4e7b-a8a9-391d5afbe2f9\") " pod="openstack/cinder-9354-account-create-5fdrf" Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.102895 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/acfb093b-f3af-4856-94b8-64a43c9c0cc1-operator-scripts\") pod \"cinder-db-create-xtxvs\" (UID: \"acfb093b-f3af-4856-94b8-64a43c9c0cc1\") " pod="openstack/cinder-db-create-xtxvs" Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.110449 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-cf3e-account-create-4p2kd" podStartSLOduration=2.110427044 podStartE2EDuration="2.110427044s" podCreationTimestamp="2025-11-25 15:19:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:19:16.105091431 +0000 UTC m=+1014.547554031" watchObservedRunningTime="2025-11-25 15:19:16.110427044 +0000 UTC m=+1014.552889654" Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.125125 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6z4w\" (UniqueName: \"kubernetes.io/projected/a8cd74cd-df01-4e7b-a8a9-391d5afbe2f9-kube-api-access-j6z4w\") pod \"cinder-9354-account-create-5fdrf\" (UID: \"a8cd74cd-df01-4e7b-a8a9-391d5afbe2f9\") " pod="openstack/cinder-9354-account-create-5fdrf" Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.136032 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j49nl\" (UniqueName: \"kubernetes.io/projected/acfb093b-f3af-4856-94b8-64a43c9c0cc1-kube-api-access-j49nl\") pod \"cinder-db-create-xtxvs\" (UID: \"acfb093b-f3af-4856-94b8-64a43c9c0cc1\") " pod="openstack/cinder-db-create-xtxvs" Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.136948 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-aca8-account-create-gm6cf"] Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.138249 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-aca8-account-create-gm6cf" Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.140580 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.160895 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-xtxvs" Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.161643 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-sppf8" podStartSLOduration=2.161600563 podStartE2EDuration="2.161600563s" podCreationTimestamp="2025-11-25 15:19:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:19:16.124439894 +0000 UTC m=+1014.566902504" watchObservedRunningTime="2025-11-25 15:19:16.161600563 +0000 UTC m=+1014.604063173" Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.188610 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-9354-account-create-5fdrf" Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.204659 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hm8hb\" (UniqueName: \"kubernetes.io/projected/2c14ab22-4e5a-43e6-be42-956608ab877b-kube-api-access-hm8hb\") pod \"barbican-db-create-hc7pc\" (UID: \"2c14ab22-4e5a-43e6-be42-956608ab877b\") " pod="openstack/barbican-db-create-hc7pc" Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.204754 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2c14ab22-4e5a-43e6-be42-956608ab877b-operator-scripts\") pod \"barbican-db-create-hc7pc\" (UID: \"2c14ab22-4e5a-43e6-be42-956608ab877b\") " pod="openstack/barbican-db-create-hc7pc" Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.208375 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-aca8-account-create-gm6cf"] Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.209450 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2c14ab22-4e5a-43e6-be42-956608ab877b-operator-scripts\") pod \"barbican-db-create-hc7pc\" (UID: \"2c14ab22-4e5a-43e6-be42-956608ab877b\") " pod="openstack/barbican-db-create-hc7pc" Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.253849 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hm8hb\" (UniqueName: \"kubernetes.io/projected/2c14ab22-4e5a-43e6-be42-956608ab877b-kube-api-access-hm8hb\") pod \"barbican-db-create-hc7pc\" (UID: \"2c14ab22-4e5a-43e6-be42-956608ab877b\") " pod="openstack/barbican-db-create-hc7pc" Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.266134 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-hc7pc" Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.276707 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-9r7gm"] Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.277884 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-9r7gm" Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.297481 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-9r7gm"] Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.306366 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c5d95f8e-951b-402b-afe8-ec8b3d090a45-operator-scripts\") pod \"barbican-aca8-account-create-gm6cf\" (UID: \"c5d95f8e-951b-402b-afe8-ec8b3d090a45\") " pod="openstack/barbican-aca8-account-create-gm6cf" Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.306418 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgg62\" (UniqueName: \"kubernetes.io/projected/c5d95f8e-951b-402b-afe8-ec8b3d090a45-kube-api-access-bgg62\") pod \"barbican-aca8-account-create-gm6cf\" (UID: \"c5d95f8e-951b-402b-afe8-ec8b3d090a45\") " pod="openstack/barbican-aca8-account-create-gm6cf" Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.407731 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c5d95f8e-951b-402b-afe8-ec8b3d090a45-operator-scripts\") pod \"barbican-aca8-account-create-gm6cf\" (UID: \"c5d95f8e-951b-402b-afe8-ec8b3d090a45\") " pod="openstack/barbican-aca8-account-create-gm6cf" Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.408143 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgg62\" (UniqueName: \"kubernetes.io/projected/c5d95f8e-951b-402b-afe8-ec8b3d090a45-kube-api-access-bgg62\") pod \"barbican-aca8-account-create-gm6cf\" (UID: \"c5d95f8e-951b-402b-afe8-ec8b3d090a45\") " pod="openstack/barbican-aca8-account-create-gm6cf" Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.408203 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8rzs\" (UniqueName: \"kubernetes.io/projected/2d380d3e-4b4c-424b-b9af-fb34e959e81c-kube-api-access-f8rzs\") pod \"neutron-db-create-9r7gm\" (UID: \"2d380d3e-4b4c-424b-b9af-fb34e959e81c\") " pod="openstack/neutron-db-create-9r7gm" Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.408243 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d380d3e-4b4c-424b-b9af-fb34e959e81c-operator-scripts\") pod \"neutron-db-create-9r7gm\" (UID: \"2d380d3e-4b4c-424b-b9af-fb34e959e81c\") " pod="openstack/neutron-db-create-9r7gm" Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.409140 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c5d95f8e-951b-402b-afe8-ec8b3d090a45-operator-scripts\") pod \"barbican-aca8-account-create-gm6cf\" (UID: \"c5d95f8e-951b-402b-afe8-ec8b3d090a45\") " pod="openstack/barbican-aca8-account-create-gm6cf" Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.449527 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgg62\" (UniqueName: \"kubernetes.io/projected/c5d95f8e-951b-402b-afe8-ec8b3d090a45-kube-api-access-bgg62\") pod \"barbican-aca8-account-create-gm6cf\" (UID: \"c5d95f8e-951b-402b-afe8-ec8b3d090a45\") " pod="openstack/barbican-aca8-account-create-gm6cf" Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.453629 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-90a3-account-create-qvq8c"] Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.455678 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-90a3-account-create-qvq8c" Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.464946 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.465376 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-90a3-account-create-qvq8c"] Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.512617 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8rzs\" (UniqueName: \"kubernetes.io/projected/2d380d3e-4b4c-424b-b9af-fb34e959e81c-kube-api-access-f8rzs\") pod \"neutron-db-create-9r7gm\" (UID: \"2d380d3e-4b4c-424b-b9af-fb34e959e81c\") " pod="openstack/neutron-db-create-9r7gm" Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.512696 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d380d3e-4b4c-424b-b9af-fb34e959e81c-operator-scripts\") pod \"neutron-db-create-9r7gm\" (UID: \"2d380d3e-4b4c-424b-b9af-fb34e959e81c\") " pod="openstack/neutron-db-create-9r7gm" Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.518427 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d380d3e-4b4c-424b-b9af-fb34e959e81c-operator-scripts\") pod \"neutron-db-create-9r7gm\" (UID: \"2d380d3e-4b4c-424b-b9af-fb34e959e81c\") " pod="openstack/neutron-db-create-9r7gm" Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.532655 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8rzs\" (UniqueName: \"kubernetes.io/projected/2d380d3e-4b4c-424b-b9af-fb34e959e81c-kube-api-access-f8rzs\") pod \"neutron-db-create-9r7gm\" (UID: \"2d380d3e-4b4c-424b-b9af-fb34e959e81c\") " pod="openstack/neutron-db-create-9r7gm" Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.595796 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-aca8-account-create-gm6cf" Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.607808 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-9r7gm" Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.615655 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrds8\" (UniqueName: \"kubernetes.io/projected/aba9e84f-bda0-4c6f-883e-ea9106b2521f-kube-api-access-lrds8\") pod \"neutron-90a3-account-create-qvq8c\" (UID: \"aba9e84f-bda0-4c6f-883e-ea9106b2521f\") " pod="openstack/neutron-90a3-account-create-qvq8c" Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.615876 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aba9e84f-bda0-4c6f-883e-ea9106b2521f-operator-scripts\") pod \"neutron-90a3-account-create-qvq8c\" (UID: \"aba9e84f-bda0-4c6f-883e-ea9106b2521f\") " pod="openstack/neutron-90a3-account-create-qvq8c" Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.717773 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrds8\" (UniqueName: \"kubernetes.io/projected/aba9e84f-bda0-4c6f-883e-ea9106b2521f-kube-api-access-lrds8\") pod \"neutron-90a3-account-create-qvq8c\" (UID: \"aba9e84f-bda0-4c6f-883e-ea9106b2521f\") " pod="openstack/neutron-90a3-account-create-qvq8c" Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.717896 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aba9e84f-bda0-4c6f-883e-ea9106b2521f-operator-scripts\") pod \"neutron-90a3-account-create-qvq8c\" (UID: \"aba9e84f-bda0-4c6f-883e-ea9106b2521f\") " pod="openstack/neutron-90a3-account-create-qvq8c" Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.719029 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aba9e84f-bda0-4c6f-883e-ea9106b2521f-operator-scripts\") pod \"neutron-90a3-account-create-qvq8c\" (UID: \"aba9e84f-bda0-4c6f-883e-ea9106b2521f\") " pod="openstack/neutron-90a3-account-create-qvq8c" Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.732287 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-9354-account-create-5fdrf"] Nov 25 15:19:16 crc kubenswrapper[4890]: W1125 15:19:16.751512 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda8cd74cd_df01_4e7b_a8a9_391d5afbe2f9.slice/crio-387d8aa9eacd58d49e790db4adc50064e6b7d77f8613ed55a238e180fe1ca6e6 WatchSource:0}: Error finding container 387d8aa9eacd58d49e790db4adc50064e6b7d77f8613ed55a238e180fe1ca6e6: Status 404 returned error can't find the container with id 387d8aa9eacd58d49e790db4adc50064e6b7d77f8613ed55a238e180fe1ca6e6 Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.752501 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrds8\" (UniqueName: \"kubernetes.io/projected/aba9e84f-bda0-4c6f-883e-ea9106b2521f-kube-api-access-lrds8\") pod \"neutron-90a3-account-create-qvq8c\" (UID: \"aba9e84f-bda0-4c6f-883e-ea9106b2521f\") " pod="openstack/neutron-90a3-account-create-qvq8c" Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.807548 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-90a3-account-create-qvq8c" Nov 25 15:19:16 crc kubenswrapper[4890]: I1125 15:19:16.826035 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-xtxvs"] Nov 25 15:19:16 crc kubenswrapper[4890]: W1125 15:19:16.874331 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podacfb093b_f3af_4856_94b8_64a43c9c0cc1.slice/crio-1a39441655b4cea76969e1e7fd436cf2ba0aafde43dc97bc0c566b83bbfdcc58 WatchSource:0}: Error finding container 1a39441655b4cea76969e1e7fd436cf2ba0aafde43dc97bc0c566b83bbfdcc58: Status 404 returned error can't find the container with id 1a39441655b4cea76969e1e7fd436cf2ba0aafde43dc97bc0c566b83bbfdcc58 Nov 25 15:19:17 crc kubenswrapper[4890]: I1125 15:19:17.083563 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-hc7pc"] Nov 25 15:19:17 crc kubenswrapper[4890]: I1125 15:19:17.129770 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-9354-account-create-5fdrf" event={"ID":"a8cd74cd-df01-4e7b-a8a9-391d5afbe2f9","Type":"ContainerStarted","Data":"387d8aa9eacd58d49e790db4adc50064e6b7d77f8613ed55a238e180fe1ca6e6"} Nov 25 15:19:17 crc kubenswrapper[4890]: I1125 15:19:17.133663 4890 generic.go:334] "Generic (PLEG): container finished" podID="290df6b3-1433-4137-87e7-5caeba2bbb59" containerID="57d487aa9b7d58ac42a98c9c84936083cef357fa3420580c62c2769103dcd5b7" exitCode=0 Nov 25 15:19:17 crc kubenswrapper[4890]: I1125 15:19:17.133799 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-sppf8" event={"ID":"290df6b3-1433-4137-87e7-5caeba2bbb59","Type":"ContainerDied","Data":"57d487aa9b7d58ac42a98c9c84936083cef357fa3420580c62c2769103dcd5b7"} Nov 25 15:19:17 crc kubenswrapper[4890]: I1125 15:19:17.156357 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-xtxvs" event={"ID":"acfb093b-f3af-4856-94b8-64a43c9c0cc1","Type":"ContainerStarted","Data":"1a39441655b4cea76969e1e7fd436cf2ba0aafde43dc97bc0c566b83bbfdcc58"} Nov 25 15:19:17 crc kubenswrapper[4890]: W1125 15:19:17.385820 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc5d95f8e_951b_402b_afe8_ec8b3d090a45.slice/crio-9ac7f25feb18fe29469f442008a65d583284315ed9188e0aec5e6924d660847b WatchSource:0}: Error finding container 9ac7f25feb18fe29469f442008a65d583284315ed9188e0aec5e6924d660847b: Status 404 returned error can't find the container with id 9ac7f25feb18fe29469f442008a65d583284315ed9188e0aec5e6924d660847b Nov 25 15:19:17 crc kubenswrapper[4890]: I1125 15:19:17.386840 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-aca8-account-create-gm6cf"] Nov 25 15:19:17 crc kubenswrapper[4890]: W1125 15:19:17.392879 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2d380d3e_4b4c_424b_b9af_fb34e959e81c.slice/crio-851e818154b0438d38dde98332ccde96f27e8fe5221271c46a649c14532a61e0 WatchSource:0}: Error finding container 851e818154b0438d38dde98332ccde96f27e8fe5221271c46a649c14532a61e0: Status 404 returned error can't find the container with id 851e818154b0438d38dde98332ccde96f27e8fe5221271c46a649c14532a61e0 Nov 25 15:19:17 crc kubenswrapper[4890]: I1125 15:19:17.401069 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-9r7gm"] Nov 25 15:19:17 crc kubenswrapper[4890]: I1125 15:19:17.563625 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-90a3-account-create-qvq8c"] Nov 25 15:19:18 crc kubenswrapper[4890]: I1125 15:19:18.165082 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-9r7gm" event={"ID":"2d380d3e-4b4c-424b-b9af-fb34e959e81c","Type":"ContainerStarted","Data":"3004fbf5a7a3bea35404dc2c53d85fe29655581adbfbcf65d2feead1f3d308ec"} Nov 25 15:19:18 crc kubenswrapper[4890]: I1125 15:19:18.165460 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-9r7gm" event={"ID":"2d380d3e-4b4c-424b-b9af-fb34e959e81c","Type":"ContainerStarted","Data":"851e818154b0438d38dde98332ccde96f27e8fe5221271c46a649c14532a61e0"} Nov 25 15:19:18 crc kubenswrapper[4890]: I1125 15:19:18.166721 4890 generic.go:334] "Generic (PLEG): container finished" podID="a8cd74cd-df01-4e7b-a8a9-391d5afbe2f9" containerID="5e84c82890d994cc8442e7f7ca92e33dce94970d4afefe933ecc2285c4373eb4" exitCode=0 Nov 25 15:19:18 crc kubenswrapper[4890]: I1125 15:19:18.166805 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-9354-account-create-5fdrf" event={"ID":"a8cd74cd-df01-4e7b-a8a9-391d5afbe2f9","Type":"ContainerDied","Data":"5e84c82890d994cc8442e7f7ca92e33dce94970d4afefe933ecc2285c4373eb4"} Nov 25 15:19:18 crc kubenswrapper[4890]: I1125 15:19:18.169231 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-90a3-account-create-qvq8c" event={"ID":"aba9e84f-bda0-4c6f-883e-ea9106b2521f","Type":"ContainerStarted","Data":"83d8df7f74e56f66d558ce883b578ee8d08a0a14374ce74454a0337e40f72975"} Nov 25 15:19:18 crc kubenswrapper[4890]: I1125 15:19:18.169266 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-90a3-account-create-qvq8c" event={"ID":"aba9e84f-bda0-4c6f-883e-ea9106b2521f","Type":"ContainerStarted","Data":"05941149d7cd734110936183168092c47962cd39b538fda7a2fe9a14398cd590"} Nov 25 15:19:18 crc kubenswrapper[4890]: I1125 15:19:18.170750 4890 generic.go:334] "Generic (PLEG): container finished" podID="2c14ab22-4e5a-43e6-be42-956608ab877b" containerID="3936a5f46f158adeb4f89a31e4321cabf8b85456daa4c610139731932f1bd706" exitCode=0 Nov 25 15:19:18 crc kubenswrapper[4890]: I1125 15:19:18.170814 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-hc7pc" event={"ID":"2c14ab22-4e5a-43e6-be42-956608ab877b","Type":"ContainerDied","Data":"3936a5f46f158adeb4f89a31e4321cabf8b85456daa4c610139731932f1bd706"} Nov 25 15:19:18 crc kubenswrapper[4890]: I1125 15:19:18.171038 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-hc7pc" event={"ID":"2c14ab22-4e5a-43e6-be42-956608ab877b","Type":"ContainerStarted","Data":"65efcf2084fe5e456379bba742b6eeef52ac0a1e27a8beb505bd8a479b4c1b07"} Nov 25 15:19:18 crc kubenswrapper[4890]: I1125 15:19:18.174225 4890 generic.go:334] "Generic (PLEG): container finished" podID="b59f5530-3d09-4573-b6f0-beb315e71d79" containerID="e893380f6bc195a09503a2f355f95510799db8ab623f49a44eedfa4e6b040962" exitCode=0 Nov 25 15:19:18 crc kubenswrapper[4890]: I1125 15:19:18.176063 4890 generic.go:334] "Generic (PLEG): container finished" podID="c5d95f8e-951b-402b-afe8-ec8b3d090a45" containerID="6788492c06d96c69fc86d50b4aa2c4f3788e2e10ff1a313014a37ac06c72e219" exitCode=0 Nov 25 15:19:18 crc kubenswrapper[4890]: I1125 15:19:18.177535 4890 generic.go:334] "Generic (PLEG): container finished" podID="acfb093b-f3af-4856-94b8-64a43c9c0cc1" containerID="9329cc99c6bd7e2b61bd28adaf2eb65dd4b8abace7166cfffeb874f0197267b5" exitCode=0 Nov 25 15:19:18 crc kubenswrapper[4890]: I1125 15:19:18.184407 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-9r7gm" podStartSLOduration=2.184384919 podStartE2EDuration="2.184384919s" podCreationTimestamp="2025-11-25 15:19:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:19:18.181691241 +0000 UTC m=+1016.624153851" watchObservedRunningTime="2025-11-25 15:19:18.184384919 +0000 UTC m=+1016.626847529" Nov 25 15:19:18 crc kubenswrapper[4890]: I1125 15:19:18.187950 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-cf3e-account-create-4p2kd" event={"ID":"b59f5530-3d09-4573-b6f0-beb315e71d79","Type":"ContainerDied","Data":"e893380f6bc195a09503a2f355f95510799db8ab623f49a44eedfa4e6b040962"} Nov 25 15:19:18 crc kubenswrapper[4890]: I1125 15:19:18.188009 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-aca8-account-create-gm6cf" event={"ID":"c5d95f8e-951b-402b-afe8-ec8b3d090a45","Type":"ContainerDied","Data":"6788492c06d96c69fc86d50b4aa2c4f3788e2e10ff1a313014a37ac06c72e219"} Nov 25 15:19:18 crc kubenswrapper[4890]: I1125 15:19:18.188026 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-aca8-account-create-gm6cf" event={"ID":"c5d95f8e-951b-402b-afe8-ec8b3d090a45","Type":"ContainerStarted","Data":"9ac7f25feb18fe29469f442008a65d583284315ed9188e0aec5e6924d660847b"} Nov 25 15:19:18 crc kubenswrapper[4890]: I1125 15:19:18.188040 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-xtxvs" event={"ID":"acfb093b-f3af-4856-94b8-64a43c9c0cc1","Type":"ContainerDied","Data":"9329cc99c6bd7e2b61bd28adaf2eb65dd4b8abace7166cfffeb874f0197267b5"} Nov 25 15:19:18 crc kubenswrapper[4890]: I1125 15:19:18.250387 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6517688d-0723-4809-ba51-bcaf48524157-etc-swift\") pod \"swift-storage-0\" (UID: \"6517688d-0723-4809-ba51-bcaf48524157\") " pod="openstack/swift-storage-0" Nov 25 15:19:18 crc kubenswrapper[4890]: E1125 15:19:18.250558 4890 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 25 15:19:18 crc kubenswrapper[4890]: E1125 15:19:18.250588 4890 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 25 15:19:18 crc kubenswrapper[4890]: E1125 15:19:18.250664 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/6517688d-0723-4809-ba51-bcaf48524157-etc-swift podName:6517688d-0723-4809-ba51-bcaf48524157 nodeName:}" failed. No retries permitted until 2025-11-25 15:19:34.250644465 +0000 UTC m=+1032.693107075 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/6517688d-0723-4809-ba51-bcaf48524157-etc-swift") pod "swift-storage-0" (UID: "6517688d-0723-4809-ba51-bcaf48524157") : configmap "swift-ring-files" not found Nov 25 15:19:18 crc kubenswrapper[4890]: I1125 15:19:18.255499 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-90a3-account-create-qvq8c" podStartSLOduration=2.255478016 podStartE2EDuration="2.255478016s" podCreationTimestamp="2025-11-25 15:19:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:19:18.234092181 +0000 UTC m=+1016.676554791" watchObservedRunningTime="2025-11-25 15:19:18.255478016 +0000 UTC m=+1016.697940626" Nov 25 15:19:18 crc kubenswrapper[4890]: I1125 15:19:18.558401 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-74f6f696b9-c9ctc" Nov 25 15:19:18 crc kubenswrapper[4890]: I1125 15:19:18.580998 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-sppf8" Nov 25 15:19:18 crc kubenswrapper[4890]: I1125 15:19:18.660041 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/290df6b3-1433-4137-87e7-5caeba2bbb59-operator-scripts\") pod \"290df6b3-1433-4137-87e7-5caeba2bbb59\" (UID: \"290df6b3-1433-4137-87e7-5caeba2bbb59\") " Nov 25 15:19:18 crc kubenswrapper[4890]: I1125 15:19:18.660181 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ms6qp\" (UniqueName: \"kubernetes.io/projected/290df6b3-1433-4137-87e7-5caeba2bbb59-kube-api-access-ms6qp\") pod \"290df6b3-1433-4137-87e7-5caeba2bbb59\" (UID: \"290df6b3-1433-4137-87e7-5caeba2bbb59\") " Nov 25 15:19:18 crc kubenswrapper[4890]: I1125 15:19:18.660551 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/290df6b3-1433-4137-87e7-5caeba2bbb59-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "290df6b3-1433-4137-87e7-5caeba2bbb59" (UID: "290df6b3-1433-4137-87e7-5caeba2bbb59"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:19:18 crc kubenswrapper[4890]: I1125 15:19:18.660917 4890 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/290df6b3-1433-4137-87e7-5caeba2bbb59-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:18 crc kubenswrapper[4890]: I1125 15:19:18.679592 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/290df6b3-1433-4137-87e7-5caeba2bbb59-kube-api-access-ms6qp" (OuterVolumeSpecName: "kube-api-access-ms6qp") pod "290df6b3-1433-4137-87e7-5caeba2bbb59" (UID: "290df6b3-1433-4137-87e7-5caeba2bbb59"). InnerVolumeSpecName "kube-api-access-ms6qp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:19:18 crc kubenswrapper[4890]: I1125 15:19:18.762526 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ms6qp\" (UniqueName: \"kubernetes.io/projected/290df6b3-1433-4137-87e7-5caeba2bbb59-kube-api-access-ms6qp\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:19 crc kubenswrapper[4890]: I1125 15:19:19.078338 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-lgp8l" Nov 25 15:19:19 crc kubenswrapper[4890]: I1125 15:19:19.130298 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6f696b9-c9ctc"] Nov 25 15:19:19 crc kubenswrapper[4890]: I1125 15:19:19.186937 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-sppf8" Nov 25 15:19:19 crc kubenswrapper[4890]: I1125 15:19:19.186936 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-sppf8" event={"ID":"290df6b3-1433-4137-87e7-5caeba2bbb59","Type":"ContainerDied","Data":"75420806b2db3463f887f51a05a015c28e2614ddb1a2792fc1ee29e7f380f299"} Nov 25 15:19:19 crc kubenswrapper[4890]: I1125 15:19:19.187064 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="75420806b2db3463f887f51a05a015c28e2614ddb1a2792fc1ee29e7f380f299" Nov 25 15:19:19 crc kubenswrapper[4890]: I1125 15:19:19.188913 4890 generic.go:334] "Generic (PLEG): container finished" podID="aba9e84f-bda0-4c6f-883e-ea9106b2521f" containerID="83d8df7f74e56f66d558ce883b578ee8d08a0a14374ce74454a0337e40f72975" exitCode=0 Nov 25 15:19:19 crc kubenswrapper[4890]: I1125 15:19:19.188987 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-90a3-account-create-qvq8c" event={"ID":"aba9e84f-bda0-4c6f-883e-ea9106b2521f","Type":"ContainerDied","Data":"83d8df7f74e56f66d558ce883b578ee8d08a0a14374ce74454a0337e40f72975"} Nov 25 15:19:19 crc kubenswrapper[4890]: I1125 15:19:19.191098 4890 generic.go:334] "Generic (PLEG): container finished" podID="2d380d3e-4b4c-424b-b9af-fb34e959e81c" containerID="3004fbf5a7a3bea35404dc2c53d85fe29655581adbfbcf65d2feead1f3d308ec" exitCode=0 Nov 25 15:19:19 crc kubenswrapper[4890]: I1125 15:19:19.191264 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-9r7gm" event={"ID":"2d380d3e-4b4c-424b-b9af-fb34e959e81c","Type":"ContainerDied","Data":"3004fbf5a7a3bea35404dc2c53d85fe29655581adbfbcf65d2feead1f3d308ec"} Nov 25 15:19:19 crc kubenswrapper[4890]: I1125 15:19:19.191482 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-74f6f696b9-c9ctc" podUID="1c1e0704-e54b-48b3-bcc3-6af856dfd5ff" containerName="dnsmasq-dns" containerID="cri-o://6a98dd24abec302a59e34c75f45a68d7b45738b22aaa348d61ca26a6fdb076a5" gracePeriod=10 Nov 25 15:19:19 crc kubenswrapper[4890]: I1125 15:19:19.404141 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-vgq94"] Nov 25 15:19:19 crc kubenswrapper[4890]: E1125 15:19:19.404525 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="290df6b3-1433-4137-87e7-5caeba2bbb59" containerName="mariadb-database-create" Nov 25 15:19:19 crc kubenswrapper[4890]: I1125 15:19:19.404545 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="290df6b3-1433-4137-87e7-5caeba2bbb59" containerName="mariadb-database-create" Nov 25 15:19:19 crc kubenswrapper[4890]: I1125 15:19:19.404803 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="290df6b3-1433-4137-87e7-5caeba2bbb59" containerName="mariadb-database-create" Nov 25 15:19:19 crc kubenswrapper[4890]: I1125 15:19:19.405370 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-vgq94" Nov 25 15:19:19 crc kubenswrapper[4890]: I1125 15:19:19.408767 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-t8zql" Nov 25 15:19:19 crc kubenswrapper[4890]: I1125 15:19:19.408994 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 25 15:19:19 crc kubenswrapper[4890]: I1125 15:19:19.409493 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 25 15:19:19 crc kubenswrapper[4890]: I1125 15:19:19.411450 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 25 15:19:19 crc kubenswrapper[4890]: I1125 15:19:19.426600 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-vgq94"] Nov 25 15:19:19 crc kubenswrapper[4890]: I1125 15:19:19.480128 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21963cd4-490e-4c7f-9a9b-d93fa2a90ebc-config-data\") pod \"keystone-db-sync-vgq94\" (UID: \"21963cd4-490e-4c7f-9a9b-d93fa2a90ebc\") " pod="openstack/keystone-db-sync-vgq94" Nov 25 15:19:19 crc kubenswrapper[4890]: I1125 15:19:19.480284 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21963cd4-490e-4c7f-9a9b-d93fa2a90ebc-combined-ca-bundle\") pod \"keystone-db-sync-vgq94\" (UID: \"21963cd4-490e-4c7f-9a9b-d93fa2a90ebc\") " pod="openstack/keystone-db-sync-vgq94" Nov 25 15:19:19 crc kubenswrapper[4890]: I1125 15:19:19.480332 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njxsq\" (UniqueName: \"kubernetes.io/projected/21963cd4-490e-4c7f-9a9b-d93fa2a90ebc-kube-api-access-njxsq\") pod \"keystone-db-sync-vgq94\" (UID: \"21963cd4-490e-4c7f-9a9b-d93fa2a90ebc\") " pod="openstack/keystone-db-sync-vgq94" Nov 25 15:19:19 crc kubenswrapper[4890]: I1125 15:19:19.582507 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21963cd4-490e-4c7f-9a9b-d93fa2a90ebc-combined-ca-bundle\") pod \"keystone-db-sync-vgq94\" (UID: \"21963cd4-490e-4c7f-9a9b-d93fa2a90ebc\") " pod="openstack/keystone-db-sync-vgq94" Nov 25 15:19:19 crc kubenswrapper[4890]: I1125 15:19:19.582629 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njxsq\" (UniqueName: \"kubernetes.io/projected/21963cd4-490e-4c7f-9a9b-d93fa2a90ebc-kube-api-access-njxsq\") pod \"keystone-db-sync-vgq94\" (UID: \"21963cd4-490e-4c7f-9a9b-d93fa2a90ebc\") " pod="openstack/keystone-db-sync-vgq94" Nov 25 15:19:19 crc kubenswrapper[4890]: I1125 15:19:19.582805 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21963cd4-490e-4c7f-9a9b-d93fa2a90ebc-config-data\") pod \"keystone-db-sync-vgq94\" (UID: \"21963cd4-490e-4c7f-9a9b-d93fa2a90ebc\") " pod="openstack/keystone-db-sync-vgq94" Nov 25 15:19:19 crc kubenswrapper[4890]: I1125 15:19:19.598986 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21963cd4-490e-4c7f-9a9b-d93fa2a90ebc-combined-ca-bundle\") pod \"keystone-db-sync-vgq94\" (UID: \"21963cd4-490e-4c7f-9a9b-d93fa2a90ebc\") " pod="openstack/keystone-db-sync-vgq94" Nov 25 15:19:19 crc kubenswrapper[4890]: I1125 15:19:19.599597 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21963cd4-490e-4c7f-9a9b-d93fa2a90ebc-config-data\") pod \"keystone-db-sync-vgq94\" (UID: \"21963cd4-490e-4c7f-9a9b-d93fa2a90ebc\") " pod="openstack/keystone-db-sync-vgq94" Nov 25 15:19:19 crc kubenswrapper[4890]: I1125 15:19:19.608604 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njxsq\" (UniqueName: \"kubernetes.io/projected/21963cd4-490e-4c7f-9a9b-d93fa2a90ebc-kube-api-access-njxsq\") pod \"keystone-db-sync-vgq94\" (UID: \"21963cd4-490e-4c7f-9a9b-d93fa2a90ebc\") " pod="openstack/keystone-db-sync-vgq94" Nov 25 15:19:19 crc kubenswrapper[4890]: I1125 15:19:19.730806 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-vgq94" Nov 25 15:19:19 crc kubenswrapper[4890]: I1125 15:19:19.789799 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-aca8-account-create-gm6cf" Nov 25 15:19:19 crc kubenswrapper[4890]: I1125 15:19:19.894965 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c5d95f8e-951b-402b-afe8-ec8b3d090a45-operator-scripts\") pod \"c5d95f8e-951b-402b-afe8-ec8b3d090a45\" (UID: \"c5d95f8e-951b-402b-afe8-ec8b3d090a45\") " Nov 25 15:19:19 crc kubenswrapper[4890]: I1125 15:19:19.895142 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bgg62\" (UniqueName: \"kubernetes.io/projected/c5d95f8e-951b-402b-afe8-ec8b3d090a45-kube-api-access-bgg62\") pod \"c5d95f8e-951b-402b-afe8-ec8b3d090a45\" (UID: \"c5d95f8e-951b-402b-afe8-ec8b3d090a45\") " Nov 25 15:19:19 crc kubenswrapper[4890]: I1125 15:19:19.895735 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5d95f8e-951b-402b-afe8-ec8b3d090a45-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c5d95f8e-951b-402b-afe8-ec8b3d090a45" (UID: "c5d95f8e-951b-402b-afe8-ec8b3d090a45"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:19:19 crc kubenswrapper[4890]: I1125 15:19:19.905736 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5d95f8e-951b-402b-afe8-ec8b3d090a45-kube-api-access-bgg62" (OuterVolumeSpecName: "kube-api-access-bgg62") pod "c5d95f8e-951b-402b-afe8-ec8b3d090a45" (UID: "c5d95f8e-951b-402b-afe8-ec8b3d090a45"). InnerVolumeSpecName "kube-api-access-bgg62". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:19:19 crc kubenswrapper[4890]: I1125 15:19:19.973715 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-cf3e-account-create-4p2kd" Nov 25 15:19:19 crc kubenswrapper[4890]: I1125 15:19:19.981460 4890 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-nrbtc" podUID="524ff107-da15-4e2f-a0c8-acca0e91b27a" containerName="ovn-controller" probeResult="failure" output=< Nov 25 15:19:19 crc kubenswrapper[4890]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 25 15:19:19 crc kubenswrapper[4890]: > Nov 25 15:19:19 crc kubenswrapper[4890]: I1125 15:19:19.983754 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-9354-account-create-5fdrf" Nov 25 15:19:19 crc kubenswrapper[4890]: I1125 15:19:19.999251 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bgg62\" (UniqueName: \"kubernetes.io/projected/c5d95f8e-951b-402b-afe8-ec8b3d090a45-kube-api-access-bgg62\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:19 crc kubenswrapper[4890]: I1125 15:19:19.999288 4890 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c5d95f8e-951b-402b-afe8-ec8b3d090a45-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:19 crc kubenswrapper[4890]: I1125 15:19:19.999480 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-xtxvs" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.018789 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-hc7pc" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.081610 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6f696b9-c9ctc" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.106151 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8cd74cd-df01-4e7b-a8a9-391d5afbe2f9-operator-scripts\") pod \"a8cd74cd-df01-4e7b-a8a9-391d5afbe2f9\" (UID: \"a8cd74cd-df01-4e7b-a8a9-391d5afbe2f9\") " Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.106261 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hm8hb\" (UniqueName: \"kubernetes.io/projected/2c14ab22-4e5a-43e6-be42-956608ab877b-kube-api-access-hm8hb\") pod \"2c14ab22-4e5a-43e6-be42-956608ab877b\" (UID: \"2c14ab22-4e5a-43e6-be42-956608ab877b\") " Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.106295 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j6z4w\" (UniqueName: \"kubernetes.io/projected/a8cd74cd-df01-4e7b-a8a9-391d5afbe2f9-kube-api-access-j6z4w\") pod \"a8cd74cd-df01-4e7b-a8a9-391d5afbe2f9\" (UID: \"a8cd74cd-df01-4e7b-a8a9-391d5afbe2f9\") " Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.106332 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2c14ab22-4e5a-43e6-be42-956608ab877b-operator-scripts\") pod \"2c14ab22-4e5a-43e6-be42-956608ab877b\" (UID: \"2c14ab22-4e5a-43e6-be42-956608ab877b\") " Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.106354 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xbmk\" (UniqueName: \"kubernetes.io/projected/b59f5530-3d09-4573-b6f0-beb315e71d79-kube-api-access-7xbmk\") pod \"b59f5530-3d09-4573-b6f0-beb315e71d79\" (UID: \"b59f5530-3d09-4573-b6f0-beb315e71d79\") " Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.106375 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/acfb093b-f3af-4856-94b8-64a43c9c0cc1-operator-scripts\") pod \"acfb093b-f3af-4856-94b8-64a43c9c0cc1\" (UID: \"acfb093b-f3af-4856-94b8-64a43c9c0cc1\") " Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.106395 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j49nl\" (UniqueName: \"kubernetes.io/projected/acfb093b-f3af-4856-94b8-64a43c9c0cc1-kube-api-access-j49nl\") pod \"acfb093b-f3af-4856-94b8-64a43c9c0cc1\" (UID: \"acfb093b-f3af-4856-94b8-64a43c9c0cc1\") " Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.106466 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b59f5530-3d09-4573-b6f0-beb315e71d79-operator-scripts\") pod \"b59f5530-3d09-4573-b6f0-beb315e71d79\" (UID: \"b59f5530-3d09-4573-b6f0-beb315e71d79\") " Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.106680 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8cd74cd-df01-4e7b-a8a9-391d5afbe2f9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a8cd74cd-df01-4e7b-a8a9-391d5afbe2f9" (UID: "a8cd74cd-df01-4e7b-a8a9-391d5afbe2f9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.106934 4890 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8cd74cd-df01-4e7b-a8a9-391d5afbe2f9-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.107596 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/acfb093b-f3af-4856-94b8-64a43c9c0cc1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "acfb093b-f3af-4856-94b8-64a43c9c0cc1" (UID: "acfb093b-f3af-4856-94b8-64a43c9c0cc1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.107997 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c14ab22-4e5a-43e6-be42-956608ab877b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2c14ab22-4e5a-43e6-be42-956608ab877b" (UID: "2c14ab22-4e5a-43e6-be42-956608ab877b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.111365 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c14ab22-4e5a-43e6-be42-956608ab877b-kube-api-access-hm8hb" (OuterVolumeSpecName: "kube-api-access-hm8hb") pod "2c14ab22-4e5a-43e6-be42-956608ab877b" (UID: "2c14ab22-4e5a-43e6-be42-956608ab877b"). InnerVolumeSpecName "kube-api-access-hm8hb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.111573 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b59f5530-3d09-4573-b6f0-beb315e71d79-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b59f5530-3d09-4573-b6f0-beb315e71d79" (UID: "b59f5530-3d09-4573-b6f0-beb315e71d79"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.112309 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b59f5530-3d09-4573-b6f0-beb315e71d79-kube-api-access-7xbmk" (OuterVolumeSpecName: "kube-api-access-7xbmk") pod "b59f5530-3d09-4573-b6f0-beb315e71d79" (UID: "b59f5530-3d09-4573-b6f0-beb315e71d79"). InnerVolumeSpecName "kube-api-access-7xbmk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.113096 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/acfb093b-f3af-4856-94b8-64a43c9c0cc1-kube-api-access-j49nl" (OuterVolumeSpecName: "kube-api-access-j49nl") pod "acfb093b-f3af-4856-94b8-64a43c9c0cc1" (UID: "acfb093b-f3af-4856-94b8-64a43c9c0cc1"). InnerVolumeSpecName "kube-api-access-j49nl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.113964 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8cd74cd-df01-4e7b-a8a9-391d5afbe2f9-kube-api-access-j6z4w" (OuterVolumeSpecName: "kube-api-access-j6z4w") pod "a8cd74cd-df01-4e7b-a8a9-391d5afbe2f9" (UID: "a8cd74cd-df01-4e7b-a8a9-391d5afbe2f9"). InnerVolumeSpecName "kube-api-access-j6z4w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.208217 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ld8g8\" (UniqueName: \"kubernetes.io/projected/1c1e0704-e54b-48b3-bcc3-6af856dfd5ff-kube-api-access-ld8g8\") pod \"1c1e0704-e54b-48b3-bcc3-6af856dfd5ff\" (UID: \"1c1e0704-e54b-48b3-bcc3-6af856dfd5ff\") " Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.208328 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1c1e0704-e54b-48b3-bcc3-6af856dfd5ff-dns-svc\") pod \"1c1e0704-e54b-48b3-bcc3-6af856dfd5ff\" (UID: \"1c1e0704-e54b-48b3-bcc3-6af856dfd5ff\") " Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.208435 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1c1e0704-e54b-48b3-bcc3-6af856dfd5ff-ovsdbserver-nb\") pod \"1c1e0704-e54b-48b3-bcc3-6af856dfd5ff\" (UID: \"1c1e0704-e54b-48b3-bcc3-6af856dfd5ff\") " Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.208551 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c1e0704-e54b-48b3-bcc3-6af856dfd5ff-config\") pod \"1c1e0704-e54b-48b3-bcc3-6af856dfd5ff\" (UID: \"1c1e0704-e54b-48b3-bcc3-6af856dfd5ff\") " Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.209873 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hm8hb\" (UniqueName: \"kubernetes.io/projected/2c14ab22-4e5a-43e6-be42-956608ab877b-kube-api-access-hm8hb\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.209902 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j6z4w\" (UniqueName: \"kubernetes.io/projected/a8cd74cd-df01-4e7b-a8a9-391d5afbe2f9-kube-api-access-j6z4w\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.209922 4890 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2c14ab22-4e5a-43e6-be42-956608ab877b-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.209935 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7xbmk\" (UniqueName: \"kubernetes.io/projected/b59f5530-3d09-4573-b6f0-beb315e71d79-kube-api-access-7xbmk\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.209947 4890 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/acfb093b-f3af-4856-94b8-64a43c9c0cc1-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.209960 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j49nl\" (UniqueName: \"kubernetes.io/projected/acfb093b-f3af-4856-94b8-64a43c9c0cc1-kube-api-access-j49nl\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.209977 4890 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b59f5530-3d09-4573-b6f0-beb315e71d79-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.217634 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-9354-account-create-5fdrf" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.217658 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-9354-account-create-5fdrf" event={"ID":"a8cd74cd-df01-4e7b-a8a9-391d5afbe2f9","Type":"ContainerDied","Data":"387d8aa9eacd58d49e790db4adc50064e6b7d77f8613ed55a238e180fe1ca6e6"} Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.217719 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="387d8aa9eacd58d49e790db4adc50064e6b7d77f8613ed55a238e180fe1ca6e6" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.219445 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c1e0704-e54b-48b3-bcc3-6af856dfd5ff-kube-api-access-ld8g8" (OuterVolumeSpecName: "kube-api-access-ld8g8") pod "1c1e0704-e54b-48b3-bcc3-6af856dfd5ff" (UID: "1c1e0704-e54b-48b3-bcc3-6af856dfd5ff"). InnerVolumeSpecName "kube-api-access-ld8g8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.221142 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-hc7pc" event={"ID":"2c14ab22-4e5a-43e6-be42-956608ab877b","Type":"ContainerDied","Data":"65efcf2084fe5e456379bba742b6eeef52ac0a1e27a8beb505bd8a479b4c1b07"} Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.221187 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="65efcf2084fe5e456379bba742b6eeef52ac0a1e27a8beb505bd8a479b4c1b07" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.221247 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-hc7pc" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.226384 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-cf3e-account-create-4p2kd" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.226402 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-cf3e-account-create-4p2kd" event={"ID":"b59f5530-3d09-4573-b6f0-beb315e71d79","Type":"ContainerDied","Data":"3e1e9c172609ba3e3cd4d962326761f5f3ec1a38b3601981ef63824853d9fc66"} Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.226458 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3e1e9c172609ba3e3cd4d962326761f5f3ec1a38b3601981ef63824853d9fc66" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.228669 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-aca8-account-create-gm6cf" event={"ID":"c5d95f8e-951b-402b-afe8-ec8b3d090a45","Type":"ContainerDied","Data":"9ac7f25feb18fe29469f442008a65d583284315ed9188e0aec5e6924d660847b"} Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.228706 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ac7f25feb18fe29469f442008a65d583284315ed9188e0aec5e6924d660847b" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.228786 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-aca8-account-create-gm6cf" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.238705 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-xtxvs" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.239512 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-xtxvs" event={"ID":"acfb093b-f3af-4856-94b8-64a43c9c0cc1","Type":"ContainerDied","Data":"1a39441655b4cea76969e1e7fd436cf2ba0aafde43dc97bc0c566b83bbfdcc58"} Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.239609 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a39441655b4cea76969e1e7fd436cf2ba0aafde43dc97bc0c566b83bbfdcc58" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.242140 4890 generic.go:334] "Generic (PLEG): container finished" podID="1c1e0704-e54b-48b3-bcc3-6af856dfd5ff" containerID="6a98dd24abec302a59e34c75f45a68d7b45738b22aaa348d61ca26a6fdb076a5" exitCode=0 Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.242267 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6f696b9-c9ctc" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.242350 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6f696b9-c9ctc" event={"ID":"1c1e0704-e54b-48b3-bcc3-6af856dfd5ff","Type":"ContainerDied","Data":"6a98dd24abec302a59e34c75f45a68d7b45738b22aaa348d61ca26a6fdb076a5"} Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.242390 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6f696b9-c9ctc" event={"ID":"1c1e0704-e54b-48b3-bcc3-6af856dfd5ff","Type":"ContainerDied","Data":"b1d3c82893a0e6bf30870137a704670554e1533d91fd8eff9d571c3b4bb1f740"} Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.242417 4890 scope.go:117] "RemoveContainer" containerID="6a98dd24abec302a59e34c75f45a68d7b45738b22aaa348d61ca26a6fdb076a5" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.264039 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c1e0704-e54b-48b3-bcc3-6af856dfd5ff-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1c1e0704-e54b-48b3-bcc3-6af856dfd5ff" (UID: "1c1e0704-e54b-48b3-bcc3-6af856dfd5ff"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.265095 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c1e0704-e54b-48b3-bcc3-6af856dfd5ff-config" (OuterVolumeSpecName: "config") pod "1c1e0704-e54b-48b3-bcc3-6af856dfd5ff" (UID: "1c1e0704-e54b-48b3-bcc3-6af856dfd5ff"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.273409 4890 scope.go:117] "RemoveContainer" containerID="d3d32a8939d3efcf680582a6c7f4ad33c857779698cbf795d61a9301d73439c2" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.284548 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c1e0704-e54b-48b3-bcc3-6af856dfd5ff-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1c1e0704-e54b-48b3-bcc3-6af856dfd5ff" (UID: "1c1e0704-e54b-48b3-bcc3-6af856dfd5ff"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.306282 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-vgq94"] Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.308133 4890 scope.go:117] "RemoveContainer" containerID="6a98dd24abec302a59e34c75f45a68d7b45738b22aaa348d61ca26a6fdb076a5" Nov 25 15:19:20 crc kubenswrapper[4890]: E1125 15:19:20.312402 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a98dd24abec302a59e34c75f45a68d7b45738b22aaa348d61ca26a6fdb076a5\": container with ID starting with 6a98dd24abec302a59e34c75f45a68d7b45738b22aaa348d61ca26a6fdb076a5 not found: ID does not exist" containerID="6a98dd24abec302a59e34c75f45a68d7b45738b22aaa348d61ca26a6fdb076a5" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.312458 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a98dd24abec302a59e34c75f45a68d7b45738b22aaa348d61ca26a6fdb076a5"} err="failed to get container status \"6a98dd24abec302a59e34c75f45a68d7b45738b22aaa348d61ca26a6fdb076a5\": rpc error: code = NotFound desc = could not find container \"6a98dd24abec302a59e34c75f45a68d7b45738b22aaa348d61ca26a6fdb076a5\": container with ID starting with 6a98dd24abec302a59e34c75f45a68d7b45738b22aaa348d61ca26a6fdb076a5 not found: ID does not exist" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.312492 4890 scope.go:117] "RemoveContainer" containerID="d3d32a8939d3efcf680582a6c7f4ad33c857779698cbf795d61a9301d73439c2" Nov 25 15:19:20 crc kubenswrapper[4890]: E1125 15:19:20.313945 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d3d32a8939d3efcf680582a6c7f4ad33c857779698cbf795d61a9301d73439c2\": container with ID starting with d3d32a8939d3efcf680582a6c7f4ad33c857779698cbf795d61a9301d73439c2 not found: ID does not exist" containerID="d3d32a8939d3efcf680582a6c7f4ad33c857779698cbf795d61a9301d73439c2" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.313976 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3d32a8939d3efcf680582a6c7f4ad33c857779698cbf795d61a9301d73439c2"} err="failed to get container status \"d3d32a8939d3efcf680582a6c7f4ad33c857779698cbf795d61a9301d73439c2\": rpc error: code = NotFound desc = could not find container \"d3d32a8939d3efcf680582a6c7f4ad33c857779698cbf795d61a9301d73439c2\": container with ID starting with d3d32a8939d3efcf680582a6c7f4ad33c857779698cbf795d61a9301d73439c2 not found: ID does not exist" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.315101 4890 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1c1e0704-e54b-48b3-bcc3-6af856dfd5ff-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.315127 4890 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c1e0704-e54b-48b3-bcc3-6af856dfd5ff-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.315136 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ld8g8\" (UniqueName: \"kubernetes.io/projected/1c1e0704-e54b-48b3-bcc3-6af856dfd5ff-kube-api-access-ld8g8\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.315147 4890 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1c1e0704-e54b-48b3-bcc3-6af856dfd5ff-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:20 crc kubenswrapper[4890]: W1125 15:19:20.324295 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod21963cd4_490e_4c7f_9a9b_d93fa2a90ebc.slice/crio-1cbc76f9f5e036ca8c4ad9b03384a46a944097733fae34917075b7bebe053992 WatchSource:0}: Error finding container 1cbc76f9f5e036ca8c4ad9b03384a46a944097733fae34917075b7bebe053992: Status 404 returned error can't find the container with id 1cbc76f9f5e036ca8c4ad9b03384a46a944097733fae34917075b7bebe053992 Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.586149 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6f696b9-c9ctc"] Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.593732 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-9r7gm" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.599011 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74f6f696b9-c9ctc"] Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.645989 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-90a3-account-create-qvq8c" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.722685 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrds8\" (UniqueName: \"kubernetes.io/projected/aba9e84f-bda0-4c6f-883e-ea9106b2521f-kube-api-access-lrds8\") pod \"aba9e84f-bda0-4c6f-883e-ea9106b2521f\" (UID: \"aba9e84f-bda0-4c6f-883e-ea9106b2521f\") " Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.722769 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aba9e84f-bda0-4c6f-883e-ea9106b2521f-operator-scripts\") pod \"aba9e84f-bda0-4c6f-883e-ea9106b2521f\" (UID: \"aba9e84f-bda0-4c6f-883e-ea9106b2521f\") " Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.722821 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f8rzs\" (UniqueName: \"kubernetes.io/projected/2d380d3e-4b4c-424b-b9af-fb34e959e81c-kube-api-access-f8rzs\") pod \"2d380d3e-4b4c-424b-b9af-fb34e959e81c\" (UID: \"2d380d3e-4b4c-424b-b9af-fb34e959e81c\") " Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.722887 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d380d3e-4b4c-424b-b9af-fb34e959e81c-operator-scripts\") pod \"2d380d3e-4b4c-424b-b9af-fb34e959e81c\" (UID: \"2d380d3e-4b4c-424b-b9af-fb34e959e81c\") " Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.723803 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d380d3e-4b4c-424b-b9af-fb34e959e81c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2d380d3e-4b4c-424b-b9af-fb34e959e81c" (UID: "2d380d3e-4b4c-424b-b9af-fb34e959e81c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.724532 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aba9e84f-bda0-4c6f-883e-ea9106b2521f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "aba9e84f-bda0-4c6f-883e-ea9106b2521f" (UID: "aba9e84f-bda0-4c6f-883e-ea9106b2521f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.728442 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d380d3e-4b4c-424b-b9af-fb34e959e81c-kube-api-access-f8rzs" (OuterVolumeSpecName: "kube-api-access-f8rzs") pod "2d380d3e-4b4c-424b-b9af-fb34e959e81c" (UID: "2d380d3e-4b4c-424b-b9af-fb34e959e81c"). InnerVolumeSpecName "kube-api-access-f8rzs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.728794 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aba9e84f-bda0-4c6f-883e-ea9106b2521f-kube-api-access-lrds8" (OuterVolumeSpecName: "kube-api-access-lrds8") pod "aba9e84f-bda0-4c6f-883e-ea9106b2521f" (UID: "aba9e84f-bda0-4c6f-883e-ea9106b2521f"). InnerVolumeSpecName "kube-api-access-lrds8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.825366 4890 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d380d3e-4b4c-424b-b9af-fb34e959e81c-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.825403 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrds8\" (UniqueName: \"kubernetes.io/projected/aba9e84f-bda0-4c6f-883e-ea9106b2521f-kube-api-access-lrds8\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.825418 4890 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aba9e84f-bda0-4c6f-883e-ea9106b2521f-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:20 crc kubenswrapper[4890]: I1125 15:19:20.825427 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f8rzs\" (UniqueName: \"kubernetes.io/projected/2d380d3e-4b4c-424b-b9af-fb34e959e81c-kube-api-access-f8rzs\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:21 crc kubenswrapper[4890]: I1125 15:19:21.253494 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-vgq94" event={"ID":"21963cd4-490e-4c7f-9a9b-d93fa2a90ebc","Type":"ContainerStarted","Data":"1cbc76f9f5e036ca8c4ad9b03384a46a944097733fae34917075b7bebe053992"} Nov 25 15:19:21 crc kubenswrapper[4890]: I1125 15:19:21.255329 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-90a3-account-create-qvq8c" Nov 25 15:19:21 crc kubenswrapper[4890]: I1125 15:19:21.255466 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-90a3-account-create-qvq8c" event={"ID":"aba9e84f-bda0-4c6f-883e-ea9106b2521f","Type":"ContainerDied","Data":"05941149d7cd734110936183168092c47962cd39b538fda7a2fe9a14398cd590"} Nov 25 15:19:21 crc kubenswrapper[4890]: I1125 15:19:21.255519 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="05941149d7cd734110936183168092c47962cd39b538fda7a2fe9a14398cd590" Nov 25 15:19:21 crc kubenswrapper[4890]: I1125 15:19:21.256935 4890 generic.go:334] "Generic (PLEG): container finished" podID="cf34ad11-7dd6-4572-b3d2-4c14127a1d76" containerID="4d2e53b036518a10a8ae51858d830607e13fedd4e8578b62ec853a3e1b0a89c2" exitCode=0 Nov 25 15:19:21 crc kubenswrapper[4890]: I1125 15:19:21.256976 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-vs9dw" event={"ID":"cf34ad11-7dd6-4572-b3d2-4c14127a1d76","Type":"ContainerDied","Data":"4d2e53b036518a10a8ae51858d830607e13fedd4e8578b62ec853a3e1b0a89c2"} Nov 25 15:19:21 crc kubenswrapper[4890]: I1125 15:19:21.258197 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-9r7gm" event={"ID":"2d380d3e-4b4c-424b-b9af-fb34e959e81c","Type":"ContainerDied","Data":"851e818154b0438d38dde98332ccde96f27e8fe5221271c46a649c14532a61e0"} Nov 25 15:19:21 crc kubenswrapper[4890]: I1125 15:19:21.258219 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="851e818154b0438d38dde98332ccde96f27e8fe5221271c46a649c14532a61e0" Nov 25 15:19:21 crc kubenswrapper[4890]: I1125 15:19:21.258250 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-9r7gm" Nov 25 15:19:22 crc kubenswrapper[4890]: I1125 15:19:22.193887 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c1e0704-e54b-48b3-bcc3-6af856dfd5ff" path="/var/lib/kubelet/pods/1c1e0704-e54b-48b3-bcc3-6af856dfd5ff/volumes" Nov 25 15:19:22 crc kubenswrapper[4890]: I1125 15:19:22.268518 4890 generic.go:334] "Generic (PLEG): container finished" podID="c14ce73e-1a95-4f41-b6b6-33478079806f" containerID="7b0b47083ae2bc3b6521792dbfc2606a74d4a262bfbeca4040d61b4873018ae8" exitCode=0 Nov 25 15:19:22 crc kubenswrapper[4890]: I1125 15:19:22.268593 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c14ce73e-1a95-4f41-b6b6-33478079806f","Type":"ContainerDied","Data":"7b0b47083ae2bc3b6521792dbfc2606a74d4a262bfbeca4040d61b4873018ae8"} Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.182046 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.493308 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-c5b26"] Nov 25 15:19:24 crc kubenswrapper[4890]: E1125 15:19:24.494089 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d380d3e-4b4c-424b-b9af-fb34e959e81c" containerName="mariadb-database-create" Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.494212 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d380d3e-4b4c-424b-b9af-fb34e959e81c" containerName="mariadb-database-create" Nov 25 15:19:24 crc kubenswrapper[4890]: E1125 15:19:24.494324 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c1e0704-e54b-48b3-bcc3-6af856dfd5ff" containerName="dnsmasq-dns" Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.494401 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c1e0704-e54b-48b3-bcc3-6af856dfd5ff" containerName="dnsmasq-dns" Nov 25 15:19:24 crc kubenswrapper[4890]: E1125 15:19:24.494472 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8cd74cd-df01-4e7b-a8a9-391d5afbe2f9" containerName="mariadb-account-create" Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.494549 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8cd74cd-df01-4e7b-a8a9-391d5afbe2f9" containerName="mariadb-account-create" Nov 25 15:19:24 crc kubenswrapper[4890]: E1125 15:19:24.494641 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c1e0704-e54b-48b3-bcc3-6af856dfd5ff" containerName="init" Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.494721 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c1e0704-e54b-48b3-bcc3-6af856dfd5ff" containerName="init" Nov 25 15:19:24 crc kubenswrapper[4890]: E1125 15:19:24.495114 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acfb093b-f3af-4856-94b8-64a43c9c0cc1" containerName="mariadb-database-create" Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.495278 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="acfb093b-f3af-4856-94b8-64a43c9c0cc1" containerName="mariadb-database-create" Nov 25 15:19:24 crc kubenswrapper[4890]: E1125 15:19:24.495442 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aba9e84f-bda0-4c6f-883e-ea9106b2521f" containerName="mariadb-account-create" Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.495586 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="aba9e84f-bda0-4c6f-883e-ea9106b2521f" containerName="mariadb-account-create" Nov 25 15:19:24 crc kubenswrapper[4890]: E1125 15:19:24.495852 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c14ab22-4e5a-43e6-be42-956608ab877b" containerName="mariadb-database-create" Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.495936 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c14ab22-4e5a-43e6-be42-956608ab877b" containerName="mariadb-database-create" Nov 25 15:19:24 crc kubenswrapper[4890]: E1125 15:19:24.496034 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b59f5530-3d09-4573-b6f0-beb315e71d79" containerName="mariadb-account-create" Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.496268 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="b59f5530-3d09-4573-b6f0-beb315e71d79" containerName="mariadb-account-create" Nov 25 15:19:24 crc kubenswrapper[4890]: E1125 15:19:24.496358 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5d95f8e-951b-402b-afe8-ec8b3d090a45" containerName="mariadb-account-create" Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.496435 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5d95f8e-951b-402b-afe8-ec8b3d090a45" containerName="mariadb-account-create" Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.497277 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8cd74cd-df01-4e7b-a8a9-391d5afbe2f9" containerName="mariadb-account-create" Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.497402 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="acfb093b-f3af-4856-94b8-64a43c9c0cc1" containerName="mariadb-database-create" Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.497486 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="b59f5530-3d09-4573-b6f0-beb315e71d79" containerName="mariadb-account-create" Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.497571 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d380d3e-4b4c-424b-b9af-fb34e959e81c" containerName="mariadb-database-create" Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.497791 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c1e0704-e54b-48b3-bcc3-6af856dfd5ff" containerName="dnsmasq-dns" Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.497885 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="aba9e84f-bda0-4c6f-883e-ea9106b2521f" containerName="mariadb-account-create" Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.497991 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5d95f8e-951b-402b-afe8-ec8b3d090a45" containerName="mariadb-account-create" Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.498070 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c14ab22-4e5a-43e6-be42-956608ab877b" containerName="mariadb-database-create" Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.498759 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-c5b26" Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.501273 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-qjthd" Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.504339 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.513959 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-c5b26"] Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.514008 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-vs9dw" Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.590372 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/cf34ad11-7dd6-4572-b3d2-4c14127a1d76-etc-swift\") pod \"cf34ad11-7dd6-4572-b3d2-4c14127a1d76\" (UID: \"cf34ad11-7dd6-4572-b3d2-4c14127a1d76\") " Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.590464 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/cf34ad11-7dd6-4572-b3d2-4c14127a1d76-swiftconf\") pod \"cf34ad11-7dd6-4572-b3d2-4c14127a1d76\" (UID: \"cf34ad11-7dd6-4572-b3d2-4c14127a1d76\") " Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.590528 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cf34ad11-7dd6-4572-b3d2-4c14127a1d76-scripts\") pod \"cf34ad11-7dd6-4572-b3d2-4c14127a1d76\" (UID: \"cf34ad11-7dd6-4572-b3d2-4c14127a1d76\") " Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.590631 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nnhvr\" (UniqueName: \"kubernetes.io/projected/cf34ad11-7dd6-4572-b3d2-4c14127a1d76-kube-api-access-nnhvr\") pod \"cf34ad11-7dd6-4572-b3d2-4c14127a1d76\" (UID: \"cf34ad11-7dd6-4572-b3d2-4c14127a1d76\") " Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.590671 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf34ad11-7dd6-4572-b3d2-4c14127a1d76-combined-ca-bundle\") pod \"cf34ad11-7dd6-4572-b3d2-4c14127a1d76\" (UID: \"cf34ad11-7dd6-4572-b3d2-4c14127a1d76\") " Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.590978 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/cf34ad11-7dd6-4572-b3d2-4c14127a1d76-ring-data-devices\") pod \"cf34ad11-7dd6-4572-b3d2-4c14127a1d76\" (UID: \"cf34ad11-7dd6-4572-b3d2-4c14127a1d76\") " Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.591219 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/cf34ad11-7dd6-4572-b3d2-4c14127a1d76-dispersionconf\") pod \"cf34ad11-7dd6-4572-b3d2-4c14127a1d76\" (UID: \"cf34ad11-7dd6-4572-b3d2-4c14127a1d76\") " Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.591599 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfc6fde5-9e4c-4224-8200-0f230e127f7e-combined-ca-bundle\") pod \"glance-db-sync-c5b26\" (UID: \"dfc6fde5-9e4c-4224-8200-0f230e127f7e\") " pod="openstack/glance-db-sync-c5b26" Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.591675 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfc6fde5-9e4c-4224-8200-0f230e127f7e-config-data\") pod \"glance-db-sync-c5b26\" (UID: \"dfc6fde5-9e4c-4224-8200-0f230e127f7e\") " pod="openstack/glance-db-sync-c5b26" Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.591737 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/dfc6fde5-9e4c-4224-8200-0f230e127f7e-db-sync-config-data\") pod \"glance-db-sync-c5b26\" (UID: \"dfc6fde5-9e4c-4224-8200-0f230e127f7e\") " pod="openstack/glance-db-sync-c5b26" Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.591846 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxv7c\" (UniqueName: \"kubernetes.io/projected/dfc6fde5-9e4c-4224-8200-0f230e127f7e-kube-api-access-fxv7c\") pod \"glance-db-sync-c5b26\" (UID: \"dfc6fde5-9e4c-4224-8200-0f230e127f7e\") " pod="openstack/glance-db-sync-c5b26" Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.592066 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf34ad11-7dd6-4572-b3d2-4c14127a1d76-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "cf34ad11-7dd6-4572-b3d2-4c14127a1d76" (UID: "cf34ad11-7dd6-4572-b3d2-4c14127a1d76"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.592733 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf34ad11-7dd6-4572-b3d2-4c14127a1d76-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "cf34ad11-7dd6-4572-b3d2-4c14127a1d76" (UID: "cf34ad11-7dd6-4572-b3d2-4c14127a1d76"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.598365 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf34ad11-7dd6-4572-b3d2-4c14127a1d76-kube-api-access-nnhvr" (OuterVolumeSpecName: "kube-api-access-nnhvr") pod "cf34ad11-7dd6-4572-b3d2-4c14127a1d76" (UID: "cf34ad11-7dd6-4572-b3d2-4c14127a1d76"). InnerVolumeSpecName "kube-api-access-nnhvr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.602444 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf34ad11-7dd6-4572-b3d2-4c14127a1d76-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "cf34ad11-7dd6-4572-b3d2-4c14127a1d76" (UID: "cf34ad11-7dd6-4572-b3d2-4c14127a1d76"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.615525 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf34ad11-7dd6-4572-b3d2-4c14127a1d76-scripts" (OuterVolumeSpecName: "scripts") pod "cf34ad11-7dd6-4572-b3d2-4c14127a1d76" (UID: "cf34ad11-7dd6-4572-b3d2-4c14127a1d76"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.617072 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf34ad11-7dd6-4572-b3d2-4c14127a1d76-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "cf34ad11-7dd6-4572-b3d2-4c14127a1d76" (UID: "cf34ad11-7dd6-4572-b3d2-4c14127a1d76"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.622307 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf34ad11-7dd6-4572-b3d2-4c14127a1d76-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cf34ad11-7dd6-4572-b3d2-4c14127a1d76" (UID: "cf34ad11-7dd6-4572-b3d2-4c14127a1d76"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.693787 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxv7c\" (UniqueName: \"kubernetes.io/projected/dfc6fde5-9e4c-4224-8200-0f230e127f7e-kube-api-access-fxv7c\") pod \"glance-db-sync-c5b26\" (UID: \"dfc6fde5-9e4c-4224-8200-0f230e127f7e\") " pod="openstack/glance-db-sync-c5b26" Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.694261 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfc6fde5-9e4c-4224-8200-0f230e127f7e-combined-ca-bundle\") pod \"glance-db-sync-c5b26\" (UID: \"dfc6fde5-9e4c-4224-8200-0f230e127f7e\") " pod="openstack/glance-db-sync-c5b26" Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.694320 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfc6fde5-9e4c-4224-8200-0f230e127f7e-config-data\") pod \"glance-db-sync-c5b26\" (UID: \"dfc6fde5-9e4c-4224-8200-0f230e127f7e\") " pod="openstack/glance-db-sync-c5b26" Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.694381 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/dfc6fde5-9e4c-4224-8200-0f230e127f7e-db-sync-config-data\") pod \"glance-db-sync-c5b26\" (UID: \"dfc6fde5-9e4c-4224-8200-0f230e127f7e\") " pod="openstack/glance-db-sync-c5b26" Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.694504 4890 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/cf34ad11-7dd6-4572-b3d2-4c14127a1d76-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.694518 4890 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/cf34ad11-7dd6-4572-b3d2-4c14127a1d76-swiftconf\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.694532 4890 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cf34ad11-7dd6-4572-b3d2-4c14127a1d76-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.694544 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nnhvr\" (UniqueName: \"kubernetes.io/projected/cf34ad11-7dd6-4572-b3d2-4c14127a1d76-kube-api-access-nnhvr\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.694555 4890 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf34ad11-7dd6-4572-b3d2-4c14127a1d76-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.694566 4890 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/cf34ad11-7dd6-4572-b3d2-4c14127a1d76-ring-data-devices\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.694577 4890 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/cf34ad11-7dd6-4572-b3d2-4c14127a1d76-dispersionconf\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.699363 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfc6fde5-9e4c-4224-8200-0f230e127f7e-combined-ca-bundle\") pod \"glance-db-sync-c5b26\" (UID: \"dfc6fde5-9e4c-4224-8200-0f230e127f7e\") " pod="openstack/glance-db-sync-c5b26" Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.700918 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfc6fde5-9e4c-4224-8200-0f230e127f7e-config-data\") pod \"glance-db-sync-c5b26\" (UID: \"dfc6fde5-9e4c-4224-8200-0f230e127f7e\") " pod="openstack/glance-db-sync-c5b26" Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.706634 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/dfc6fde5-9e4c-4224-8200-0f230e127f7e-db-sync-config-data\") pod \"glance-db-sync-c5b26\" (UID: \"dfc6fde5-9e4c-4224-8200-0f230e127f7e\") " pod="openstack/glance-db-sync-c5b26" Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.718415 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxv7c\" (UniqueName: \"kubernetes.io/projected/dfc6fde5-9e4c-4224-8200-0f230e127f7e-kube-api-access-fxv7c\") pod \"glance-db-sync-c5b26\" (UID: \"dfc6fde5-9e4c-4224-8200-0f230e127f7e\") " pod="openstack/glance-db-sync-c5b26" Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.832269 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-c5b26" Nov 25 15:19:24 crc kubenswrapper[4890]: I1125 15:19:24.972405 4890 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-nrbtc" podUID="524ff107-da15-4e2f-a0c8-acca0e91b27a" containerName="ovn-controller" probeResult="failure" output=< Nov 25 15:19:24 crc kubenswrapper[4890]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 25 15:19:24 crc kubenswrapper[4890]: > Nov 25 15:19:25 crc kubenswrapper[4890]: I1125 15:19:25.313069 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-vgq94" event={"ID":"21963cd4-490e-4c7f-9a9b-d93fa2a90ebc","Type":"ContainerStarted","Data":"14a4299c5a0c5640a64d516b2a7b493115f14487442b9308ee75e07c09e6b8ab"} Nov 25 15:19:25 crc kubenswrapper[4890]: I1125 15:19:25.315935 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-vs9dw" event={"ID":"cf34ad11-7dd6-4572-b3d2-4c14127a1d76","Type":"ContainerDied","Data":"ee321cbe36fab6fbdc08de675467b8b88bd16887233dd57e6eb443ed15a9617b"} Nov 25 15:19:25 crc kubenswrapper[4890]: I1125 15:19:25.315974 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee321cbe36fab6fbdc08de675467b8b88bd16887233dd57e6eb443ed15a9617b" Nov 25 15:19:25 crc kubenswrapper[4890]: I1125 15:19:25.316016 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-vs9dw" Nov 25 15:19:25 crc kubenswrapper[4890]: I1125 15:19:25.319338 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c14ce73e-1a95-4f41-b6b6-33478079806f","Type":"ContainerStarted","Data":"f5a82450da67557d7cb5285c4f393e18175e4939d9428d0b24e1f390918f2959"} Nov 25 15:19:25 crc kubenswrapper[4890]: I1125 15:19:25.320278 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:19:25 crc kubenswrapper[4890]: I1125 15:19:25.362529 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-vgq94" podStartSLOduration=2.163361892 podStartE2EDuration="6.362506491s" podCreationTimestamp="2025-11-25 15:19:19 +0000 UTC" firstStartedPulling="2025-11-25 15:19:20.330595388 +0000 UTC m=+1018.773057998" lastFinishedPulling="2025-11-25 15:19:24.529739987 +0000 UTC m=+1022.972202597" observedRunningTime="2025-11-25 15:19:25.330589913 +0000 UTC m=+1023.773052523" watchObservedRunningTime="2025-11-25 15:19:25.362506491 +0000 UTC m=+1023.804969101" Nov 25 15:19:25 crc kubenswrapper[4890]: I1125 15:19:25.368523 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=-9223371935.486269 podStartE2EDuration="1m41.368507261s" podCreationTimestamp="2025-11-25 15:17:44 +0000 UTC" firstStartedPulling="2025-11-25 15:17:46.378628362 +0000 UTC m=+924.821090972" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:19:25.36766164 +0000 UTC m=+1023.810124250" watchObservedRunningTime="2025-11-25 15:19:25.368507261 +0000 UTC m=+1023.810969871" Nov 25 15:19:25 crc kubenswrapper[4890]: I1125 15:19:25.383823 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-c5b26"] Nov 25 15:19:26 crc kubenswrapper[4890]: I1125 15:19:26.327360 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-c5b26" event={"ID":"dfc6fde5-9e4c-4224-8200-0f230e127f7e","Type":"ContainerStarted","Data":"6aa17f501d336d60c4f2eccff69fa11460eb1feac76b5ad7dc2b5e14dee3ae4a"} Nov 25 15:19:29 crc kubenswrapper[4890]: I1125 15:19:29.957047 4890 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-nrbtc" podUID="524ff107-da15-4e2f-a0c8-acca0e91b27a" containerName="ovn-controller" probeResult="failure" output=< Nov 25 15:19:29 crc kubenswrapper[4890]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 25 15:19:29 crc kubenswrapper[4890]: > Nov 25 15:19:33 crc kubenswrapper[4890]: I1125 15:19:33.386639 4890 generic.go:334] "Generic (PLEG): container finished" podID="21963cd4-490e-4c7f-9a9b-d93fa2a90ebc" containerID="14a4299c5a0c5640a64d516b2a7b493115f14487442b9308ee75e07c09e6b8ab" exitCode=0 Nov 25 15:19:33 crc kubenswrapper[4890]: I1125 15:19:33.386727 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-vgq94" event={"ID":"21963cd4-490e-4c7f-9a9b-d93fa2a90ebc","Type":"ContainerDied","Data":"14a4299c5a0c5640a64d516b2a7b493115f14487442b9308ee75e07c09e6b8ab"} Nov 25 15:19:34 crc kubenswrapper[4890]: I1125 15:19:34.268231 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6517688d-0723-4809-ba51-bcaf48524157-etc-swift\") pod \"swift-storage-0\" (UID: \"6517688d-0723-4809-ba51-bcaf48524157\") " pod="openstack/swift-storage-0" Nov 25 15:19:34 crc kubenswrapper[4890]: I1125 15:19:34.277602 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6517688d-0723-4809-ba51-bcaf48524157-etc-swift\") pod \"swift-storage-0\" (UID: \"6517688d-0723-4809-ba51-bcaf48524157\") " pod="openstack/swift-storage-0" Nov 25 15:19:34 crc kubenswrapper[4890]: I1125 15:19:34.536321 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 25 15:19:34 crc kubenswrapper[4890]: I1125 15:19:34.957309 4890 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-nrbtc" podUID="524ff107-da15-4e2f-a0c8-acca0e91b27a" containerName="ovn-controller" probeResult="failure" output=< Nov 25 15:19:34 crc kubenswrapper[4890]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 25 15:19:34 crc kubenswrapper[4890]: > Nov 25 15:19:35 crc kubenswrapper[4890]: I1125 15:19:35.789332 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:19:37 crc kubenswrapper[4890]: I1125 15:19:37.545422 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-vgq94" Nov 25 15:19:37 crc kubenswrapper[4890]: I1125 15:19:37.638813 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-njxsq\" (UniqueName: \"kubernetes.io/projected/21963cd4-490e-4c7f-9a9b-d93fa2a90ebc-kube-api-access-njxsq\") pod \"21963cd4-490e-4c7f-9a9b-d93fa2a90ebc\" (UID: \"21963cd4-490e-4c7f-9a9b-d93fa2a90ebc\") " Nov 25 15:19:37 crc kubenswrapper[4890]: I1125 15:19:37.639407 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21963cd4-490e-4c7f-9a9b-d93fa2a90ebc-config-data\") pod \"21963cd4-490e-4c7f-9a9b-d93fa2a90ebc\" (UID: \"21963cd4-490e-4c7f-9a9b-d93fa2a90ebc\") " Nov 25 15:19:37 crc kubenswrapper[4890]: I1125 15:19:37.639432 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21963cd4-490e-4c7f-9a9b-d93fa2a90ebc-combined-ca-bundle\") pod \"21963cd4-490e-4c7f-9a9b-d93fa2a90ebc\" (UID: \"21963cd4-490e-4c7f-9a9b-d93fa2a90ebc\") " Nov 25 15:19:37 crc kubenswrapper[4890]: I1125 15:19:37.658017 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21963cd4-490e-4c7f-9a9b-d93fa2a90ebc-kube-api-access-njxsq" (OuterVolumeSpecName: "kube-api-access-njxsq") pod "21963cd4-490e-4c7f-9a9b-d93fa2a90ebc" (UID: "21963cd4-490e-4c7f-9a9b-d93fa2a90ebc"). InnerVolumeSpecName "kube-api-access-njxsq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:19:37 crc kubenswrapper[4890]: I1125 15:19:37.689525 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21963cd4-490e-4c7f-9a9b-d93fa2a90ebc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "21963cd4-490e-4c7f-9a9b-d93fa2a90ebc" (UID: "21963cd4-490e-4c7f-9a9b-d93fa2a90ebc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:19:37 crc kubenswrapper[4890]: I1125 15:19:37.690834 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21963cd4-490e-4c7f-9a9b-d93fa2a90ebc-config-data" (OuterVolumeSpecName: "config-data") pod "21963cd4-490e-4c7f-9a9b-d93fa2a90ebc" (UID: "21963cd4-490e-4c7f-9a9b-d93fa2a90ebc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:19:37 crc kubenswrapper[4890]: I1125 15:19:37.747635 4890 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21963cd4-490e-4c7f-9a9b-d93fa2a90ebc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:37 crc kubenswrapper[4890]: I1125 15:19:37.747692 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-njxsq\" (UniqueName: \"kubernetes.io/projected/21963cd4-490e-4c7f-9a9b-d93fa2a90ebc-kube-api-access-njxsq\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:37 crc kubenswrapper[4890]: I1125 15:19:37.747707 4890 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21963cd4-490e-4c7f-9a9b-d93fa2a90ebc-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:38 crc kubenswrapper[4890]: I1125 15:19:38.051939 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 25 15:19:38 crc kubenswrapper[4890]: W1125 15:19:38.203837 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6517688d_0723_4809_ba51_bcaf48524157.slice/crio-d8b3e0278e28754f75aff4a8f155922fc9eee1714f552bcc985d33871ee90c54 WatchSource:0}: Error finding container d8b3e0278e28754f75aff4a8f155922fc9eee1714f552bcc985d33871ee90c54: Status 404 returned error can't find the container with id d8b3e0278e28754f75aff4a8f155922fc9eee1714f552bcc985d33871ee90c54 Nov 25 15:19:38 crc kubenswrapper[4890]: I1125 15:19:38.434417 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-vgq94" event={"ID":"21963cd4-490e-4c7f-9a9b-d93fa2a90ebc","Type":"ContainerDied","Data":"1cbc76f9f5e036ca8c4ad9b03384a46a944097733fae34917075b7bebe053992"} Nov 25 15:19:38 crc kubenswrapper[4890]: I1125 15:19:38.434777 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1cbc76f9f5e036ca8c4ad9b03384a46a944097733fae34917075b7bebe053992" Nov 25 15:19:38 crc kubenswrapper[4890]: I1125 15:19:38.434698 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-vgq94" Nov 25 15:19:38 crc kubenswrapper[4890]: I1125 15:19:38.437818 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6517688d-0723-4809-ba51-bcaf48524157","Type":"ContainerStarted","Data":"d8b3e0278e28754f75aff4a8f155922fc9eee1714f552bcc985d33871ee90c54"} Nov 25 15:19:38 crc kubenswrapper[4890]: I1125 15:19:38.758788 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f877ddd87-7g2r9"] Nov 25 15:19:38 crc kubenswrapper[4890]: E1125 15:19:38.759207 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf34ad11-7dd6-4572-b3d2-4c14127a1d76" containerName="swift-ring-rebalance" Nov 25 15:19:38 crc kubenswrapper[4890]: I1125 15:19:38.759223 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf34ad11-7dd6-4572-b3d2-4c14127a1d76" containerName="swift-ring-rebalance" Nov 25 15:19:38 crc kubenswrapper[4890]: E1125 15:19:38.759291 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21963cd4-490e-4c7f-9a9b-d93fa2a90ebc" containerName="keystone-db-sync" Nov 25 15:19:38 crc kubenswrapper[4890]: I1125 15:19:38.759300 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="21963cd4-490e-4c7f-9a9b-d93fa2a90ebc" containerName="keystone-db-sync" Nov 25 15:19:38 crc kubenswrapper[4890]: I1125 15:19:38.759512 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf34ad11-7dd6-4572-b3d2-4c14127a1d76" containerName="swift-ring-rebalance" Nov 25 15:19:38 crc kubenswrapper[4890]: I1125 15:19:38.759531 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="21963cd4-490e-4c7f-9a9b-d93fa2a90ebc" containerName="keystone-db-sync" Nov 25 15:19:38 crc kubenswrapper[4890]: I1125 15:19:38.760860 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f877ddd87-7g2r9" Nov 25 15:19:38 crc kubenswrapper[4890]: I1125 15:19:38.776136 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f877ddd87-7g2r9"] Nov 25 15:19:38 crc kubenswrapper[4890]: I1125 15:19:38.826290 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-dblm6"] Nov 25 15:19:38 crc kubenswrapper[4890]: I1125 15:19:38.828504 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dblm6" Nov 25 15:19:38 crc kubenswrapper[4890]: I1125 15:19:38.832362 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 25 15:19:38 crc kubenswrapper[4890]: I1125 15:19:38.832622 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 25 15:19:38 crc kubenswrapper[4890]: I1125 15:19:38.832735 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 25 15:19:38 crc kubenswrapper[4890]: I1125 15:19:38.832928 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-t8zql" Nov 25 15:19:38 crc kubenswrapper[4890]: I1125 15:19:38.833051 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 25 15:19:38 crc kubenswrapper[4890]: I1125 15:19:38.838432 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-dblm6"] Nov 25 15:19:38 crc kubenswrapper[4890]: I1125 15:19:38.878241 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/181307eb-d901-4255-8dab-64b90259f746-ovsdbserver-nb\") pod \"dnsmasq-dns-f877ddd87-7g2r9\" (UID: \"181307eb-d901-4255-8dab-64b90259f746\") " pod="openstack/dnsmasq-dns-f877ddd87-7g2r9" Nov 25 15:19:38 crc kubenswrapper[4890]: I1125 15:19:38.878304 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b803ac45-3f40-4188-a0cc-df469b30f1ea-fernet-keys\") pod \"keystone-bootstrap-dblm6\" (UID: \"b803ac45-3f40-4188-a0cc-df469b30f1ea\") " pod="openstack/keystone-bootstrap-dblm6" Nov 25 15:19:38 crc kubenswrapper[4890]: I1125 15:19:38.878356 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b803ac45-3f40-4188-a0cc-df469b30f1ea-scripts\") pod \"keystone-bootstrap-dblm6\" (UID: \"b803ac45-3f40-4188-a0cc-df469b30f1ea\") " pod="openstack/keystone-bootstrap-dblm6" Nov 25 15:19:38 crc kubenswrapper[4890]: I1125 15:19:38.878385 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b803ac45-3f40-4188-a0cc-df469b30f1ea-combined-ca-bundle\") pod \"keystone-bootstrap-dblm6\" (UID: \"b803ac45-3f40-4188-a0cc-df469b30f1ea\") " pod="openstack/keystone-bootstrap-dblm6" Nov 25 15:19:38 crc kubenswrapper[4890]: I1125 15:19:38.878422 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rp6l8\" (UniqueName: \"kubernetes.io/projected/b803ac45-3f40-4188-a0cc-df469b30f1ea-kube-api-access-rp6l8\") pod \"keystone-bootstrap-dblm6\" (UID: \"b803ac45-3f40-4188-a0cc-df469b30f1ea\") " pod="openstack/keystone-bootstrap-dblm6" Nov 25 15:19:38 crc kubenswrapper[4890]: I1125 15:19:38.878442 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/181307eb-d901-4255-8dab-64b90259f746-dns-svc\") pod \"dnsmasq-dns-f877ddd87-7g2r9\" (UID: \"181307eb-d901-4255-8dab-64b90259f746\") " pod="openstack/dnsmasq-dns-f877ddd87-7g2r9" Nov 25 15:19:38 crc kubenswrapper[4890]: I1125 15:19:38.878464 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/181307eb-d901-4255-8dab-64b90259f746-ovsdbserver-sb\") pod \"dnsmasq-dns-f877ddd87-7g2r9\" (UID: \"181307eb-d901-4255-8dab-64b90259f746\") " pod="openstack/dnsmasq-dns-f877ddd87-7g2r9" Nov 25 15:19:38 crc kubenswrapper[4890]: I1125 15:19:38.878483 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b803ac45-3f40-4188-a0cc-df469b30f1ea-credential-keys\") pod \"keystone-bootstrap-dblm6\" (UID: \"b803ac45-3f40-4188-a0cc-df469b30f1ea\") " pod="openstack/keystone-bootstrap-dblm6" Nov 25 15:19:38 crc kubenswrapper[4890]: I1125 15:19:38.878510 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/181307eb-d901-4255-8dab-64b90259f746-config\") pod \"dnsmasq-dns-f877ddd87-7g2r9\" (UID: \"181307eb-d901-4255-8dab-64b90259f746\") " pod="openstack/dnsmasq-dns-f877ddd87-7g2r9" Nov 25 15:19:38 crc kubenswrapper[4890]: I1125 15:19:38.878530 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhg79\" (UniqueName: \"kubernetes.io/projected/181307eb-d901-4255-8dab-64b90259f746-kube-api-access-fhg79\") pod \"dnsmasq-dns-f877ddd87-7g2r9\" (UID: \"181307eb-d901-4255-8dab-64b90259f746\") " pod="openstack/dnsmasq-dns-f877ddd87-7g2r9" Nov 25 15:19:38 crc kubenswrapper[4890]: I1125 15:19:38.878560 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b803ac45-3f40-4188-a0cc-df469b30f1ea-config-data\") pod \"keystone-bootstrap-dblm6\" (UID: \"b803ac45-3f40-4188-a0cc-df469b30f1ea\") " pod="openstack/keystone-bootstrap-dblm6" Nov 25 15:19:38 crc kubenswrapper[4890]: I1125 15:19:38.984755 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/181307eb-d901-4255-8dab-64b90259f746-config\") pod \"dnsmasq-dns-f877ddd87-7g2r9\" (UID: \"181307eb-d901-4255-8dab-64b90259f746\") " pod="openstack/dnsmasq-dns-f877ddd87-7g2r9" Nov 25 15:19:38 crc kubenswrapper[4890]: I1125 15:19:38.984834 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhg79\" (UniqueName: \"kubernetes.io/projected/181307eb-d901-4255-8dab-64b90259f746-kube-api-access-fhg79\") pod \"dnsmasq-dns-f877ddd87-7g2r9\" (UID: \"181307eb-d901-4255-8dab-64b90259f746\") " pod="openstack/dnsmasq-dns-f877ddd87-7g2r9" Nov 25 15:19:38 crc kubenswrapper[4890]: I1125 15:19:38.984866 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b803ac45-3f40-4188-a0cc-df469b30f1ea-config-data\") pod \"keystone-bootstrap-dblm6\" (UID: \"b803ac45-3f40-4188-a0cc-df469b30f1ea\") " pod="openstack/keystone-bootstrap-dblm6" Nov 25 15:19:38 crc kubenswrapper[4890]: I1125 15:19:38.984944 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/181307eb-d901-4255-8dab-64b90259f746-ovsdbserver-nb\") pod \"dnsmasq-dns-f877ddd87-7g2r9\" (UID: \"181307eb-d901-4255-8dab-64b90259f746\") " pod="openstack/dnsmasq-dns-f877ddd87-7g2r9" Nov 25 15:19:38 crc kubenswrapper[4890]: I1125 15:19:38.984973 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b803ac45-3f40-4188-a0cc-df469b30f1ea-fernet-keys\") pod \"keystone-bootstrap-dblm6\" (UID: \"b803ac45-3f40-4188-a0cc-df469b30f1ea\") " pod="openstack/keystone-bootstrap-dblm6" Nov 25 15:19:38 crc kubenswrapper[4890]: I1125 15:19:38.985026 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b803ac45-3f40-4188-a0cc-df469b30f1ea-scripts\") pod \"keystone-bootstrap-dblm6\" (UID: \"b803ac45-3f40-4188-a0cc-df469b30f1ea\") " pod="openstack/keystone-bootstrap-dblm6" Nov 25 15:19:38 crc kubenswrapper[4890]: I1125 15:19:38.985054 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b803ac45-3f40-4188-a0cc-df469b30f1ea-combined-ca-bundle\") pod \"keystone-bootstrap-dblm6\" (UID: \"b803ac45-3f40-4188-a0cc-df469b30f1ea\") " pod="openstack/keystone-bootstrap-dblm6" Nov 25 15:19:38 crc kubenswrapper[4890]: I1125 15:19:38.985099 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rp6l8\" (UniqueName: \"kubernetes.io/projected/b803ac45-3f40-4188-a0cc-df469b30f1ea-kube-api-access-rp6l8\") pod \"keystone-bootstrap-dblm6\" (UID: \"b803ac45-3f40-4188-a0cc-df469b30f1ea\") " pod="openstack/keystone-bootstrap-dblm6" Nov 25 15:19:38 crc kubenswrapper[4890]: I1125 15:19:38.985121 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/181307eb-d901-4255-8dab-64b90259f746-dns-svc\") pod \"dnsmasq-dns-f877ddd87-7g2r9\" (UID: \"181307eb-d901-4255-8dab-64b90259f746\") " pod="openstack/dnsmasq-dns-f877ddd87-7g2r9" Nov 25 15:19:38 crc kubenswrapper[4890]: I1125 15:19:38.985140 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/181307eb-d901-4255-8dab-64b90259f746-ovsdbserver-sb\") pod \"dnsmasq-dns-f877ddd87-7g2r9\" (UID: \"181307eb-d901-4255-8dab-64b90259f746\") " pod="openstack/dnsmasq-dns-f877ddd87-7g2r9" Nov 25 15:19:38 crc kubenswrapper[4890]: I1125 15:19:38.985182 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b803ac45-3f40-4188-a0cc-df469b30f1ea-credential-keys\") pod \"keystone-bootstrap-dblm6\" (UID: \"b803ac45-3f40-4188-a0cc-df469b30f1ea\") " pod="openstack/keystone-bootstrap-dblm6" Nov 25 15:19:38 crc kubenswrapper[4890]: I1125 15:19:38.989086 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/181307eb-d901-4255-8dab-64b90259f746-ovsdbserver-nb\") pod \"dnsmasq-dns-f877ddd87-7g2r9\" (UID: \"181307eb-d901-4255-8dab-64b90259f746\") " pod="openstack/dnsmasq-dns-f877ddd87-7g2r9" Nov 25 15:19:38 crc kubenswrapper[4890]: I1125 15:19:38.990690 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/181307eb-d901-4255-8dab-64b90259f746-dns-svc\") pod \"dnsmasq-dns-f877ddd87-7g2r9\" (UID: \"181307eb-d901-4255-8dab-64b90259f746\") " pod="openstack/dnsmasq-dns-f877ddd87-7g2r9" Nov 25 15:19:38 crc kubenswrapper[4890]: I1125 15:19:38.991489 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/181307eb-d901-4255-8dab-64b90259f746-config\") pod \"dnsmasq-dns-f877ddd87-7g2r9\" (UID: \"181307eb-d901-4255-8dab-64b90259f746\") " pod="openstack/dnsmasq-dns-f877ddd87-7g2r9" Nov 25 15:19:38 crc kubenswrapper[4890]: I1125 15:19:38.992554 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-wlvxc"] Nov 25 15:19:38 crc kubenswrapper[4890]: I1125 15:19:38.994748 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-wlvxc" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.005099 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.005512 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-bgrtc" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.005641 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.006636 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b803ac45-3f40-4188-a0cc-df469b30f1ea-scripts\") pod \"keystone-bootstrap-dblm6\" (UID: \"b803ac45-3f40-4188-a0cc-df469b30f1ea\") " pod="openstack/keystone-bootstrap-dblm6" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.008213 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b803ac45-3f40-4188-a0cc-df469b30f1ea-config-data\") pod \"keystone-bootstrap-dblm6\" (UID: \"b803ac45-3f40-4188-a0cc-df469b30f1ea\") " pod="openstack/keystone-bootstrap-dblm6" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.010470 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/181307eb-d901-4255-8dab-64b90259f746-ovsdbserver-sb\") pod \"dnsmasq-dns-f877ddd87-7g2r9\" (UID: \"181307eb-d901-4255-8dab-64b90259f746\") " pod="openstack/dnsmasq-dns-f877ddd87-7g2r9" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.015850 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b803ac45-3f40-4188-a0cc-df469b30f1ea-credential-keys\") pod \"keystone-bootstrap-dblm6\" (UID: \"b803ac45-3f40-4188-a0cc-df469b30f1ea\") " pod="openstack/keystone-bootstrap-dblm6" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.019700 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b803ac45-3f40-4188-a0cc-df469b30f1ea-combined-ca-bundle\") pod \"keystone-bootstrap-dblm6\" (UID: \"b803ac45-3f40-4188-a0cc-df469b30f1ea\") " pod="openstack/keystone-bootstrap-dblm6" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.024227 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b803ac45-3f40-4188-a0cc-df469b30f1ea-fernet-keys\") pod \"keystone-bootstrap-dblm6\" (UID: \"b803ac45-3f40-4188-a0cc-df469b30f1ea\") " pod="openstack/keystone-bootstrap-dblm6" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.024992 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhg79\" (UniqueName: \"kubernetes.io/projected/181307eb-d901-4255-8dab-64b90259f746-kube-api-access-fhg79\") pod \"dnsmasq-dns-f877ddd87-7g2r9\" (UID: \"181307eb-d901-4255-8dab-64b90259f746\") " pod="openstack/dnsmasq-dns-f877ddd87-7g2r9" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.029882 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rp6l8\" (UniqueName: \"kubernetes.io/projected/b803ac45-3f40-4188-a0cc-df469b30f1ea-kube-api-access-rp6l8\") pod \"keystone-bootstrap-dblm6\" (UID: \"b803ac45-3f40-4188-a0cc-df469b30f1ea\") " pod="openstack/keystone-bootstrap-dblm6" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.032622 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-wlvxc"] Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.069279 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-bf56cc7bf-tn5lx"] Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.071277 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-bf56cc7bf-tn5lx" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.074699 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.074994 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-htt2h" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.075337 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.075608 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.091789 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqntl\" (UniqueName: \"kubernetes.io/projected/d64cf59f-da15-4183-be46-13510f3f6efb-kube-api-access-tqntl\") pod \"horizon-bf56cc7bf-tn5lx\" (UID: \"d64cf59f-da15-4183-be46-13510f3f6efb\") " pod="openstack/horizon-bf56cc7bf-tn5lx" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.091846 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/449c7dd2-e3c8-48c2-8770-c1d6a77fce7b-scripts\") pod \"cinder-db-sync-wlvxc\" (UID: \"449c7dd2-e3c8-48c2-8770-c1d6a77fce7b\") " pod="openstack/cinder-db-sync-wlvxc" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.091903 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/449c7dd2-e3c8-48c2-8770-c1d6a77fce7b-db-sync-config-data\") pod \"cinder-db-sync-wlvxc\" (UID: \"449c7dd2-e3c8-48c2-8770-c1d6a77fce7b\") " pod="openstack/cinder-db-sync-wlvxc" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.091936 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d64cf59f-da15-4183-be46-13510f3f6efb-config-data\") pod \"horizon-bf56cc7bf-tn5lx\" (UID: \"d64cf59f-da15-4183-be46-13510f3f6efb\") " pod="openstack/horizon-bf56cc7bf-tn5lx" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.091968 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/449c7dd2-e3c8-48c2-8770-c1d6a77fce7b-config-data\") pod \"cinder-db-sync-wlvxc\" (UID: \"449c7dd2-e3c8-48c2-8770-c1d6a77fce7b\") " pod="openstack/cinder-db-sync-wlvxc" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.091998 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/449c7dd2-e3c8-48c2-8770-c1d6a77fce7b-etc-machine-id\") pod \"cinder-db-sync-wlvxc\" (UID: \"449c7dd2-e3c8-48c2-8770-c1d6a77fce7b\") " pod="openstack/cinder-db-sync-wlvxc" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.092054 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/449c7dd2-e3c8-48c2-8770-c1d6a77fce7b-combined-ca-bundle\") pod \"cinder-db-sync-wlvxc\" (UID: \"449c7dd2-e3c8-48c2-8770-c1d6a77fce7b\") " pod="openstack/cinder-db-sync-wlvxc" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.092084 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d64cf59f-da15-4183-be46-13510f3f6efb-horizon-secret-key\") pod \"horizon-bf56cc7bf-tn5lx\" (UID: \"d64cf59f-da15-4183-be46-13510f3f6efb\") " pod="openstack/horizon-bf56cc7bf-tn5lx" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.092128 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d64cf59f-da15-4183-be46-13510f3f6efb-logs\") pod \"horizon-bf56cc7bf-tn5lx\" (UID: \"d64cf59f-da15-4183-be46-13510f3f6efb\") " pod="openstack/horizon-bf56cc7bf-tn5lx" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.092156 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d64cf59f-da15-4183-be46-13510f3f6efb-scripts\") pod \"horizon-bf56cc7bf-tn5lx\" (UID: \"d64cf59f-da15-4183-be46-13510f3f6efb\") " pod="openstack/horizon-bf56cc7bf-tn5lx" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.092217 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q98k8\" (UniqueName: \"kubernetes.io/projected/449c7dd2-e3c8-48c2-8770-c1d6a77fce7b-kube-api-access-q98k8\") pod \"cinder-db-sync-wlvxc\" (UID: \"449c7dd2-e3c8-48c2-8770-c1d6a77fce7b\") " pod="openstack/cinder-db-sync-wlvxc" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.096127 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-bf56cc7bf-tn5lx"] Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.114465 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f877ddd87-7g2r9" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.175350 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f877ddd87-7g2r9"] Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.175738 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dblm6" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.194394 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqntl\" (UniqueName: \"kubernetes.io/projected/d64cf59f-da15-4183-be46-13510f3f6efb-kube-api-access-tqntl\") pod \"horizon-bf56cc7bf-tn5lx\" (UID: \"d64cf59f-da15-4183-be46-13510f3f6efb\") " pod="openstack/horizon-bf56cc7bf-tn5lx" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.194446 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/449c7dd2-e3c8-48c2-8770-c1d6a77fce7b-scripts\") pod \"cinder-db-sync-wlvxc\" (UID: \"449c7dd2-e3c8-48c2-8770-c1d6a77fce7b\") " pod="openstack/cinder-db-sync-wlvxc" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.194492 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/449c7dd2-e3c8-48c2-8770-c1d6a77fce7b-db-sync-config-data\") pod \"cinder-db-sync-wlvxc\" (UID: \"449c7dd2-e3c8-48c2-8770-c1d6a77fce7b\") " pod="openstack/cinder-db-sync-wlvxc" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.194516 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d64cf59f-da15-4183-be46-13510f3f6efb-config-data\") pod \"horizon-bf56cc7bf-tn5lx\" (UID: \"d64cf59f-da15-4183-be46-13510f3f6efb\") " pod="openstack/horizon-bf56cc7bf-tn5lx" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.194538 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/449c7dd2-e3c8-48c2-8770-c1d6a77fce7b-config-data\") pod \"cinder-db-sync-wlvxc\" (UID: \"449c7dd2-e3c8-48c2-8770-c1d6a77fce7b\") " pod="openstack/cinder-db-sync-wlvxc" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.194555 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/449c7dd2-e3c8-48c2-8770-c1d6a77fce7b-etc-machine-id\") pod \"cinder-db-sync-wlvxc\" (UID: \"449c7dd2-e3c8-48c2-8770-c1d6a77fce7b\") " pod="openstack/cinder-db-sync-wlvxc" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.194596 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/449c7dd2-e3c8-48c2-8770-c1d6a77fce7b-combined-ca-bundle\") pod \"cinder-db-sync-wlvxc\" (UID: \"449c7dd2-e3c8-48c2-8770-c1d6a77fce7b\") " pod="openstack/cinder-db-sync-wlvxc" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.194616 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d64cf59f-da15-4183-be46-13510f3f6efb-horizon-secret-key\") pod \"horizon-bf56cc7bf-tn5lx\" (UID: \"d64cf59f-da15-4183-be46-13510f3f6efb\") " pod="openstack/horizon-bf56cc7bf-tn5lx" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.194647 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d64cf59f-da15-4183-be46-13510f3f6efb-logs\") pod \"horizon-bf56cc7bf-tn5lx\" (UID: \"d64cf59f-da15-4183-be46-13510f3f6efb\") " pod="openstack/horizon-bf56cc7bf-tn5lx" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.194664 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d64cf59f-da15-4183-be46-13510f3f6efb-scripts\") pod \"horizon-bf56cc7bf-tn5lx\" (UID: \"d64cf59f-da15-4183-be46-13510f3f6efb\") " pod="openstack/horizon-bf56cc7bf-tn5lx" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.194688 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q98k8\" (UniqueName: \"kubernetes.io/projected/449c7dd2-e3c8-48c2-8770-c1d6a77fce7b-kube-api-access-q98k8\") pod \"cinder-db-sync-wlvxc\" (UID: \"449c7dd2-e3c8-48c2-8770-c1d6a77fce7b\") " pod="openstack/cinder-db-sync-wlvxc" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.196721 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/449c7dd2-e3c8-48c2-8770-c1d6a77fce7b-etc-machine-id\") pod \"cinder-db-sync-wlvxc\" (UID: \"449c7dd2-e3c8-48c2-8770-c1d6a77fce7b\") " pod="openstack/cinder-db-sync-wlvxc" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.198823 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d64cf59f-da15-4183-be46-13510f3f6efb-config-data\") pod \"horizon-bf56cc7bf-tn5lx\" (UID: \"d64cf59f-da15-4183-be46-13510f3f6efb\") " pod="openstack/horizon-bf56cc7bf-tn5lx" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.201246 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d64cf59f-da15-4183-be46-13510f3f6efb-logs\") pod \"horizon-bf56cc7bf-tn5lx\" (UID: \"d64cf59f-da15-4183-be46-13510f3f6efb\") " pod="openstack/horizon-bf56cc7bf-tn5lx" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.212153 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d64cf59f-da15-4183-be46-13510f3f6efb-scripts\") pod \"horizon-bf56cc7bf-tn5lx\" (UID: \"d64cf59f-da15-4183-be46-13510f3f6efb\") " pod="openstack/horizon-bf56cc7bf-tn5lx" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.218484 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/449c7dd2-e3c8-48c2-8770-c1d6a77fce7b-scripts\") pod \"cinder-db-sync-wlvxc\" (UID: \"449c7dd2-e3c8-48c2-8770-c1d6a77fce7b\") " pod="openstack/cinder-db-sync-wlvxc" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.219053 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/449c7dd2-e3c8-48c2-8770-c1d6a77fce7b-config-data\") pod \"cinder-db-sync-wlvxc\" (UID: \"449c7dd2-e3c8-48c2-8770-c1d6a77fce7b\") " pod="openstack/cinder-db-sync-wlvxc" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.224533 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-mg8z4"] Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.226068 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-mg8z4" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.234720 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d64cf59f-da15-4183-be46-13510f3f6efb-horizon-secret-key\") pod \"horizon-bf56cc7bf-tn5lx\" (UID: \"d64cf59f-da15-4183-be46-13510f3f6efb\") " pod="openstack/horizon-bf56cc7bf-tn5lx" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.237514 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.240272 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-mg8z4"] Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.248069 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/449c7dd2-e3c8-48c2-8770-c1d6a77fce7b-combined-ca-bundle\") pod \"cinder-db-sync-wlvxc\" (UID: \"449c7dd2-e3c8-48c2-8770-c1d6a77fce7b\") " pod="openstack/cinder-db-sync-wlvxc" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.252888 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/449c7dd2-e3c8-48c2-8770-c1d6a77fce7b-db-sync-config-data\") pod \"cinder-db-sync-wlvxc\" (UID: \"449c7dd2-e3c8-48c2-8770-c1d6a77fce7b\") " pod="openstack/cinder-db-sync-wlvxc" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.253353 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqntl\" (UniqueName: \"kubernetes.io/projected/d64cf59f-da15-4183-be46-13510f3f6efb-kube-api-access-tqntl\") pod \"horizon-bf56cc7bf-tn5lx\" (UID: \"d64cf59f-da15-4183-be46-13510f3f6efb\") " pod="openstack/horizon-bf56cc7bf-tn5lx" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.237803 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-xzflj" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.252703 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.269058 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q98k8\" (UniqueName: \"kubernetes.io/projected/449c7dd2-e3c8-48c2-8770-c1d6a77fce7b-kube-api-access-q98k8\") pod \"cinder-db-sync-wlvxc\" (UID: \"449c7dd2-e3c8-48c2-8770-c1d6a77fce7b\") " pod="openstack/cinder-db-sync-wlvxc" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.302418 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-bf56cc7bf-tn5lx" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.418653 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-j7w9k"] Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.419015 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hj6k7\" (UniqueName: \"kubernetes.io/projected/13af51fd-267a-4b92-b08b-49db0d79a866-kube-api-access-hj6k7\") pod \"neutron-db-sync-mg8z4\" (UID: \"13af51fd-267a-4b92-b08b-49db0d79a866\") " pod="openstack/neutron-db-sync-mg8z4" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.419091 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/13af51fd-267a-4b92-b08b-49db0d79a866-config\") pod \"neutron-db-sync-mg8z4\" (UID: \"13af51fd-267a-4b92-b08b-49db0d79a866\") " pod="openstack/neutron-db-sync-mg8z4" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.419249 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13af51fd-267a-4b92-b08b-49db0d79a866-combined-ca-bundle\") pod \"neutron-db-sync-mg8z4\" (UID: \"13af51fd-267a-4b92-b08b-49db0d79a866\") " pod="openstack/neutron-db-sync-mg8z4" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.421771 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-j7w9k" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.425736 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.431135 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.439968 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-bgppv" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.463245 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-68dcc9cf6f-lm2sb"] Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.474093 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68dcc9cf6f-lm2sb" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.521094 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hj6k7\" (UniqueName: \"kubernetes.io/projected/13af51fd-267a-4b92-b08b-49db0d79a866-kube-api-access-hj6k7\") pod \"neutron-db-sync-mg8z4\" (UID: \"13af51fd-267a-4b92-b08b-49db0d79a866\") " pod="openstack/neutron-db-sync-mg8z4" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.521132 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/13af51fd-267a-4b92-b08b-49db0d79a866-config\") pod \"neutron-db-sync-mg8z4\" (UID: \"13af51fd-267a-4b92-b08b-49db0d79a866\") " pod="openstack/neutron-db-sync-mg8z4" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.521189 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13af51fd-267a-4b92-b08b-49db0d79a866-combined-ca-bundle\") pod \"neutron-db-sync-mg8z4\" (UID: \"13af51fd-267a-4b92-b08b-49db0d79a866\") " pod="openstack/neutron-db-sync-mg8z4" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.526060 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7f7df64d4f-k6jfd"] Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.534089 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13af51fd-267a-4b92-b08b-49db0d79a866-combined-ca-bundle\") pod \"neutron-db-sync-mg8z4\" (UID: \"13af51fd-267a-4b92-b08b-49db0d79a866\") " pod="openstack/neutron-db-sync-mg8z4" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.538660 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-wlvxc" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.539625 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/13af51fd-267a-4b92-b08b-49db0d79a866-config\") pod \"neutron-db-sync-mg8z4\" (UID: \"13af51fd-267a-4b92-b08b-49db0d79a866\") " pod="openstack/neutron-db-sync-mg8z4" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.549782 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-c5b26" event={"ID":"dfc6fde5-9e4c-4224-8200-0f230e127f7e","Type":"ContainerStarted","Data":"30fb995c1452e7fb88486c70e38cdf87240b7d5488b0d1aaa70d5a725f84c6bf"} Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.549838 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.551916 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f7df64d4f-k6jfd" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.553139 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-j7w9k"] Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.553203 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-68dcc9cf6f-lm2sb"] Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.553278 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.558072 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-f8ctf"] Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.559423 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-f8ctf" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.568690 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7f7df64d4f-k6jfd"] Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.570389 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.570699 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.570696 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.570885 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hj6k7\" (UniqueName: \"kubernetes.io/projected/13af51fd-267a-4b92-b08b-49db0d79a866-kube-api-access-hj6k7\") pod \"neutron-db-sync-mg8z4\" (UID: \"13af51fd-267a-4b92-b08b-49db0d79a866\") " pod="openstack/neutron-db-sync-mg8z4" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.571003 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-44nwj" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.583991 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.589566 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-f8ctf"] Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.604997 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-c5b26" podStartSLOduration=2.732364938 podStartE2EDuration="15.604976491s" podCreationTimestamp="2025-11-25 15:19:24 +0000 UTC" firstStartedPulling="2025-11-25 15:19:25.387579708 +0000 UTC m=+1023.830042318" lastFinishedPulling="2025-11-25 15:19:38.260191261 +0000 UTC m=+1036.702653871" observedRunningTime="2025-11-25 15:19:39.583582587 +0000 UTC m=+1038.026045217" watchObservedRunningTime="2025-11-25 15:19:39.604976491 +0000 UTC m=+1038.047439101" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.625060 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/997a078a-52bd-4eb2-8fde-f80d05d9b18c-scripts\") pod \"placement-db-sync-j7w9k\" (UID: \"997a078a-52bd-4eb2-8fde-f80d05d9b18c\") " pod="openstack/placement-db-sync-j7w9k" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.625101 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d760f8c7-b47e-455f-81c3-0ce064da5c96-ovsdbserver-sb\") pod \"dnsmasq-dns-68dcc9cf6f-lm2sb\" (UID: \"d760f8c7-b47e-455f-81c3-0ce064da5c96\") " pod="openstack/dnsmasq-dns-68dcc9cf6f-lm2sb" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.625140 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b246a631-f563-4f3f-aef3-caa0d92520c6-logs\") pod \"horizon-7f7df64d4f-k6jfd\" (UID: \"b246a631-f563-4f3f-aef3-caa0d92520c6\") " pod="openstack/horizon-7f7df64d4f-k6jfd" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.626600 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/997a078a-52bd-4eb2-8fde-f80d05d9b18c-logs\") pod \"placement-db-sync-j7w9k\" (UID: \"997a078a-52bd-4eb2-8fde-f80d05d9b18c\") " pod="openstack/placement-db-sync-j7w9k" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.626636 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/997a078a-52bd-4eb2-8fde-f80d05d9b18c-combined-ca-bundle\") pod \"placement-db-sync-j7w9k\" (UID: \"997a078a-52bd-4eb2-8fde-f80d05d9b18c\") " pod="openstack/placement-db-sync-j7w9k" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.626656 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d760f8c7-b47e-455f-81c3-0ce064da5c96-dns-svc\") pod \"dnsmasq-dns-68dcc9cf6f-lm2sb\" (UID: \"d760f8c7-b47e-455f-81c3-0ce064da5c96\") " pod="openstack/dnsmasq-dns-68dcc9cf6f-lm2sb" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.626707 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d760f8c7-b47e-455f-81c3-0ce064da5c96-config\") pod \"dnsmasq-dns-68dcc9cf6f-lm2sb\" (UID: \"d760f8c7-b47e-455f-81c3-0ce064da5c96\") " pod="openstack/dnsmasq-dns-68dcc9cf6f-lm2sb" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.626730 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/997a078a-52bd-4eb2-8fde-f80d05d9b18c-config-data\") pod \"placement-db-sync-j7w9k\" (UID: \"997a078a-52bd-4eb2-8fde-f80d05d9b18c\") " pod="openstack/placement-db-sync-j7w9k" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.626747 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b246a631-f563-4f3f-aef3-caa0d92520c6-scripts\") pod \"horizon-7f7df64d4f-k6jfd\" (UID: \"b246a631-f563-4f3f-aef3-caa0d92520c6\") " pod="openstack/horizon-7f7df64d4f-k6jfd" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.626762 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4q26\" (UniqueName: \"kubernetes.io/projected/d760f8c7-b47e-455f-81c3-0ce064da5c96-kube-api-access-p4q26\") pod \"dnsmasq-dns-68dcc9cf6f-lm2sb\" (UID: \"d760f8c7-b47e-455f-81c3-0ce064da5c96\") " pod="openstack/dnsmasq-dns-68dcc9cf6f-lm2sb" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.626799 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d760f8c7-b47e-455f-81c3-0ce064da5c96-ovsdbserver-nb\") pod \"dnsmasq-dns-68dcc9cf6f-lm2sb\" (UID: \"d760f8c7-b47e-455f-81c3-0ce064da5c96\") " pod="openstack/dnsmasq-dns-68dcc9cf6f-lm2sb" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.626814 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b246a631-f563-4f3f-aef3-caa0d92520c6-config-data\") pod \"horizon-7f7df64d4f-k6jfd\" (UID: \"b246a631-f563-4f3f-aef3-caa0d92520c6\") " pod="openstack/horizon-7f7df64d4f-k6jfd" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.626835 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pw2ct\" (UniqueName: \"kubernetes.io/projected/997a078a-52bd-4eb2-8fde-f80d05d9b18c-kube-api-access-pw2ct\") pod \"placement-db-sync-j7w9k\" (UID: \"997a078a-52bd-4eb2-8fde-f80d05d9b18c\") " pod="openstack/placement-db-sync-j7w9k" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.626870 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skrsm\" (UniqueName: \"kubernetes.io/projected/b246a631-f563-4f3f-aef3-caa0d92520c6-kube-api-access-skrsm\") pod \"horizon-7f7df64d4f-k6jfd\" (UID: \"b246a631-f563-4f3f-aef3-caa0d92520c6\") " pod="openstack/horizon-7f7df64d4f-k6jfd" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.626895 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b246a631-f563-4f3f-aef3-caa0d92520c6-horizon-secret-key\") pod \"horizon-7f7df64d4f-k6jfd\" (UID: \"b246a631-f563-4f3f-aef3-caa0d92520c6\") " pod="openstack/horizon-7f7df64d4f-k6jfd" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.686340 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-mg8z4" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.728514 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fdcacd06-b857-46cf-8516-345e271cf882-db-sync-config-data\") pod \"barbican-db-sync-f8ctf\" (UID: \"fdcacd06-b857-46cf-8516-345e271cf882\") " pod="openstack/barbican-db-sync-f8ctf" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.728553 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdcacd06-b857-46cf-8516-345e271cf882-combined-ca-bundle\") pod \"barbican-db-sync-f8ctf\" (UID: \"fdcacd06-b857-46cf-8516-345e271cf882\") " pod="openstack/barbican-db-sync-f8ctf" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.728579 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a886f74-d737-46b2-9647-232deec47e87-scripts\") pod \"ceilometer-0\" (UID: \"1a886f74-d737-46b2-9647-232deec47e87\") " pod="openstack/ceilometer-0" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.728602 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d760f8c7-b47e-455f-81c3-0ce064da5c96-config\") pod \"dnsmasq-dns-68dcc9cf6f-lm2sb\" (UID: \"d760f8c7-b47e-455f-81c3-0ce064da5c96\") " pod="openstack/dnsmasq-dns-68dcc9cf6f-lm2sb" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.728624 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/997a078a-52bd-4eb2-8fde-f80d05d9b18c-config-data\") pod \"placement-db-sync-j7w9k\" (UID: \"997a078a-52bd-4eb2-8fde-f80d05d9b18c\") " pod="openstack/placement-db-sync-j7w9k" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.728644 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b246a631-f563-4f3f-aef3-caa0d92520c6-scripts\") pod \"horizon-7f7df64d4f-k6jfd\" (UID: \"b246a631-f563-4f3f-aef3-caa0d92520c6\") " pod="openstack/horizon-7f7df64d4f-k6jfd" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.728663 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4q26\" (UniqueName: \"kubernetes.io/projected/d760f8c7-b47e-455f-81c3-0ce064da5c96-kube-api-access-p4q26\") pod \"dnsmasq-dns-68dcc9cf6f-lm2sb\" (UID: \"d760f8c7-b47e-455f-81c3-0ce064da5c96\") " pod="openstack/dnsmasq-dns-68dcc9cf6f-lm2sb" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.728691 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d760f8c7-b47e-455f-81c3-0ce064da5c96-ovsdbserver-nb\") pod \"dnsmasq-dns-68dcc9cf6f-lm2sb\" (UID: \"d760f8c7-b47e-455f-81c3-0ce064da5c96\") " pod="openstack/dnsmasq-dns-68dcc9cf6f-lm2sb" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.728707 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b246a631-f563-4f3f-aef3-caa0d92520c6-config-data\") pod \"horizon-7f7df64d4f-k6jfd\" (UID: \"b246a631-f563-4f3f-aef3-caa0d92520c6\") " pod="openstack/horizon-7f7df64d4f-k6jfd" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.728728 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pw2ct\" (UniqueName: \"kubernetes.io/projected/997a078a-52bd-4eb2-8fde-f80d05d9b18c-kube-api-access-pw2ct\") pod \"placement-db-sync-j7w9k\" (UID: \"997a078a-52bd-4eb2-8fde-f80d05d9b18c\") " pod="openstack/placement-db-sync-j7w9k" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.728755 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a886f74-d737-46b2-9647-232deec47e87-run-httpd\") pod \"ceilometer-0\" (UID: \"1a886f74-d737-46b2-9647-232deec47e87\") " pod="openstack/ceilometer-0" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.728772 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skrsm\" (UniqueName: \"kubernetes.io/projected/b246a631-f563-4f3f-aef3-caa0d92520c6-kube-api-access-skrsm\") pod \"horizon-7f7df64d4f-k6jfd\" (UID: \"b246a631-f563-4f3f-aef3-caa0d92520c6\") " pod="openstack/horizon-7f7df64d4f-k6jfd" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.728798 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b246a631-f563-4f3f-aef3-caa0d92520c6-horizon-secret-key\") pod \"horizon-7f7df64d4f-k6jfd\" (UID: \"b246a631-f563-4f3f-aef3-caa0d92520c6\") " pod="openstack/horizon-7f7df64d4f-k6jfd" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.728828 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a886f74-d737-46b2-9647-232deec47e87-config-data\") pod \"ceilometer-0\" (UID: \"1a886f74-d737-46b2-9647-232deec47e87\") " pod="openstack/ceilometer-0" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.728847 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1a886f74-d737-46b2-9647-232deec47e87-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1a886f74-d737-46b2-9647-232deec47e87\") " pod="openstack/ceilometer-0" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.728865 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/997a078a-52bd-4eb2-8fde-f80d05d9b18c-scripts\") pod \"placement-db-sync-j7w9k\" (UID: \"997a078a-52bd-4eb2-8fde-f80d05d9b18c\") " pod="openstack/placement-db-sync-j7w9k" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.728881 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d760f8c7-b47e-455f-81c3-0ce064da5c96-ovsdbserver-sb\") pod \"dnsmasq-dns-68dcc9cf6f-lm2sb\" (UID: \"d760f8c7-b47e-455f-81c3-0ce064da5c96\") " pod="openstack/dnsmasq-dns-68dcc9cf6f-lm2sb" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.728899 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rk2gl\" (UniqueName: \"kubernetes.io/projected/1a886f74-d737-46b2-9647-232deec47e87-kube-api-access-rk2gl\") pod \"ceilometer-0\" (UID: \"1a886f74-d737-46b2-9647-232deec47e87\") " pod="openstack/ceilometer-0" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.728923 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a886f74-d737-46b2-9647-232deec47e87-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1a886f74-d737-46b2-9647-232deec47e87\") " pod="openstack/ceilometer-0" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.728942 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szt2s\" (UniqueName: \"kubernetes.io/projected/fdcacd06-b857-46cf-8516-345e271cf882-kube-api-access-szt2s\") pod \"barbican-db-sync-f8ctf\" (UID: \"fdcacd06-b857-46cf-8516-345e271cf882\") " pod="openstack/barbican-db-sync-f8ctf" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.728964 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b246a631-f563-4f3f-aef3-caa0d92520c6-logs\") pod \"horizon-7f7df64d4f-k6jfd\" (UID: \"b246a631-f563-4f3f-aef3-caa0d92520c6\") " pod="openstack/horizon-7f7df64d4f-k6jfd" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.728989 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a886f74-d737-46b2-9647-232deec47e87-log-httpd\") pod \"ceilometer-0\" (UID: \"1a886f74-d737-46b2-9647-232deec47e87\") " pod="openstack/ceilometer-0" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.729007 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/997a078a-52bd-4eb2-8fde-f80d05d9b18c-logs\") pod \"placement-db-sync-j7w9k\" (UID: \"997a078a-52bd-4eb2-8fde-f80d05d9b18c\") " pod="openstack/placement-db-sync-j7w9k" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.729021 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/997a078a-52bd-4eb2-8fde-f80d05d9b18c-combined-ca-bundle\") pod \"placement-db-sync-j7w9k\" (UID: \"997a078a-52bd-4eb2-8fde-f80d05d9b18c\") " pod="openstack/placement-db-sync-j7w9k" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.729040 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d760f8c7-b47e-455f-81c3-0ce064da5c96-dns-svc\") pod \"dnsmasq-dns-68dcc9cf6f-lm2sb\" (UID: \"d760f8c7-b47e-455f-81c3-0ce064da5c96\") " pod="openstack/dnsmasq-dns-68dcc9cf6f-lm2sb" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.732391 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d760f8c7-b47e-455f-81c3-0ce064da5c96-config\") pod \"dnsmasq-dns-68dcc9cf6f-lm2sb\" (UID: \"d760f8c7-b47e-455f-81c3-0ce064da5c96\") " pod="openstack/dnsmasq-dns-68dcc9cf6f-lm2sb" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.734000 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d760f8c7-b47e-455f-81c3-0ce064da5c96-dns-svc\") pod \"dnsmasq-dns-68dcc9cf6f-lm2sb\" (UID: \"d760f8c7-b47e-455f-81c3-0ce064da5c96\") " pod="openstack/dnsmasq-dns-68dcc9cf6f-lm2sb" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.734565 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b246a631-f563-4f3f-aef3-caa0d92520c6-logs\") pod \"horizon-7f7df64d4f-k6jfd\" (UID: \"b246a631-f563-4f3f-aef3-caa0d92520c6\") " pod="openstack/horizon-7f7df64d4f-k6jfd" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.735286 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d760f8c7-b47e-455f-81c3-0ce064da5c96-ovsdbserver-nb\") pod \"dnsmasq-dns-68dcc9cf6f-lm2sb\" (UID: \"d760f8c7-b47e-455f-81c3-0ce064da5c96\") " pod="openstack/dnsmasq-dns-68dcc9cf6f-lm2sb" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.735518 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/997a078a-52bd-4eb2-8fde-f80d05d9b18c-logs\") pod \"placement-db-sync-j7w9k\" (UID: \"997a078a-52bd-4eb2-8fde-f80d05d9b18c\") " pod="openstack/placement-db-sync-j7w9k" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.740818 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b246a631-f563-4f3f-aef3-caa0d92520c6-scripts\") pod \"horizon-7f7df64d4f-k6jfd\" (UID: \"b246a631-f563-4f3f-aef3-caa0d92520c6\") " pod="openstack/horizon-7f7df64d4f-k6jfd" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.741326 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b246a631-f563-4f3f-aef3-caa0d92520c6-config-data\") pod \"horizon-7f7df64d4f-k6jfd\" (UID: \"b246a631-f563-4f3f-aef3-caa0d92520c6\") " pod="openstack/horizon-7f7df64d4f-k6jfd" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.743810 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/997a078a-52bd-4eb2-8fde-f80d05d9b18c-config-data\") pod \"placement-db-sync-j7w9k\" (UID: \"997a078a-52bd-4eb2-8fde-f80d05d9b18c\") " pod="openstack/placement-db-sync-j7w9k" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.744777 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b246a631-f563-4f3f-aef3-caa0d92520c6-horizon-secret-key\") pod \"horizon-7f7df64d4f-k6jfd\" (UID: \"b246a631-f563-4f3f-aef3-caa0d92520c6\") " pod="openstack/horizon-7f7df64d4f-k6jfd" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.748599 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d760f8c7-b47e-455f-81c3-0ce064da5c96-ovsdbserver-sb\") pod \"dnsmasq-dns-68dcc9cf6f-lm2sb\" (UID: \"d760f8c7-b47e-455f-81c3-0ce064da5c96\") " pod="openstack/dnsmasq-dns-68dcc9cf6f-lm2sb" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.751530 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/997a078a-52bd-4eb2-8fde-f80d05d9b18c-combined-ca-bundle\") pod \"placement-db-sync-j7w9k\" (UID: \"997a078a-52bd-4eb2-8fde-f80d05d9b18c\") " pod="openstack/placement-db-sync-j7w9k" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.760524 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skrsm\" (UniqueName: \"kubernetes.io/projected/b246a631-f563-4f3f-aef3-caa0d92520c6-kube-api-access-skrsm\") pod \"horizon-7f7df64d4f-k6jfd\" (UID: \"b246a631-f563-4f3f-aef3-caa0d92520c6\") " pod="openstack/horizon-7f7df64d4f-k6jfd" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.764776 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/997a078a-52bd-4eb2-8fde-f80d05d9b18c-scripts\") pod \"placement-db-sync-j7w9k\" (UID: \"997a078a-52bd-4eb2-8fde-f80d05d9b18c\") " pod="openstack/placement-db-sync-j7w9k" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.778243 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f7df64d4f-k6jfd" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.782038 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pw2ct\" (UniqueName: \"kubernetes.io/projected/997a078a-52bd-4eb2-8fde-f80d05d9b18c-kube-api-access-pw2ct\") pod \"placement-db-sync-j7w9k\" (UID: \"997a078a-52bd-4eb2-8fde-f80d05d9b18c\") " pod="openstack/placement-db-sync-j7w9k" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.807883 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4q26\" (UniqueName: \"kubernetes.io/projected/d760f8c7-b47e-455f-81c3-0ce064da5c96-kube-api-access-p4q26\") pod \"dnsmasq-dns-68dcc9cf6f-lm2sb\" (UID: \"d760f8c7-b47e-455f-81c3-0ce064da5c96\") " pod="openstack/dnsmasq-dns-68dcc9cf6f-lm2sb" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.832946 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a886f74-d737-46b2-9647-232deec47e87-scripts\") pod \"ceilometer-0\" (UID: \"1a886f74-d737-46b2-9647-232deec47e87\") " pod="openstack/ceilometer-0" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.834303 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a886f74-d737-46b2-9647-232deec47e87-run-httpd\") pod \"ceilometer-0\" (UID: \"1a886f74-d737-46b2-9647-232deec47e87\") " pod="openstack/ceilometer-0" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.834380 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a886f74-d737-46b2-9647-232deec47e87-config-data\") pod \"ceilometer-0\" (UID: \"1a886f74-d737-46b2-9647-232deec47e87\") " pod="openstack/ceilometer-0" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.834410 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1a886f74-d737-46b2-9647-232deec47e87-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1a886f74-d737-46b2-9647-232deec47e87\") " pod="openstack/ceilometer-0" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.834445 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rk2gl\" (UniqueName: \"kubernetes.io/projected/1a886f74-d737-46b2-9647-232deec47e87-kube-api-access-rk2gl\") pod \"ceilometer-0\" (UID: \"1a886f74-d737-46b2-9647-232deec47e87\") " pod="openstack/ceilometer-0" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.834480 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a886f74-d737-46b2-9647-232deec47e87-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1a886f74-d737-46b2-9647-232deec47e87\") " pod="openstack/ceilometer-0" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.834503 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szt2s\" (UniqueName: \"kubernetes.io/projected/fdcacd06-b857-46cf-8516-345e271cf882-kube-api-access-szt2s\") pod \"barbican-db-sync-f8ctf\" (UID: \"fdcacd06-b857-46cf-8516-345e271cf882\") " pod="openstack/barbican-db-sync-f8ctf" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.834549 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a886f74-d737-46b2-9647-232deec47e87-log-httpd\") pod \"ceilometer-0\" (UID: \"1a886f74-d737-46b2-9647-232deec47e87\") " pod="openstack/ceilometer-0" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.834608 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fdcacd06-b857-46cf-8516-345e271cf882-db-sync-config-data\") pod \"barbican-db-sync-f8ctf\" (UID: \"fdcacd06-b857-46cf-8516-345e271cf882\") " pod="openstack/barbican-db-sync-f8ctf" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.834627 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdcacd06-b857-46cf-8516-345e271cf882-combined-ca-bundle\") pod \"barbican-db-sync-f8ctf\" (UID: \"fdcacd06-b857-46cf-8516-345e271cf882\") " pod="openstack/barbican-db-sync-f8ctf" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.836107 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a886f74-d737-46b2-9647-232deec47e87-log-httpd\") pod \"ceilometer-0\" (UID: \"1a886f74-d737-46b2-9647-232deec47e87\") " pod="openstack/ceilometer-0" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.836753 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a886f74-d737-46b2-9647-232deec47e87-run-httpd\") pod \"ceilometer-0\" (UID: \"1a886f74-d737-46b2-9647-232deec47e87\") " pod="openstack/ceilometer-0" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.846495 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a886f74-d737-46b2-9647-232deec47e87-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1a886f74-d737-46b2-9647-232deec47e87\") " pod="openstack/ceilometer-0" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.846844 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fdcacd06-b857-46cf-8516-345e271cf882-db-sync-config-data\") pod \"barbican-db-sync-f8ctf\" (UID: \"fdcacd06-b857-46cf-8516-345e271cf882\") " pod="openstack/barbican-db-sync-f8ctf" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.847259 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1a886f74-d737-46b2-9647-232deec47e87-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1a886f74-d737-46b2-9647-232deec47e87\") " pod="openstack/ceilometer-0" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.848196 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a886f74-d737-46b2-9647-232deec47e87-scripts\") pod \"ceilometer-0\" (UID: \"1a886f74-d737-46b2-9647-232deec47e87\") " pod="openstack/ceilometer-0" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.867907 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szt2s\" (UniqueName: \"kubernetes.io/projected/fdcacd06-b857-46cf-8516-345e271cf882-kube-api-access-szt2s\") pod \"barbican-db-sync-f8ctf\" (UID: \"fdcacd06-b857-46cf-8516-345e271cf882\") " pod="openstack/barbican-db-sync-f8ctf" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.872389 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a886f74-d737-46b2-9647-232deec47e87-config-data\") pod \"ceilometer-0\" (UID: \"1a886f74-d737-46b2-9647-232deec47e87\") " pod="openstack/ceilometer-0" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.874821 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdcacd06-b857-46cf-8516-345e271cf882-combined-ca-bundle\") pod \"barbican-db-sync-f8ctf\" (UID: \"fdcacd06-b857-46cf-8516-345e271cf882\") " pod="openstack/barbican-db-sync-f8ctf" Nov 25 15:19:39 crc kubenswrapper[4890]: I1125 15:19:39.877664 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rk2gl\" (UniqueName: \"kubernetes.io/projected/1a886f74-d737-46b2-9647-232deec47e87-kube-api-access-rk2gl\") pod \"ceilometer-0\" (UID: \"1a886f74-d737-46b2-9647-232deec47e87\") " pod="openstack/ceilometer-0" Nov 25 15:19:40 crc kubenswrapper[4890]: I1125 15:19:40.003413 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-5tg6b" Nov 25 15:19:40 crc kubenswrapper[4890]: I1125 15:19:40.036613 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-j7w9k" Nov 25 15:19:40 crc kubenswrapper[4890]: I1125 15:19:40.054247 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68dcc9cf6f-lm2sb" Nov 25 15:19:40 crc kubenswrapper[4890]: I1125 15:19:40.061309 4890 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-nrbtc" podUID="524ff107-da15-4e2f-a0c8-acca0e91b27a" containerName="ovn-controller" probeResult="failure" output=< Nov 25 15:19:40 crc kubenswrapper[4890]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 25 15:19:40 crc kubenswrapper[4890]: > Nov 25 15:19:40 crc kubenswrapper[4890]: I1125 15:19:40.095045 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-5tg6b" Nov 25 15:19:40 crc kubenswrapper[4890]: I1125 15:19:40.113035 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f877ddd87-7g2r9"] Nov 25 15:19:40 crc kubenswrapper[4890]: I1125 15:19:40.115331 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 15:19:40 crc kubenswrapper[4890]: I1125 15:19:40.116579 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-f8ctf" Nov 25 15:19:40 crc kubenswrapper[4890]: I1125 15:19:40.233125 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-dblm6"] Nov 25 15:19:40 crc kubenswrapper[4890]: W1125 15:19:40.279470 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb803ac45_3f40_4188_a0cc_df469b30f1ea.slice/crio-9082a1f88dc7b7666d50eccc3dd2b60d2075e9f08cec6a5df2fda5a7c6090379 WatchSource:0}: Error finding container 9082a1f88dc7b7666d50eccc3dd2b60d2075e9f08cec6a5df2fda5a7c6090379: Status 404 returned error can't find the container with id 9082a1f88dc7b7666d50eccc3dd2b60d2075e9f08cec6a5df2fda5a7c6090379 Nov 25 15:19:40 crc kubenswrapper[4890]: I1125 15:19:40.316225 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-bf56cc7bf-tn5lx"] Nov 25 15:19:40 crc kubenswrapper[4890]: I1125 15:19:40.338190 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-nrbtc-config-pzg8w"] Nov 25 15:19:40 crc kubenswrapper[4890]: I1125 15:19:40.339595 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-nrbtc-config-pzg8w" Nov 25 15:19:40 crc kubenswrapper[4890]: I1125 15:19:40.352405 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 25 15:19:40 crc kubenswrapper[4890]: I1125 15:19:40.368281 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-nrbtc-config-pzg8w"] Nov 25 15:19:40 crc kubenswrapper[4890]: W1125 15:19:40.386278 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd64cf59f_da15_4183_be46_13510f3f6efb.slice/crio-14b8c029bcac1c3a368275ab4b80a168117208e12013f7f2dafc0bad80497751 WatchSource:0}: Error finding container 14b8c029bcac1c3a368275ab4b80a168117208e12013f7f2dafc0bad80497751: Status 404 returned error can't find the container with id 14b8c029bcac1c3a368275ab4b80a168117208e12013f7f2dafc0bad80497751 Nov 25 15:19:40 crc kubenswrapper[4890]: I1125 15:19:40.452628 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-mg8z4"] Nov 25 15:19:40 crc kubenswrapper[4890]: I1125 15:19:40.472250 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/18de5476-1cdf-40f7-b5f9-1adf7d471ba7-var-log-ovn\") pod \"ovn-controller-nrbtc-config-pzg8w\" (UID: \"18de5476-1cdf-40f7-b5f9-1adf7d471ba7\") " pod="openstack/ovn-controller-nrbtc-config-pzg8w" Nov 25 15:19:40 crc kubenswrapper[4890]: I1125 15:19:40.472303 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtqwc\" (UniqueName: \"kubernetes.io/projected/18de5476-1cdf-40f7-b5f9-1adf7d471ba7-kube-api-access-wtqwc\") pod \"ovn-controller-nrbtc-config-pzg8w\" (UID: \"18de5476-1cdf-40f7-b5f9-1adf7d471ba7\") " pod="openstack/ovn-controller-nrbtc-config-pzg8w" Nov 25 15:19:40 crc kubenswrapper[4890]: I1125 15:19:40.472333 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/18de5476-1cdf-40f7-b5f9-1adf7d471ba7-scripts\") pod \"ovn-controller-nrbtc-config-pzg8w\" (UID: \"18de5476-1cdf-40f7-b5f9-1adf7d471ba7\") " pod="openstack/ovn-controller-nrbtc-config-pzg8w" Nov 25 15:19:40 crc kubenswrapper[4890]: I1125 15:19:40.472421 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/18de5476-1cdf-40f7-b5f9-1adf7d471ba7-var-run-ovn\") pod \"ovn-controller-nrbtc-config-pzg8w\" (UID: \"18de5476-1cdf-40f7-b5f9-1adf7d471ba7\") " pod="openstack/ovn-controller-nrbtc-config-pzg8w" Nov 25 15:19:40 crc kubenswrapper[4890]: I1125 15:19:40.472450 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/18de5476-1cdf-40f7-b5f9-1adf7d471ba7-additional-scripts\") pod \"ovn-controller-nrbtc-config-pzg8w\" (UID: \"18de5476-1cdf-40f7-b5f9-1adf7d471ba7\") " pod="openstack/ovn-controller-nrbtc-config-pzg8w" Nov 25 15:19:40 crc kubenswrapper[4890]: I1125 15:19:40.472471 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/18de5476-1cdf-40f7-b5f9-1adf7d471ba7-var-run\") pod \"ovn-controller-nrbtc-config-pzg8w\" (UID: \"18de5476-1cdf-40f7-b5f9-1adf7d471ba7\") " pod="openstack/ovn-controller-nrbtc-config-pzg8w" Nov 25 15:19:40 crc kubenswrapper[4890]: I1125 15:19:40.482230 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-wlvxc"] Nov 25 15:19:40 crc kubenswrapper[4890]: W1125 15:19:40.523406 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod449c7dd2_e3c8_48c2_8770_c1d6a77fce7b.slice/crio-42047d72e03e641622835078e3ab9955e61599a2e2d9caf1245c6703ca3426d5 WatchSource:0}: Error finding container 42047d72e03e641622835078e3ab9955e61599a2e2d9caf1245c6703ca3426d5: Status 404 returned error can't find the container with id 42047d72e03e641622835078e3ab9955e61599a2e2d9caf1245c6703ca3426d5 Nov 25 15:19:40 crc kubenswrapper[4890]: I1125 15:19:40.568350 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-wlvxc" event={"ID":"449c7dd2-e3c8-48c2-8770-c1d6a77fce7b","Type":"ContainerStarted","Data":"42047d72e03e641622835078e3ab9955e61599a2e2d9caf1245c6703ca3426d5"} Nov 25 15:19:40 crc kubenswrapper[4890]: I1125 15:19:40.579852 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/18de5476-1cdf-40f7-b5f9-1adf7d471ba7-scripts\") pod \"ovn-controller-nrbtc-config-pzg8w\" (UID: \"18de5476-1cdf-40f7-b5f9-1adf7d471ba7\") " pod="openstack/ovn-controller-nrbtc-config-pzg8w" Nov 25 15:19:40 crc kubenswrapper[4890]: I1125 15:19:40.582474 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/18de5476-1cdf-40f7-b5f9-1adf7d471ba7-var-run-ovn\") pod \"ovn-controller-nrbtc-config-pzg8w\" (UID: \"18de5476-1cdf-40f7-b5f9-1adf7d471ba7\") " pod="openstack/ovn-controller-nrbtc-config-pzg8w" Nov 25 15:19:40 crc kubenswrapper[4890]: I1125 15:19:40.582678 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/18de5476-1cdf-40f7-b5f9-1adf7d471ba7-additional-scripts\") pod \"ovn-controller-nrbtc-config-pzg8w\" (UID: \"18de5476-1cdf-40f7-b5f9-1adf7d471ba7\") " pod="openstack/ovn-controller-nrbtc-config-pzg8w" Nov 25 15:19:40 crc kubenswrapper[4890]: I1125 15:19:40.582794 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/18de5476-1cdf-40f7-b5f9-1adf7d471ba7-var-run\") pod \"ovn-controller-nrbtc-config-pzg8w\" (UID: \"18de5476-1cdf-40f7-b5f9-1adf7d471ba7\") " pod="openstack/ovn-controller-nrbtc-config-pzg8w" Nov 25 15:19:40 crc kubenswrapper[4890]: I1125 15:19:40.582952 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/18de5476-1cdf-40f7-b5f9-1adf7d471ba7-var-log-ovn\") pod \"ovn-controller-nrbtc-config-pzg8w\" (UID: \"18de5476-1cdf-40f7-b5f9-1adf7d471ba7\") " pod="openstack/ovn-controller-nrbtc-config-pzg8w" Nov 25 15:19:40 crc kubenswrapper[4890]: I1125 15:19:40.583063 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtqwc\" (UniqueName: \"kubernetes.io/projected/18de5476-1cdf-40f7-b5f9-1adf7d471ba7-kube-api-access-wtqwc\") pod \"ovn-controller-nrbtc-config-pzg8w\" (UID: \"18de5476-1cdf-40f7-b5f9-1adf7d471ba7\") " pod="openstack/ovn-controller-nrbtc-config-pzg8w" Nov 25 15:19:40 crc kubenswrapper[4890]: I1125 15:19:40.583705 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/18de5476-1cdf-40f7-b5f9-1adf7d471ba7-scripts\") pod \"ovn-controller-nrbtc-config-pzg8w\" (UID: \"18de5476-1cdf-40f7-b5f9-1adf7d471ba7\") " pod="openstack/ovn-controller-nrbtc-config-pzg8w" Nov 25 15:19:40 crc kubenswrapper[4890]: I1125 15:19:40.584011 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/18de5476-1cdf-40f7-b5f9-1adf7d471ba7-var-run\") pod \"ovn-controller-nrbtc-config-pzg8w\" (UID: \"18de5476-1cdf-40f7-b5f9-1adf7d471ba7\") " pod="openstack/ovn-controller-nrbtc-config-pzg8w" Nov 25 15:19:40 crc kubenswrapper[4890]: I1125 15:19:40.585402 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/18de5476-1cdf-40f7-b5f9-1adf7d471ba7-additional-scripts\") pod \"ovn-controller-nrbtc-config-pzg8w\" (UID: \"18de5476-1cdf-40f7-b5f9-1adf7d471ba7\") " pod="openstack/ovn-controller-nrbtc-config-pzg8w" Nov 25 15:19:40 crc kubenswrapper[4890]: I1125 15:19:40.585479 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/18de5476-1cdf-40f7-b5f9-1adf7d471ba7-var-log-ovn\") pod \"ovn-controller-nrbtc-config-pzg8w\" (UID: \"18de5476-1cdf-40f7-b5f9-1adf7d471ba7\") " pod="openstack/ovn-controller-nrbtc-config-pzg8w" Nov 25 15:19:40 crc kubenswrapper[4890]: I1125 15:19:40.589073 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-mg8z4" event={"ID":"13af51fd-267a-4b92-b08b-49db0d79a866","Type":"ContainerStarted","Data":"49f5d63e0c337e5d81d61fd78d2701e12b7d8929b5365a29732cf565ade40ad1"} Nov 25 15:19:40 crc kubenswrapper[4890]: I1125 15:19:40.591709 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/18de5476-1cdf-40f7-b5f9-1adf7d471ba7-var-run-ovn\") pod \"ovn-controller-nrbtc-config-pzg8w\" (UID: \"18de5476-1cdf-40f7-b5f9-1adf7d471ba7\") " pod="openstack/ovn-controller-nrbtc-config-pzg8w" Nov 25 15:19:40 crc kubenswrapper[4890]: I1125 15:19:40.595394 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-bf56cc7bf-tn5lx" event={"ID":"d64cf59f-da15-4183-be46-13510f3f6efb","Type":"ContainerStarted","Data":"14b8c029bcac1c3a368275ab4b80a168117208e12013f7f2dafc0bad80497751"} Nov 25 15:19:40 crc kubenswrapper[4890]: I1125 15:19:40.614974 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtqwc\" (UniqueName: \"kubernetes.io/projected/18de5476-1cdf-40f7-b5f9-1adf7d471ba7-kube-api-access-wtqwc\") pod \"ovn-controller-nrbtc-config-pzg8w\" (UID: \"18de5476-1cdf-40f7-b5f9-1adf7d471ba7\") " pod="openstack/ovn-controller-nrbtc-config-pzg8w" Nov 25 15:19:40 crc kubenswrapper[4890]: I1125 15:19:40.619371 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dblm6" event={"ID":"b803ac45-3f40-4188-a0cc-df469b30f1ea","Type":"ContainerStarted","Data":"9082a1f88dc7b7666d50eccc3dd2b60d2075e9f08cec6a5df2fda5a7c6090379"} Nov 25 15:19:40 crc kubenswrapper[4890]: I1125 15:19:40.644081 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f877ddd87-7g2r9" event={"ID":"181307eb-d901-4255-8dab-64b90259f746","Type":"ContainerStarted","Data":"cc9f3544a839e2fdf57070ee2c331b0c882957e8aa035cc280634af8ea48152c"} Nov 25 15:19:40 crc kubenswrapper[4890]: I1125 15:19:40.655224 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7f7df64d4f-k6jfd"] Nov 25 15:19:40 crc kubenswrapper[4890]: I1125 15:19:40.703875 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-nrbtc-config-pzg8w" Nov 25 15:19:40 crc kubenswrapper[4890]: I1125 15:19:40.850893 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-j7w9k"] Nov 25 15:19:40 crc kubenswrapper[4890]: I1125 15:19:40.976694 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-68dcc9cf6f-lm2sb"] Nov 25 15:19:41 crc kubenswrapper[4890]: I1125 15:19:41.114891 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:19:41 crc kubenswrapper[4890]: I1125 15:19:41.132254 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-f8ctf"] Nov 25 15:19:41 crc kubenswrapper[4890]: I1125 15:19:41.359966 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-nrbtc-config-pzg8w"] Nov 25 15:19:41 crc kubenswrapper[4890]: W1125 15:19:41.554731 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfdcacd06_b857_46cf_8516_345e271cf882.slice/crio-507bac2bdb8d51b224e50337dc96073d3ec3a50ad9762887ab9180a4379d584d WatchSource:0}: Error finding container 507bac2bdb8d51b224e50337dc96073d3ec3a50ad9762887ab9180a4379d584d: Status 404 returned error can't find the container with id 507bac2bdb8d51b224e50337dc96073d3ec3a50ad9762887ab9180a4379d584d Nov 25 15:19:41 crc kubenswrapper[4890]: I1125 15:19:41.677731 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68dcc9cf6f-lm2sb" event={"ID":"d760f8c7-b47e-455f-81c3-0ce064da5c96","Type":"ContainerStarted","Data":"7d98a127dfc030e0dd214a9f59a1f1f03c0fc1a527250179ecd5b339fb28e6a8"} Nov 25 15:19:41 crc kubenswrapper[4890]: I1125 15:19:41.679974 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f7df64d4f-k6jfd" event={"ID":"b246a631-f563-4f3f-aef3-caa0d92520c6","Type":"ContainerStarted","Data":"ed9e23083e7e95e1e0c845406cfae1a4059af7e8375cd11d1c0879b29033bb3f"} Nov 25 15:19:41 crc kubenswrapper[4890]: I1125 15:19:41.682116 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a886f74-d737-46b2-9647-232deec47e87","Type":"ContainerStarted","Data":"bda5bba1161b8bf25188069ccbc59156bbc3aaaeae183aafe126e72ca564a143"} Nov 25 15:19:41 crc kubenswrapper[4890]: I1125 15:19:41.685816 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-j7w9k" event={"ID":"997a078a-52bd-4eb2-8fde-f80d05d9b18c","Type":"ContainerStarted","Data":"cf4e901c7a76d094a83862a406a6fb9c94053a7cf4e304013cf95d3d582eae9f"} Nov 25 15:19:41 crc kubenswrapper[4890]: I1125 15:19:41.688578 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-f8ctf" event={"ID":"fdcacd06-b857-46cf-8516-345e271cf882","Type":"ContainerStarted","Data":"507bac2bdb8d51b224e50337dc96073d3ec3a50ad9762887ab9180a4379d584d"} Nov 25 15:19:41 crc kubenswrapper[4890]: I1125 15:19:41.693976 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dblm6" event={"ID":"b803ac45-3f40-4188-a0cc-df469b30f1ea","Type":"ContainerStarted","Data":"90c930538f8ad9440ca2e8fefaa6d6abf1268230e5c9ba07b74a553ef0141b4d"} Nov 25 15:19:41 crc kubenswrapper[4890]: I1125 15:19:41.697520 4890 generic.go:334] "Generic (PLEG): container finished" podID="181307eb-d901-4255-8dab-64b90259f746" containerID="6afe12fbfb7f772a710739c53712e17640022c0af2cf911f23f8b87c484d195d" exitCode=0 Nov 25 15:19:41 crc kubenswrapper[4890]: I1125 15:19:41.697648 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f877ddd87-7g2r9" event={"ID":"181307eb-d901-4255-8dab-64b90259f746","Type":"ContainerDied","Data":"6afe12fbfb7f772a710739c53712e17640022c0af2cf911f23f8b87c484d195d"} Nov 25 15:19:41 crc kubenswrapper[4890]: I1125 15:19:41.702939 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-mg8z4" event={"ID":"13af51fd-267a-4b92-b08b-49db0d79a866","Type":"ContainerStarted","Data":"57a8dfaf17074cb0e7b19059e430746740989c7d1433da6ed075a05661853b0a"} Nov 25 15:19:41 crc kubenswrapper[4890]: I1125 15:19:41.704371 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-nrbtc-config-pzg8w" event={"ID":"18de5476-1cdf-40f7-b5f9-1adf7d471ba7","Type":"ContainerStarted","Data":"e36773440c73e7ee679bcfb0d69716d859efe62dec1da214ecc744df8fd16f81"} Nov 25 15:19:41 crc kubenswrapper[4890]: I1125 15:19:41.723050 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-dblm6" podStartSLOduration=3.723016827 podStartE2EDuration="3.723016827s" podCreationTimestamp="2025-11-25 15:19:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:19:41.722099654 +0000 UTC m=+1040.164562264" watchObservedRunningTime="2025-11-25 15:19:41.723016827 +0000 UTC m=+1040.165479597" Nov 25 15:19:41 crc kubenswrapper[4890]: I1125 15:19:41.793770 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-mg8z4" podStartSLOduration=2.793748515 podStartE2EDuration="2.793748515s" podCreationTimestamp="2025-11-25 15:19:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:19:41.792612316 +0000 UTC m=+1040.235074926" watchObservedRunningTime="2025-11-25 15:19:41.793748515 +0000 UTC m=+1040.236211115" Nov 25 15:19:42 crc kubenswrapper[4890]: I1125 15:19:42.088755 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-bf56cc7bf-tn5lx"] Nov 25 15:19:42 crc kubenswrapper[4890]: I1125 15:19:42.147210 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-568c5cd49-gz9n4"] Nov 25 15:19:42 crc kubenswrapper[4890]: I1125 15:19:42.149766 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-568c5cd49-gz9n4" Nov 25 15:19:42 crc kubenswrapper[4890]: I1125 15:19:42.192666 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-568c5cd49-gz9n4"] Nov 25 15:19:42 crc kubenswrapper[4890]: I1125 15:19:42.224411 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e40cb582-ce6c-4064-8c1b-16fb4f7f3e62-scripts\") pod \"horizon-568c5cd49-gz9n4\" (UID: \"e40cb582-ce6c-4064-8c1b-16fb4f7f3e62\") " pod="openstack/horizon-568c5cd49-gz9n4" Nov 25 15:19:42 crc kubenswrapper[4890]: I1125 15:19:42.224455 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e40cb582-ce6c-4064-8c1b-16fb4f7f3e62-config-data\") pod \"horizon-568c5cd49-gz9n4\" (UID: \"e40cb582-ce6c-4064-8c1b-16fb4f7f3e62\") " pod="openstack/horizon-568c5cd49-gz9n4" Nov 25 15:19:42 crc kubenswrapper[4890]: I1125 15:19:42.224527 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e40cb582-ce6c-4064-8c1b-16fb4f7f3e62-logs\") pod \"horizon-568c5cd49-gz9n4\" (UID: \"e40cb582-ce6c-4064-8c1b-16fb4f7f3e62\") " pod="openstack/horizon-568c5cd49-gz9n4" Nov 25 15:19:42 crc kubenswrapper[4890]: I1125 15:19:42.224623 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e40cb582-ce6c-4064-8c1b-16fb4f7f3e62-horizon-secret-key\") pod \"horizon-568c5cd49-gz9n4\" (UID: \"e40cb582-ce6c-4064-8c1b-16fb4f7f3e62\") " pod="openstack/horizon-568c5cd49-gz9n4" Nov 25 15:19:42 crc kubenswrapper[4890]: I1125 15:19:42.224652 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5q48k\" (UniqueName: \"kubernetes.io/projected/e40cb582-ce6c-4064-8c1b-16fb4f7f3e62-kube-api-access-5q48k\") pod \"horizon-568c5cd49-gz9n4\" (UID: \"e40cb582-ce6c-4064-8c1b-16fb4f7f3e62\") " pod="openstack/horizon-568c5cd49-gz9n4" Nov 25 15:19:42 crc kubenswrapper[4890]: I1125 15:19:42.283892 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f877ddd87-7g2r9" Nov 25 15:19:42 crc kubenswrapper[4890]: I1125 15:19:42.291136 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:19:42 crc kubenswrapper[4890]: I1125 15:19:42.329100 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e40cb582-ce6c-4064-8c1b-16fb4f7f3e62-horizon-secret-key\") pod \"horizon-568c5cd49-gz9n4\" (UID: \"e40cb582-ce6c-4064-8c1b-16fb4f7f3e62\") " pod="openstack/horizon-568c5cd49-gz9n4" Nov 25 15:19:42 crc kubenswrapper[4890]: I1125 15:19:42.329234 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5q48k\" (UniqueName: \"kubernetes.io/projected/e40cb582-ce6c-4064-8c1b-16fb4f7f3e62-kube-api-access-5q48k\") pod \"horizon-568c5cd49-gz9n4\" (UID: \"e40cb582-ce6c-4064-8c1b-16fb4f7f3e62\") " pod="openstack/horizon-568c5cd49-gz9n4" Nov 25 15:19:42 crc kubenswrapper[4890]: I1125 15:19:42.329312 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e40cb582-ce6c-4064-8c1b-16fb4f7f3e62-scripts\") pod \"horizon-568c5cd49-gz9n4\" (UID: \"e40cb582-ce6c-4064-8c1b-16fb4f7f3e62\") " pod="openstack/horizon-568c5cd49-gz9n4" Nov 25 15:19:42 crc kubenswrapper[4890]: I1125 15:19:42.329364 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e40cb582-ce6c-4064-8c1b-16fb4f7f3e62-config-data\") pod \"horizon-568c5cd49-gz9n4\" (UID: \"e40cb582-ce6c-4064-8c1b-16fb4f7f3e62\") " pod="openstack/horizon-568c5cd49-gz9n4" Nov 25 15:19:42 crc kubenswrapper[4890]: I1125 15:19:42.329451 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e40cb582-ce6c-4064-8c1b-16fb4f7f3e62-logs\") pod \"horizon-568c5cd49-gz9n4\" (UID: \"e40cb582-ce6c-4064-8c1b-16fb4f7f3e62\") " pod="openstack/horizon-568c5cd49-gz9n4" Nov 25 15:19:42 crc kubenswrapper[4890]: I1125 15:19:42.332711 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e40cb582-ce6c-4064-8c1b-16fb4f7f3e62-logs\") pod \"horizon-568c5cd49-gz9n4\" (UID: \"e40cb582-ce6c-4064-8c1b-16fb4f7f3e62\") " pod="openstack/horizon-568c5cd49-gz9n4" Nov 25 15:19:42 crc kubenswrapper[4890]: I1125 15:19:42.333872 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e40cb582-ce6c-4064-8c1b-16fb4f7f3e62-config-data\") pod \"horizon-568c5cd49-gz9n4\" (UID: \"e40cb582-ce6c-4064-8c1b-16fb4f7f3e62\") " pod="openstack/horizon-568c5cd49-gz9n4" Nov 25 15:19:42 crc kubenswrapper[4890]: I1125 15:19:42.334505 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e40cb582-ce6c-4064-8c1b-16fb4f7f3e62-scripts\") pod \"horizon-568c5cd49-gz9n4\" (UID: \"e40cb582-ce6c-4064-8c1b-16fb4f7f3e62\") " pod="openstack/horizon-568c5cd49-gz9n4" Nov 25 15:19:42 crc kubenswrapper[4890]: I1125 15:19:42.346981 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e40cb582-ce6c-4064-8c1b-16fb4f7f3e62-horizon-secret-key\") pod \"horizon-568c5cd49-gz9n4\" (UID: \"e40cb582-ce6c-4064-8c1b-16fb4f7f3e62\") " pod="openstack/horizon-568c5cd49-gz9n4" Nov 25 15:19:42 crc kubenswrapper[4890]: I1125 15:19:42.364087 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5q48k\" (UniqueName: \"kubernetes.io/projected/e40cb582-ce6c-4064-8c1b-16fb4f7f3e62-kube-api-access-5q48k\") pod \"horizon-568c5cd49-gz9n4\" (UID: \"e40cb582-ce6c-4064-8c1b-16fb4f7f3e62\") " pod="openstack/horizon-568c5cd49-gz9n4" Nov 25 15:19:42 crc kubenswrapper[4890]: I1125 15:19:42.434380 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/181307eb-d901-4255-8dab-64b90259f746-dns-svc\") pod \"181307eb-d901-4255-8dab-64b90259f746\" (UID: \"181307eb-d901-4255-8dab-64b90259f746\") " Nov 25 15:19:42 crc kubenswrapper[4890]: I1125 15:19:42.434631 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/181307eb-d901-4255-8dab-64b90259f746-config\") pod \"181307eb-d901-4255-8dab-64b90259f746\" (UID: \"181307eb-d901-4255-8dab-64b90259f746\") " Nov 25 15:19:42 crc kubenswrapper[4890]: I1125 15:19:42.434744 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/181307eb-d901-4255-8dab-64b90259f746-ovsdbserver-sb\") pod \"181307eb-d901-4255-8dab-64b90259f746\" (UID: \"181307eb-d901-4255-8dab-64b90259f746\") " Nov 25 15:19:42 crc kubenswrapper[4890]: I1125 15:19:42.434812 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fhg79\" (UniqueName: \"kubernetes.io/projected/181307eb-d901-4255-8dab-64b90259f746-kube-api-access-fhg79\") pod \"181307eb-d901-4255-8dab-64b90259f746\" (UID: \"181307eb-d901-4255-8dab-64b90259f746\") " Nov 25 15:19:42 crc kubenswrapper[4890]: I1125 15:19:42.434855 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/181307eb-d901-4255-8dab-64b90259f746-ovsdbserver-nb\") pod \"181307eb-d901-4255-8dab-64b90259f746\" (UID: \"181307eb-d901-4255-8dab-64b90259f746\") " Nov 25 15:19:42 crc kubenswrapper[4890]: I1125 15:19:42.455208 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/181307eb-d901-4255-8dab-64b90259f746-kube-api-access-fhg79" (OuterVolumeSpecName: "kube-api-access-fhg79") pod "181307eb-d901-4255-8dab-64b90259f746" (UID: "181307eb-d901-4255-8dab-64b90259f746"). InnerVolumeSpecName "kube-api-access-fhg79". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:19:42 crc kubenswrapper[4890]: I1125 15:19:42.549024 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fhg79\" (UniqueName: \"kubernetes.io/projected/181307eb-d901-4255-8dab-64b90259f746-kube-api-access-fhg79\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:42 crc kubenswrapper[4890]: I1125 15:19:42.551708 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-568c5cd49-gz9n4" Nov 25 15:19:42 crc kubenswrapper[4890]: I1125 15:19:42.609540 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/181307eb-d901-4255-8dab-64b90259f746-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "181307eb-d901-4255-8dab-64b90259f746" (UID: "181307eb-d901-4255-8dab-64b90259f746"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:19:42 crc kubenswrapper[4890]: I1125 15:19:42.621507 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/181307eb-d901-4255-8dab-64b90259f746-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "181307eb-d901-4255-8dab-64b90259f746" (UID: "181307eb-d901-4255-8dab-64b90259f746"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:19:42 crc kubenswrapper[4890]: I1125 15:19:42.631135 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/181307eb-d901-4255-8dab-64b90259f746-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "181307eb-d901-4255-8dab-64b90259f746" (UID: "181307eb-d901-4255-8dab-64b90259f746"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:19:42 crc kubenswrapper[4890]: I1125 15:19:42.640155 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/181307eb-d901-4255-8dab-64b90259f746-config" (OuterVolumeSpecName: "config") pod "181307eb-d901-4255-8dab-64b90259f746" (UID: "181307eb-d901-4255-8dab-64b90259f746"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:19:42 crc kubenswrapper[4890]: I1125 15:19:42.652826 4890 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/181307eb-d901-4255-8dab-64b90259f746-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:42 crc kubenswrapper[4890]: I1125 15:19:42.652874 4890 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/181307eb-d901-4255-8dab-64b90259f746-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:42 crc kubenswrapper[4890]: I1125 15:19:42.652889 4890 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/181307eb-d901-4255-8dab-64b90259f746-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:42 crc kubenswrapper[4890]: I1125 15:19:42.652901 4890 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/181307eb-d901-4255-8dab-64b90259f746-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:42 crc kubenswrapper[4890]: I1125 15:19:42.731342 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6517688d-0723-4809-ba51-bcaf48524157","Type":"ContainerStarted","Data":"12833165b6b97b3108d1af7de8d7dbea76e7ff7133db3619e117274d433fc0f1"} Nov 25 15:19:42 crc kubenswrapper[4890]: I1125 15:19:42.748794 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f877ddd87-7g2r9" event={"ID":"181307eb-d901-4255-8dab-64b90259f746","Type":"ContainerDied","Data":"cc9f3544a839e2fdf57070ee2c331b0c882957e8aa035cc280634af8ea48152c"} Nov 25 15:19:42 crc kubenswrapper[4890]: I1125 15:19:42.748853 4890 scope.go:117] "RemoveContainer" containerID="6afe12fbfb7f772a710739c53712e17640022c0af2cf911f23f8b87c484d195d" Nov 25 15:19:42 crc kubenswrapper[4890]: I1125 15:19:42.749004 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f877ddd87-7g2r9" Nov 25 15:19:42 crc kubenswrapper[4890]: I1125 15:19:42.761611 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-nrbtc-config-pzg8w" event={"ID":"18de5476-1cdf-40f7-b5f9-1adf7d471ba7","Type":"ContainerStarted","Data":"f28158f6c45d4687b53cc873efafe9bd6ddd6a433e6e23b764315cf714baff16"} Nov 25 15:19:42 crc kubenswrapper[4890]: I1125 15:19:42.768464 4890 generic.go:334] "Generic (PLEG): container finished" podID="d760f8c7-b47e-455f-81c3-0ce064da5c96" containerID="fecd5109953cc4cec007d516b879961539f951212be4adc3272e23b74098f3c3" exitCode=0 Nov 25 15:19:42 crc kubenswrapper[4890]: I1125 15:19:42.769658 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68dcc9cf6f-lm2sb" event={"ID":"d760f8c7-b47e-455f-81c3-0ce064da5c96","Type":"ContainerDied","Data":"fecd5109953cc4cec007d516b879961539f951212be4adc3272e23b74098f3c3"} Nov 25 15:19:42 crc kubenswrapper[4890]: I1125 15:19:42.783139 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-nrbtc-config-pzg8w" podStartSLOduration=2.783123028 podStartE2EDuration="2.783123028s" podCreationTimestamp="2025-11-25 15:19:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:19:42.776120912 +0000 UTC m=+1041.218583512" watchObservedRunningTime="2025-11-25 15:19:42.783123028 +0000 UTC m=+1041.225585638" Nov 25 15:19:42 crc kubenswrapper[4890]: I1125 15:19:42.872605 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f877ddd87-7g2r9"] Nov 25 15:19:42 crc kubenswrapper[4890]: I1125 15:19:42.893178 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f877ddd87-7g2r9"] Nov 25 15:19:43 crc kubenswrapper[4890]: I1125 15:19:43.334739 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-568c5cd49-gz9n4"] Nov 25 15:19:43 crc kubenswrapper[4890]: W1125 15:19:43.382033 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode40cb582_ce6c_4064_8c1b_16fb4f7f3e62.slice/crio-85ed31693c35ff51013b13e0efe5e290540726e59a59b4df9cd1eee9522cfbcd WatchSource:0}: Error finding container 85ed31693c35ff51013b13e0efe5e290540726e59a59b4df9cd1eee9522cfbcd: Status 404 returned error can't find the container with id 85ed31693c35ff51013b13e0efe5e290540726e59a59b4df9cd1eee9522cfbcd Nov 25 15:19:43 crc kubenswrapper[4890]: I1125 15:19:43.792251 4890 generic.go:334] "Generic (PLEG): container finished" podID="18de5476-1cdf-40f7-b5f9-1adf7d471ba7" containerID="f28158f6c45d4687b53cc873efafe9bd6ddd6a433e6e23b764315cf714baff16" exitCode=0 Nov 25 15:19:43 crc kubenswrapper[4890]: I1125 15:19:43.792334 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-nrbtc-config-pzg8w" event={"ID":"18de5476-1cdf-40f7-b5f9-1adf7d471ba7","Type":"ContainerDied","Data":"f28158f6c45d4687b53cc873efafe9bd6ddd6a433e6e23b764315cf714baff16"} Nov 25 15:19:43 crc kubenswrapper[4890]: I1125 15:19:43.805609 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-568c5cd49-gz9n4" event={"ID":"e40cb582-ce6c-4064-8c1b-16fb4f7f3e62","Type":"ContainerStarted","Data":"85ed31693c35ff51013b13e0efe5e290540726e59a59b4df9cd1eee9522cfbcd"} Nov 25 15:19:43 crc kubenswrapper[4890]: I1125 15:19:43.830562 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68dcc9cf6f-lm2sb" event={"ID":"d760f8c7-b47e-455f-81c3-0ce064da5c96","Type":"ContainerStarted","Data":"c2caff9e64a05d375605099a37e8642c11e8fcf517f2776b1cc179c5b8d8d3e5"} Nov 25 15:19:43 crc kubenswrapper[4890]: I1125 15:19:43.830625 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-68dcc9cf6f-lm2sb" Nov 25 15:19:43 crc kubenswrapper[4890]: I1125 15:19:43.846720 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6517688d-0723-4809-ba51-bcaf48524157","Type":"ContainerStarted","Data":"e1431cd3734854e213f25f7f931e5dc7b07167bc8bc69249b6e5bed1eae6e40a"} Nov 25 15:19:43 crc kubenswrapper[4890]: I1125 15:19:43.846801 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6517688d-0723-4809-ba51-bcaf48524157","Type":"ContainerStarted","Data":"a00c74ff09ff4451771f3018f8118e6db56e60bfa3526eda7ec9d560d2334db1"} Nov 25 15:19:43 crc kubenswrapper[4890]: I1125 15:19:43.872576 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-68dcc9cf6f-lm2sb" podStartSLOduration=4.872548465 podStartE2EDuration="4.872548465s" podCreationTimestamp="2025-11-25 15:19:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:19:43.865012015 +0000 UTC m=+1042.307474625" watchObservedRunningTime="2025-11-25 15:19:43.872548465 +0000 UTC m=+1042.315011075" Nov 25 15:19:44 crc kubenswrapper[4890]: I1125 15:19:44.187545 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="181307eb-d901-4255-8dab-64b90259f746" path="/var/lib/kubelet/pods/181307eb-d901-4255-8dab-64b90259f746/volumes" Nov 25 15:19:44 crc kubenswrapper[4890]: I1125 15:19:44.882921 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6517688d-0723-4809-ba51-bcaf48524157","Type":"ContainerStarted","Data":"05c5d9dcf054e8409df3ffbc5e3e78f902e4eff454f57fb72569c59b3537f8da"} Nov 25 15:19:44 crc kubenswrapper[4890]: I1125 15:19:44.993305 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-nrbtc" Nov 25 15:19:45 crc kubenswrapper[4890]: I1125 15:19:45.278397 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-nrbtc-config-pzg8w" Nov 25 15:19:45 crc kubenswrapper[4890]: I1125 15:19:45.355314 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/18de5476-1cdf-40f7-b5f9-1adf7d471ba7-var-run\") pod \"18de5476-1cdf-40f7-b5f9-1adf7d471ba7\" (UID: \"18de5476-1cdf-40f7-b5f9-1adf7d471ba7\") " Nov 25 15:19:45 crc kubenswrapper[4890]: I1125 15:19:45.355430 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/18de5476-1cdf-40f7-b5f9-1adf7d471ba7-var-run" (OuterVolumeSpecName: "var-run") pod "18de5476-1cdf-40f7-b5f9-1adf7d471ba7" (UID: "18de5476-1cdf-40f7-b5f9-1adf7d471ba7"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 15:19:45 crc kubenswrapper[4890]: I1125 15:19:45.355481 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/18de5476-1cdf-40f7-b5f9-1adf7d471ba7-var-log-ovn\") pod \"18de5476-1cdf-40f7-b5f9-1adf7d471ba7\" (UID: \"18de5476-1cdf-40f7-b5f9-1adf7d471ba7\") " Nov 25 15:19:45 crc kubenswrapper[4890]: I1125 15:19:45.355525 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/18de5476-1cdf-40f7-b5f9-1adf7d471ba7-var-run-ovn\") pod \"18de5476-1cdf-40f7-b5f9-1adf7d471ba7\" (UID: \"18de5476-1cdf-40f7-b5f9-1adf7d471ba7\") " Nov 25 15:19:45 crc kubenswrapper[4890]: I1125 15:19:45.355557 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/18de5476-1cdf-40f7-b5f9-1adf7d471ba7-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "18de5476-1cdf-40f7-b5f9-1adf7d471ba7" (UID: "18de5476-1cdf-40f7-b5f9-1adf7d471ba7"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 15:19:45 crc kubenswrapper[4890]: I1125 15:19:45.355566 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/18de5476-1cdf-40f7-b5f9-1adf7d471ba7-scripts\") pod \"18de5476-1cdf-40f7-b5f9-1adf7d471ba7\" (UID: \"18de5476-1cdf-40f7-b5f9-1adf7d471ba7\") " Nov 25 15:19:45 crc kubenswrapper[4890]: I1125 15:19:45.355578 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/18de5476-1cdf-40f7-b5f9-1adf7d471ba7-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "18de5476-1cdf-40f7-b5f9-1adf7d471ba7" (UID: "18de5476-1cdf-40f7-b5f9-1adf7d471ba7"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 15:19:45 crc kubenswrapper[4890]: I1125 15:19:45.355622 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/18de5476-1cdf-40f7-b5f9-1adf7d471ba7-additional-scripts\") pod \"18de5476-1cdf-40f7-b5f9-1adf7d471ba7\" (UID: \"18de5476-1cdf-40f7-b5f9-1adf7d471ba7\") " Nov 25 15:19:45 crc kubenswrapper[4890]: I1125 15:19:45.355642 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wtqwc\" (UniqueName: \"kubernetes.io/projected/18de5476-1cdf-40f7-b5f9-1adf7d471ba7-kube-api-access-wtqwc\") pod \"18de5476-1cdf-40f7-b5f9-1adf7d471ba7\" (UID: \"18de5476-1cdf-40f7-b5f9-1adf7d471ba7\") " Nov 25 15:19:45 crc kubenswrapper[4890]: I1125 15:19:45.356365 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18de5476-1cdf-40f7-b5f9-1adf7d471ba7-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "18de5476-1cdf-40f7-b5f9-1adf7d471ba7" (UID: "18de5476-1cdf-40f7-b5f9-1adf7d471ba7"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:19:45 crc kubenswrapper[4890]: I1125 15:19:45.356692 4890 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/18de5476-1cdf-40f7-b5f9-1adf7d471ba7-var-run\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:45 crc kubenswrapper[4890]: I1125 15:19:45.356715 4890 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/18de5476-1cdf-40f7-b5f9-1adf7d471ba7-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:45 crc kubenswrapper[4890]: I1125 15:19:45.356727 4890 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/18de5476-1cdf-40f7-b5f9-1adf7d471ba7-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:45 crc kubenswrapper[4890]: I1125 15:19:45.356738 4890 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/18de5476-1cdf-40f7-b5f9-1adf7d471ba7-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:45 crc kubenswrapper[4890]: I1125 15:19:45.356835 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18de5476-1cdf-40f7-b5f9-1adf7d471ba7-scripts" (OuterVolumeSpecName: "scripts") pod "18de5476-1cdf-40f7-b5f9-1adf7d471ba7" (UID: "18de5476-1cdf-40f7-b5f9-1adf7d471ba7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:19:45 crc kubenswrapper[4890]: I1125 15:19:45.366265 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18de5476-1cdf-40f7-b5f9-1adf7d471ba7-kube-api-access-wtqwc" (OuterVolumeSpecName: "kube-api-access-wtqwc") pod "18de5476-1cdf-40f7-b5f9-1adf7d471ba7" (UID: "18de5476-1cdf-40f7-b5f9-1adf7d471ba7"). InnerVolumeSpecName "kube-api-access-wtqwc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:19:45 crc kubenswrapper[4890]: I1125 15:19:45.458732 4890 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/18de5476-1cdf-40f7-b5f9-1adf7d471ba7-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:45 crc kubenswrapper[4890]: I1125 15:19:45.458793 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wtqwc\" (UniqueName: \"kubernetes.io/projected/18de5476-1cdf-40f7-b5f9-1adf7d471ba7-kube-api-access-wtqwc\") on node \"crc\" DevicePath \"\"" Nov 25 15:19:45 crc kubenswrapper[4890]: I1125 15:19:45.896589 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-nrbtc-config-pzg8w"] Nov 25 15:19:45 crc kubenswrapper[4890]: I1125 15:19:45.907645 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-nrbtc-config-pzg8w" event={"ID":"18de5476-1cdf-40f7-b5f9-1adf7d471ba7","Type":"ContainerDied","Data":"e36773440c73e7ee679bcfb0d69716d859efe62dec1da214ecc744df8fd16f81"} Nov 25 15:19:45 crc kubenswrapper[4890]: I1125 15:19:45.908423 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e36773440c73e7ee679bcfb0d69716d859efe62dec1da214ecc744df8fd16f81" Nov 25 15:19:45 crc kubenswrapper[4890]: I1125 15:19:45.907716 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-nrbtc-config-pzg8w" Nov 25 15:19:45 crc kubenswrapper[4890]: I1125 15:19:45.910703 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-nrbtc-config-pzg8w"] Nov 25 15:19:45 crc kubenswrapper[4890]: I1125 15:19:45.993368 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-nrbtc-config-f5xmh"] Nov 25 15:19:45 crc kubenswrapper[4890]: E1125 15:19:45.994170 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18de5476-1cdf-40f7-b5f9-1adf7d471ba7" containerName="ovn-config" Nov 25 15:19:45 crc kubenswrapper[4890]: I1125 15:19:45.994188 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="18de5476-1cdf-40f7-b5f9-1adf7d471ba7" containerName="ovn-config" Nov 25 15:19:45 crc kubenswrapper[4890]: E1125 15:19:45.994208 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="181307eb-d901-4255-8dab-64b90259f746" containerName="init" Nov 25 15:19:45 crc kubenswrapper[4890]: I1125 15:19:45.994215 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="181307eb-d901-4255-8dab-64b90259f746" containerName="init" Nov 25 15:19:45 crc kubenswrapper[4890]: I1125 15:19:45.994405 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="18de5476-1cdf-40f7-b5f9-1adf7d471ba7" containerName="ovn-config" Nov 25 15:19:45 crc kubenswrapper[4890]: I1125 15:19:45.994417 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="181307eb-d901-4255-8dab-64b90259f746" containerName="init" Nov 25 15:19:45 crc kubenswrapper[4890]: I1125 15:19:45.995052 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-nrbtc-config-f5xmh" Nov 25 15:19:45 crc kubenswrapper[4890]: I1125 15:19:45.997366 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 25 15:19:46 crc kubenswrapper[4890]: I1125 15:19:46.001969 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-nrbtc-config-f5xmh"] Nov 25 15:19:46 crc kubenswrapper[4890]: I1125 15:19:46.073040 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/84169bad-df45-41ac-92ca-356122dec8f3-var-log-ovn\") pod \"ovn-controller-nrbtc-config-f5xmh\" (UID: \"84169bad-df45-41ac-92ca-356122dec8f3\") " pod="openstack/ovn-controller-nrbtc-config-f5xmh" Nov 25 15:19:46 crc kubenswrapper[4890]: I1125 15:19:46.073105 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52d55\" (UniqueName: \"kubernetes.io/projected/84169bad-df45-41ac-92ca-356122dec8f3-kube-api-access-52d55\") pod \"ovn-controller-nrbtc-config-f5xmh\" (UID: \"84169bad-df45-41ac-92ca-356122dec8f3\") " pod="openstack/ovn-controller-nrbtc-config-f5xmh" Nov 25 15:19:46 crc kubenswrapper[4890]: I1125 15:19:46.073143 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/84169bad-df45-41ac-92ca-356122dec8f3-var-run\") pod \"ovn-controller-nrbtc-config-f5xmh\" (UID: \"84169bad-df45-41ac-92ca-356122dec8f3\") " pod="openstack/ovn-controller-nrbtc-config-f5xmh" Nov 25 15:19:46 crc kubenswrapper[4890]: I1125 15:19:46.073231 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/84169bad-df45-41ac-92ca-356122dec8f3-var-run-ovn\") pod \"ovn-controller-nrbtc-config-f5xmh\" (UID: \"84169bad-df45-41ac-92ca-356122dec8f3\") " pod="openstack/ovn-controller-nrbtc-config-f5xmh" Nov 25 15:19:46 crc kubenswrapper[4890]: I1125 15:19:46.073341 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/84169bad-df45-41ac-92ca-356122dec8f3-scripts\") pod \"ovn-controller-nrbtc-config-f5xmh\" (UID: \"84169bad-df45-41ac-92ca-356122dec8f3\") " pod="openstack/ovn-controller-nrbtc-config-f5xmh" Nov 25 15:19:46 crc kubenswrapper[4890]: I1125 15:19:46.073371 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/84169bad-df45-41ac-92ca-356122dec8f3-additional-scripts\") pod \"ovn-controller-nrbtc-config-f5xmh\" (UID: \"84169bad-df45-41ac-92ca-356122dec8f3\") " pod="openstack/ovn-controller-nrbtc-config-f5xmh" Nov 25 15:19:46 crc kubenswrapper[4890]: I1125 15:19:46.174153 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/84169bad-df45-41ac-92ca-356122dec8f3-additional-scripts\") pod \"ovn-controller-nrbtc-config-f5xmh\" (UID: \"84169bad-df45-41ac-92ca-356122dec8f3\") " pod="openstack/ovn-controller-nrbtc-config-f5xmh" Nov 25 15:19:46 crc kubenswrapper[4890]: I1125 15:19:46.174256 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/84169bad-df45-41ac-92ca-356122dec8f3-var-log-ovn\") pod \"ovn-controller-nrbtc-config-f5xmh\" (UID: \"84169bad-df45-41ac-92ca-356122dec8f3\") " pod="openstack/ovn-controller-nrbtc-config-f5xmh" Nov 25 15:19:46 crc kubenswrapper[4890]: I1125 15:19:46.174279 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52d55\" (UniqueName: \"kubernetes.io/projected/84169bad-df45-41ac-92ca-356122dec8f3-kube-api-access-52d55\") pod \"ovn-controller-nrbtc-config-f5xmh\" (UID: \"84169bad-df45-41ac-92ca-356122dec8f3\") " pod="openstack/ovn-controller-nrbtc-config-f5xmh" Nov 25 15:19:46 crc kubenswrapper[4890]: I1125 15:19:46.174313 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/84169bad-df45-41ac-92ca-356122dec8f3-var-run\") pod \"ovn-controller-nrbtc-config-f5xmh\" (UID: \"84169bad-df45-41ac-92ca-356122dec8f3\") " pod="openstack/ovn-controller-nrbtc-config-f5xmh" Nov 25 15:19:46 crc kubenswrapper[4890]: I1125 15:19:46.174356 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/84169bad-df45-41ac-92ca-356122dec8f3-var-run-ovn\") pod \"ovn-controller-nrbtc-config-f5xmh\" (UID: \"84169bad-df45-41ac-92ca-356122dec8f3\") " pod="openstack/ovn-controller-nrbtc-config-f5xmh" Nov 25 15:19:46 crc kubenswrapper[4890]: I1125 15:19:46.174400 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/84169bad-df45-41ac-92ca-356122dec8f3-scripts\") pod \"ovn-controller-nrbtc-config-f5xmh\" (UID: \"84169bad-df45-41ac-92ca-356122dec8f3\") " pod="openstack/ovn-controller-nrbtc-config-f5xmh" Nov 25 15:19:46 crc kubenswrapper[4890]: I1125 15:19:46.174750 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/84169bad-df45-41ac-92ca-356122dec8f3-var-log-ovn\") pod \"ovn-controller-nrbtc-config-f5xmh\" (UID: \"84169bad-df45-41ac-92ca-356122dec8f3\") " pod="openstack/ovn-controller-nrbtc-config-f5xmh" Nov 25 15:19:46 crc kubenswrapper[4890]: I1125 15:19:46.175413 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/84169bad-df45-41ac-92ca-356122dec8f3-var-run\") pod \"ovn-controller-nrbtc-config-f5xmh\" (UID: \"84169bad-df45-41ac-92ca-356122dec8f3\") " pod="openstack/ovn-controller-nrbtc-config-f5xmh" Nov 25 15:19:46 crc kubenswrapper[4890]: I1125 15:19:46.175653 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/84169bad-df45-41ac-92ca-356122dec8f3-var-run-ovn\") pod \"ovn-controller-nrbtc-config-f5xmh\" (UID: \"84169bad-df45-41ac-92ca-356122dec8f3\") " pod="openstack/ovn-controller-nrbtc-config-f5xmh" Nov 25 15:19:46 crc kubenswrapper[4890]: I1125 15:19:46.175835 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/84169bad-df45-41ac-92ca-356122dec8f3-additional-scripts\") pod \"ovn-controller-nrbtc-config-f5xmh\" (UID: \"84169bad-df45-41ac-92ca-356122dec8f3\") " pod="openstack/ovn-controller-nrbtc-config-f5xmh" Nov 25 15:19:46 crc kubenswrapper[4890]: I1125 15:19:46.176692 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/84169bad-df45-41ac-92ca-356122dec8f3-scripts\") pod \"ovn-controller-nrbtc-config-f5xmh\" (UID: \"84169bad-df45-41ac-92ca-356122dec8f3\") " pod="openstack/ovn-controller-nrbtc-config-f5xmh" Nov 25 15:19:46 crc kubenswrapper[4890]: I1125 15:19:46.191189 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18de5476-1cdf-40f7-b5f9-1adf7d471ba7" path="/var/lib/kubelet/pods/18de5476-1cdf-40f7-b5f9-1adf7d471ba7/volumes" Nov 25 15:19:46 crc kubenswrapper[4890]: I1125 15:19:46.196418 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52d55\" (UniqueName: \"kubernetes.io/projected/84169bad-df45-41ac-92ca-356122dec8f3-kube-api-access-52d55\") pod \"ovn-controller-nrbtc-config-f5xmh\" (UID: \"84169bad-df45-41ac-92ca-356122dec8f3\") " pod="openstack/ovn-controller-nrbtc-config-f5xmh" Nov 25 15:19:46 crc kubenswrapper[4890]: I1125 15:19:46.314468 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-nrbtc-config-f5xmh" Nov 25 15:19:46 crc kubenswrapper[4890]: I1125 15:19:46.862884 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-nrbtc-config-f5xmh"] Nov 25 15:19:46 crc kubenswrapper[4890]: W1125 15:19:46.877591 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod84169bad_df45_41ac_92ca_356122dec8f3.slice/crio-76823c9bff48e184abe27bff50d1ec3e5677a49a05a2cd32ea24cf1836802ff0 WatchSource:0}: Error finding container 76823c9bff48e184abe27bff50d1ec3e5677a49a05a2cd32ea24cf1836802ff0: Status 404 returned error can't find the container with id 76823c9bff48e184abe27bff50d1ec3e5677a49a05a2cd32ea24cf1836802ff0 Nov 25 15:19:46 crc kubenswrapper[4890]: I1125 15:19:46.921272 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-nrbtc-config-f5xmh" event={"ID":"84169bad-df45-41ac-92ca-356122dec8f3","Type":"ContainerStarted","Data":"76823c9bff48e184abe27bff50d1ec3e5677a49a05a2cd32ea24cf1836802ff0"} Nov 25 15:19:47 crc kubenswrapper[4890]: I1125 15:19:47.940313 4890 generic.go:334] "Generic (PLEG): container finished" podID="84169bad-df45-41ac-92ca-356122dec8f3" containerID="9400aa82e81acded3c8c82ed72b89555165e618f7269b989ffebeca39eb383dc" exitCode=0 Nov 25 15:19:47 crc kubenswrapper[4890]: I1125 15:19:47.940361 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-nrbtc-config-f5xmh" event={"ID":"84169bad-df45-41ac-92ca-356122dec8f3","Type":"ContainerDied","Data":"9400aa82e81acded3c8c82ed72b89555165e618f7269b989ffebeca39eb383dc"} Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.053546 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7f7df64d4f-k6jfd"] Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.105889 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-556858788d-s2ppl"] Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.109218 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-556858788d-s2ppl" Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.112057 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.125492 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-556858788d-s2ppl"] Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.220610 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f-scripts\") pod \"horizon-556858788d-s2ppl\" (UID: \"1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f\") " pod="openstack/horizon-556858788d-s2ppl" Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.220732 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f-config-data\") pod \"horizon-556858788d-s2ppl\" (UID: \"1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f\") " pod="openstack/horizon-556858788d-s2ppl" Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.220949 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f-horizon-tls-certs\") pod \"horizon-556858788d-s2ppl\" (UID: \"1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f\") " pod="openstack/horizon-556858788d-s2ppl" Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.221041 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f-horizon-secret-key\") pod \"horizon-556858788d-s2ppl\" (UID: \"1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f\") " pod="openstack/horizon-556858788d-s2ppl" Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.221118 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f-combined-ca-bundle\") pod \"horizon-556858788d-s2ppl\" (UID: \"1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f\") " pod="openstack/horizon-556858788d-s2ppl" Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.221257 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcc2n\" (UniqueName: \"kubernetes.io/projected/1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f-kube-api-access-zcc2n\") pod \"horizon-556858788d-s2ppl\" (UID: \"1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f\") " pod="openstack/horizon-556858788d-s2ppl" Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.221338 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f-logs\") pod \"horizon-556858788d-s2ppl\" (UID: \"1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f\") " pod="openstack/horizon-556858788d-s2ppl" Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.229385 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-568c5cd49-gz9n4"] Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.259394 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7d75bbdb6d-5dvcn"] Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.266954 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7d75bbdb6d-5dvcn"] Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.267132 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d75bbdb6d-5dvcn" Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.323318 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcc2n\" (UniqueName: \"kubernetes.io/projected/1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f-kube-api-access-zcc2n\") pod \"horizon-556858788d-s2ppl\" (UID: \"1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f\") " pod="openstack/horizon-556858788d-s2ppl" Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.323409 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f-logs\") pod \"horizon-556858788d-s2ppl\" (UID: \"1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f\") " pod="openstack/horizon-556858788d-s2ppl" Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.323442 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f-scripts\") pod \"horizon-556858788d-s2ppl\" (UID: \"1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f\") " pod="openstack/horizon-556858788d-s2ppl" Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.323483 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f-config-data\") pod \"horizon-556858788d-s2ppl\" (UID: \"1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f\") " pod="openstack/horizon-556858788d-s2ppl" Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.323566 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f-horizon-tls-certs\") pod \"horizon-556858788d-s2ppl\" (UID: \"1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f\") " pod="openstack/horizon-556858788d-s2ppl" Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.323607 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f-horizon-secret-key\") pod \"horizon-556858788d-s2ppl\" (UID: \"1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f\") " pod="openstack/horizon-556858788d-s2ppl" Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.323633 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f-combined-ca-bundle\") pod \"horizon-556858788d-s2ppl\" (UID: \"1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f\") " pod="openstack/horizon-556858788d-s2ppl" Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.324780 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f-logs\") pod \"horizon-556858788d-s2ppl\" (UID: \"1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f\") " pod="openstack/horizon-556858788d-s2ppl" Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.326137 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f-scripts\") pod \"horizon-556858788d-s2ppl\" (UID: \"1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f\") " pod="openstack/horizon-556858788d-s2ppl" Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.333148 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f-config-data\") pod \"horizon-556858788d-s2ppl\" (UID: \"1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f\") " pod="openstack/horizon-556858788d-s2ppl" Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.333396 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f-horizon-secret-key\") pod \"horizon-556858788d-s2ppl\" (UID: \"1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f\") " pod="openstack/horizon-556858788d-s2ppl" Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.338961 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f-horizon-tls-certs\") pod \"horizon-556858788d-s2ppl\" (UID: \"1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f\") " pod="openstack/horizon-556858788d-s2ppl" Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.346146 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f-combined-ca-bundle\") pod \"horizon-556858788d-s2ppl\" (UID: \"1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f\") " pod="openstack/horizon-556858788d-s2ppl" Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.346532 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcc2n\" (UniqueName: \"kubernetes.io/projected/1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f-kube-api-access-zcc2n\") pod \"horizon-556858788d-s2ppl\" (UID: \"1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f\") " pod="openstack/horizon-556858788d-s2ppl" Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.425744 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bf94c7c-4031-4868-bb7a-39c84abc96ea-logs\") pod \"horizon-7d75bbdb6d-5dvcn\" (UID: \"2bf94c7c-4031-4868-bb7a-39c84abc96ea\") " pod="openstack/horizon-7d75bbdb6d-5dvcn" Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.425825 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjj84\" (UniqueName: \"kubernetes.io/projected/2bf94c7c-4031-4868-bb7a-39c84abc96ea-kube-api-access-vjj84\") pod \"horizon-7d75bbdb6d-5dvcn\" (UID: \"2bf94c7c-4031-4868-bb7a-39c84abc96ea\") " pod="openstack/horizon-7d75bbdb6d-5dvcn" Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.425907 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2bf94c7c-4031-4868-bb7a-39c84abc96ea-config-data\") pod \"horizon-7d75bbdb6d-5dvcn\" (UID: \"2bf94c7c-4031-4868-bb7a-39c84abc96ea\") " pod="openstack/horizon-7d75bbdb6d-5dvcn" Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.425975 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2bf94c7c-4031-4868-bb7a-39c84abc96ea-scripts\") pod \"horizon-7d75bbdb6d-5dvcn\" (UID: \"2bf94c7c-4031-4868-bb7a-39c84abc96ea\") " pod="openstack/horizon-7d75bbdb6d-5dvcn" Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.426137 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bf94c7c-4031-4868-bb7a-39c84abc96ea-combined-ca-bundle\") pod \"horizon-7d75bbdb6d-5dvcn\" (UID: \"2bf94c7c-4031-4868-bb7a-39c84abc96ea\") " pod="openstack/horizon-7d75bbdb6d-5dvcn" Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.426266 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2bf94c7c-4031-4868-bb7a-39c84abc96ea-horizon-secret-key\") pod \"horizon-7d75bbdb6d-5dvcn\" (UID: \"2bf94c7c-4031-4868-bb7a-39c84abc96ea\") " pod="openstack/horizon-7d75bbdb6d-5dvcn" Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.426354 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bf94c7c-4031-4868-bb7a-39c84abc96ea-horizon-tls-certs\") pod \"horizon-7d75bbdb6d-5dvcn\" (UID: \"2bf94c7c-4031-4868-bb7a-39c84abc96ea\") " pod="openstack/horizon-7d75bbdb6d-5dvcn" Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.459557 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-556858788d-s2ppl" Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.527730 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bf94c7c-4031-4868-bb7a-39c84abc96ea-horizon-tls-certs\") pod \"horizon-7d75bbdb6d-5dvcn\" (UID: \"2bf94c7c-4031-4868-bb7a-39c84abc96ea\") " pod="openstack/horizon-7d75bbdb6d-5dvcn" Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.528050 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bf94c7c-4031-4868-bb7a-39c84abc96ea-logs\") pod \"horizon-7d75bbdb6d-5dvcn\" (UID: \"2bf94c7c-4031-4868-bb7a-39c84abc96ea\") " pod="openstack/horizon-7d75bbdb6d-5dvcn" Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.528171 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjj84\" (UniqueName: \"kubernetes.io/projected/2bf94c7c-4031-4868-bb7a-39c84abc96ea-kube-api-access-vjj84\") pod \"horizon-7d75bbdb6d-5dvcn\" (UID: \"2bf94c7c-4031-4868-bb7a-39c84abc96ea\") " pod="openstack/horizon-7d75bbdb6d-5dvcn" Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.528281 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2bf94c7c-4031-4868-bb7a-39c84abc96ea-config-data\") pod \"horizon-7d75bbdb6d-5dvcn\" (UID: \"2bf94c7c-4031-4868-bb7a-39c84abc96ea\") " pod="openstack/horizon-7d75bbdb6d-5dvcn" Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.528405 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2bf94c7c-4031-4868-bb7a-39c84abc96ea-scripts\") pod \"horizon-7d75bbdb6d-5dvcn\" (UID: \"2bf94c7c-4031-4868-bb7a-39c84abc96ea\") " pod="openstack/horizon-7d75bbdb6d-5dvcn" Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.528508 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bf94c7c-4031-4868-bb7a-39c84abc96ea-combined-ca-bundle\") pod \"horizon-7d75bbdb6d-5dvcn\" (UID: \"2bf94c7c-4031-4868-bb7a-39c84abc96ea\") " pod="openstack/horizon-7d75bbdb6d-5dvcn" Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.528618 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2bf94c7c-4031-4868-bb7a-39c84abc96ea-horizon-secret-key\") pod \"horizon-7d75bbdb6d-5dvcn\" (UID: \"2bf94c7c-4031-4868-bb7a-39c84abc96ea\") " pod="openstack/horizon-7d75bbdb6d-5dvcn" Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.530588 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2bf94c7c-4031-4868-bb7a-39c84abc96ea-config-data\") pod \"horizon-7d75bbdb6d-5dvcn\" (UID: \"2bf94c7c-4031-4868-bb7a-39c84abc96ea\") " pod="openstack/horizon-7d75bbdb6d-5dvcn" Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.530995 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bf94c7c-4031-4868-bb7a-39c84abc96ea-logs\") pod \"horizon-7d75bbdb6d-5dvcn\" (UID: \"2bf94c7c-4031-4868-bb7a-39c84abc96ea\") " pod="openstack/horizon-7d75bbdb6d-5dvcn" Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.533646 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2bf94c7c-4031-4868-bb7a-39c84abc96ea-scripts\") pod \"horizon-7d75bbdb6d-5dvcn\" (UID: \"2bf94c7c-4031-4868-bb7a-39c84abc96ea\") " pod="openstack/horizon-7d75bbdb6d-5dvcn" Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.537393 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bf94c7c-4031-4868-bb7a-39c84abc96ea-horizon-tls-certs\") pod \"horizon-7d75bbdb6d-5dvcn\" (UID: \"2bf94c7c-4031-4868-bb7a-39c84abc96ea\") " pod="openstack/horizon-7d75bbdb6d-5dvcn" Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.541147 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bf94c7c-4031-4868-bb7a-39c84abc96ea-combined-ca-bundle\") pod \"horizon-7d75bbdb6d-5dvcn\" (UID: \"2bf94c7c-4031-4868-bb7a-39c84abc96ea\") " pod="openstack/horizon-7d75bbdb6d-5dvcn" Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.558680 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2bf94c7c-4031-4868-bb7a-39c84abc96ea-horizon-secret-key\") pod \"horizon-7d75bbdb6d-5dvcn\" (UID: \"2bf94c7c-4031-4868-bb7a-39c84abc96ea\") " pod="openstack/horizon-7d75bbdb6d-5dvcn" Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.568356 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjj84\" (UniqueName: \"kubernetes.io/projected/2bf94c7c-4031-4868-bb7a-39c84abc96ea-kube-api-access-vjj84\") pod \"horizon-7d75bbdb6d-5dvcn\" (UID: \"2bf94c7c-4031-4868-bb7a-39c84abc96ea\") " pod="openstack/horizon-7d75bbdb6d-5dvcn" Nov 25 15:19:48 crc kubenswrapper[4890]: I1125 15:19:48.603075 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d75bbdb6d-5dvcn" Nov 25 15:19:58 crc kubenswrapper[4890]: I1125 15:19:50.056420 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-68dcc9cf6f-lm2sb" Nov 25 15:19:58 crc kubenswrapper[4890]: I1125 15:19:50.120982 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-lgp8l"] Nov 25 15:19:58 crc kubenswrapper[4890]: I1125 15:19:50.121235 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-lgp8l" podUID="df163dc8-8261-4fdf-b446-bba9e8e1d8eb" containerName="dnsmasq-dns" containerID="cri-o://6a49096a7d81634a7d99776fad685186d59a98d07acb82c3cd930e0645292325" gracePeriod=10 Nov 25 15:19:58 crc kubenswrapper[4890]: I1125 15:19:51.989637 4890 generic.go:334] "Generic (PLEG): container finished" podID="df163dc8-8261-4fdf-b446-bba9e8e1d8eb" containerID="6a49096a7d81634a7d99776fad685186d59a98d07acb82c3cd930e0645292325" exitCode=0 Nov 25 15:19:58 crc kubenswrapper[4890]: I1125 15:19:51.989668 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-lgp8l" event={"ID":"df163dc8-8261-4fdf-b446-bba9e8e1d8eb","Type":"ContainerDied","Data":"6a49096a7d81634a7d99776fad685186d59a98d07acb82c3cd930e0645292325"} Nov 25 15:19:58 crc kubenswrapper[4890]: I1125 15:19:54.077022 4890 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-lgp8l" podUID="df163dc8-8261-4fdf-b446-bba9e8e1d8eb" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.116:5353: connect: connection refused" Nov 25 15:20:00 crc kubenswrapper[4890]: I1125 15:20:00.054643 4890 generic.go:334] "Generic (PLEG): container finished" podID="b803ac45-3f40-4188-a0cc-df469b30f1ea" containerID="90c930538f8ad9440ca2e8fefaa6d6abf1268230e5c9ba07b74a553ef0141b4d" exitCode=0 Nov 25 15:20:00 crc kubenswrapper[4890]: I1125 15:20:00.054815 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dblm6" event={"ID":"b803ac45-3f40-4188-a0cc-df469b30f1ea","Type":"ContainerDied","Data":"90c930538f8ad9440ca2e8fefaa6d6abf1268230e5c9ba07b74a553ef0141b4d"} Nov 25 15:20:03 crc kubenswrapper[4890]: E1125 15:20:03.659747 4890 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Nov 25 15:20:03 crc kubenswrapper[4890]: E1125 15:20:03.661610 4890 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n658h675h54fh5f8h678hc9h8dh85h544h688h5fh5c4h5c8h596h656h595h665hb8h654h597h548h56fh6dh5d6h557h684h645hch645h5b9h79h5fq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-skrsm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-7f7df64d4f-k6jfd_openstack(b246a631-f563-4f3f-aef3-caa0d92520c6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 15:20:03 crc kubenswrapper[4890]: E1125 15:20:03.665104 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-7f7df64d4f-k6jfd" podUID="b246a631-f563-4f3f-aef3-caa0d92520c6" Nov 25 15:20:04 crc kubenswrapper[4890]: I1125 15:20:04.077016 4890 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-lgp8l" podUID="df163dc8-8261-4fdf-b446-bba9e8e1d8eb" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.116:5353: i/o timeout" Nov 25 15:20:04 crc kubenswrapper[4890]: E1125 15:20:04.605273 4890 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Nov 25 15:20:04 crc kubenswrapper[4890]: E1125 15:20:04.605619 4890 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-szt2s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-f8ctf_openstack(fdcacd06-b857-46cf-8516-345e271cf882): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 15:20:04 crc kubenswrapper[4890]: E1125 15:20:04.606996 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-f8ctf" podUID="fdcacd06-b857-46cf-8516-345e271cf882" Nov 25 15:20:05 crc kubenswrapper[4890]: E1125 15:20:05.107574 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-f8ctf" podUID="fdcacd06-b857-46cf-8516-345e271cf882" Nov 25 15:20:09 crc kubenswrapper[4890]: I1125 15:20:09.077981 4890 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-lgp8l" podUID="df163dc8-8261-4fdf-b446-bba9e8e1d8eb" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.116:5353: i/o timeout" Nov 25 15:20:09 crc kubenswrapper[4890]: I1125 15:20:09.078653 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-lgp8l" Nov 25 15:20:12 crc kubenswrapper[4890]: E1125 15:20:12.022751 4890 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Nov 25 15:20:12 crc kubenswrapper[4890]: E1125 15:20:12.023268 4890 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pw2ct,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-j7w9k_openstack(997a078a-52bd-4eb2-8fde-f80d05d9b18c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 15:20:12 crc kubenswrapper[4890]: E1125 15:20:12.024489 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-j7w9k" podUID="997a078a-52bd-4eb2-8fde-f80d05d9b18c" Nov 25 15:20:12 crc kubenswrapper[4890]: E1125 15:20:12.154938 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-j7w9k" podUID="997a078a-52bd-4eb2-8fde-f80d05d9b18c" Nov 25 15:20:12 crc kubenswrapper[4890]: E1125 15:20:12.780697 4890 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Nov 25 15:20:12 crc kubenswrapper[4890]: E1125 15:20:12.780866 4890 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nf5h678h566h5f4h59bhd6h54bh574hbch5dh59bhdch689h544h649h566h9h666hf6h5ffh67fh59ch66chb9h68fh549hd7h55h659h595h578h568q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tqntl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-bf56cc7bf-tn5lx_openstack(d64cf59f-da15-4183-be46-13510f3f6efb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 15:20:12 crc kubenswrapper[4890]: E1125 15:20:12.783101 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-bf56cc7bf-tn5lx" podUID="d64cf59f-da15-4183-be46-13510f3f6efb" Nov 25 15:20:13 crc kubenswrapper[4890]: E1125 15:20:13.296635 4890 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Nov 25 15:20:13 crc kubenswrapper[4890]: E1125 15:20:13.297232 4890 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n56h557h599hch666h545h559h58fh68fh55dh559h5hfh59bh59fh56hd5h597h97h95hfdh66fh55fhd6h696hc5h5b8h64ch574hc6h599h58q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5q48k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-568c5cd49-gz9n4_openstack(e40cb582-ce6c-4064-8c1b-16fb4f7f3e62): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 15:20:13 crc kubenswrapper[4890]: E1125 15:20:13.300657 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-568c5cd49-gz9n4" podUID="e40cb582-ce6c-4064-8c1b-16fb4f7f3e62" Nov 25 15:20:14 crc kubenswrapper[4890]: I1125 15:20:14.081573 4890 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-lgp8l" podUID="df163dc8-8261-4fdf-b446-bba9e8e1d8eb" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.116:5353: i/o timeout" Nov 25 15:20:19 crc kubenswrapper[4890]: I1125 15:20:19.082448 4890 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-lgp8l" podUID="df163dc8-8261-4fdf-b446-bba9e8e1d8eb" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.116:5353: i/o timeout" Nov 25 15:20:24 crc kubenswrapper[4890]: I1125 15:20:24.084294 4890 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-lgp8l" podUID="df163dc8-8261-4fdf-b446-bba9e8e1d8eb" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.116:5353: i/o timeout" Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.502755 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-nrbtc-config-f5xmh" Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.513769 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-lgp8l" Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.525434 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dblm6" Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.646237 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-52d55\" (UniqueName: \"kubernetes.io/projected/84169bad-df45-41ac-92ca-356122dec8f3-kube-api-access-52d55\") pod \"84169bad-df45-41ac-92ca-356122dec8f3\" (UID: \"84169bad-df45-41ac-92ca-356122dec8f3\") " Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.646291 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b803ac45-3f40-4188-a0cc-df469b30f1ea-scripts\") pod \"b803ac45-3f40-4188-a0cc-df469b30f1ea\" (UID: \"b803ac45-3f40-4188-a0cc-df469b30f1ea\") " Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.646351 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/84169bad-df45-41ac-92ca-356122dec8f3-var-log-ovn\") pod \"84169bad-df45-41ac-92ca-356122dec8f3\" (UID: \"84169bad-df45-41ac-92ca-356122dec8f3\") " Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.646389 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/df163dc8-8261-4fdf-b446-bba9e8e1d8eb-dns-svc\") pod \"df163dc8-8261-4fdf-b446-bba9e8e1d8eb\" (UID: \"df163dc8-8261-4fdf-b446-bba9e8e1d8eb\") " Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.646421 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b803ac45-3f40-4188-a0cc-df469b30f1ea-combined-ca-bundle\") pod \"b803ac45-3f40-4188-a0cc-df469b30f1ea\" (UID: \"b803ac45-3f40-4188-a0cc-df469b30f1ea\") " Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.646447 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rp6l8\" (UniqueName: \"kubernetes.io/projected/b803ac45-3f40-4188-a0cc-df469b30f1ea-kube-api-access-rp6l8\") pod \"b803ac45-3f40-4188-a0cc-df469b30f1ea\" (UID: \"b803ac45-3f40-4188-a0cc-df469b30f1ea\") " Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.646506 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/84169bad-df45-41ac-92ca-356122dec8f3-var-run-ovn\") pod \"84169bad-df45-41ac-92ca-356122dec8f3\" (UID: \"84169bad-df45-41ac-92ca-356122dec8f3\") " Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.646507 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/84169bad-df45-41ac-92ca-356122dec8f3-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "84169bad-df45-41ac-92ca-356122dec8f3" (UID: "84169bad-df45-41ac-92ca-356122dec8f3"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.646530 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/84169bad-df45-41ac-92ca-356122dec8f3-additional-scripts\") pod \"84169bad-df45-41ac-92ca-356122dec8f3\" (UID: \"84169bad-df45-41ac-92ca-356122dec8f3\") " Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.646560 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/84169bad-df45-41ac-92ca-356122dec8f3-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "84169bad-df45-41ac-92ca-356122dec8f3" (UID: "84169bad-df45-41ac-92ca-356122dec8f3"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.646579 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v28mf\" (UniqueName: \"kubernetes.io/projected/df163dc8-8261-4fdf-b446-bba9e8e1d8eb-kube-api-access-v28mf\") pod \"df163dc8-8261-4fdf-b446-bba9e8e1d8eb\" (UID: \"df163dc8-8261-4fdf-b446-bba9e8e1d8eb\") " Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.646618 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df163dc8-8261-4fdf-b446-bba9e8e1d8eb-config\") pod \"df163dc8-8261-4fdf-b446-bba9e8e1d8eb\" (UID: \"df163dc8-8261-4fdf-b446-bba9e8e1d8eb\") " Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.646678 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/84169bad-df45-41ac-92ca-356122dec8f3-var-run\") pod \"84169bad-df45-41ac-92ca-356122dec8f3\" (UID: \"84169bad-df45-41ac-92ca-356122dec8f3\") " Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.646708 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b803ac45-3f40-4188-a0cc-df469b30f1ea-credential-keys\") pod \"b803ac45-3f40-4188-a0cc-df469b30f1ea\" (UID: \"b803ac45-3f40-4188-a0cc-df469b30f1ea\") " Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.646801 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/df163dc8-8261-4fdf-b446-bba9e8e1d8eb-ovsdbserver-sb\") pod \"df163dc8-8261-4fdf-b446-bba9e8e1d8eb\" (UID: \"df163dc8-8261-4fdf-b446-bba9e8e1d8eb\") " Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.646833 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b803ac45-3f40-4188-a0cc-df469b30f1ea-config-data\") pod \"b803ac45-3f40-4188-a0cc-df469b30f1ea\" (UID: \"b803ac45-3f40-4188-a0cc-df469b30f1ea\") " Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.646842 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/84169bad-df45-41ac-92ca-356122dec8f3-var-run" (OuterVolumeSpecName: "var-run") pod "84169bad-df45-41ac-92ca-356122dec8f3" (UID: "84169bad-df45-41ac-92ca-356122dec8f3"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.646898 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b803ac45-3f40-4188-a0cc-df469b30f1ea-fernet-keys\") pod \"b803ac45-3f40-4188-a0cc-df469b30f1ea\" (UID: \"b803ac45-3f40-4188-a0cc-df469b30f1ea\") " Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.646929 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/84169bad-df45-41ac-92ca-356122dec8f3-scripts\") pod \"84169bad-df45-41ac-92ca-356122dec8f3\" (UID: \"84169bad-df45-41ac-92ca-356122dec8f3\") " Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.646967 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/df163dc8-8261-4fdf-b446-bba9e8e1d8eb-ovsdbserver-nb\") pod \"df163dc8-8261-4fdf-b446-bba9e8e1d8eb\" (UID: \"df163dc8-8261-4fdf-b446-bba9e8e1d8eb\") " Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.647134 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84169bad-df45-41ac-92ca-356122dec8f3-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "84169bad-df45-41ac-92ca-356122dec8f3" (UID: "84169bad-df45-41ac-92ca-356122dec8f3"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.647470 4890 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/84169bad-df45-41ac-92ca-356122dec8f3-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.647488 4890 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/84169bad-df45-41ac-92ca-356122dec8f3-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.647499 4890 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/84169bad-df45-41ac-92ca-356122dec8f3-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.647511 4890 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/84169bad-df45-41ac-92ca-356122dec8f3-var-run\") on node \"crc\" DevicePath \"\"" Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.653352 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b803ac45-3f40-4188-a0cc-df469b30f1ea-scripts" (OuterVolumeSpecName: "scripts") pod "b803ac45-3f40-4188-a0cc-df469b30f1ea" (UID: "b803ac45-3f40-4188-a0cc-df469b30f1ea"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.653863 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df163dc8-8261-4fdf-b446-bba9e8e1d8eb-kube-api-access-v28mf" (OuterVolumeSpecName: "kube-api-access-v28mf") pod "df163dc8-8261-4fdf-b446-bba9e8e1d8eb" (UID: "df163dc8-8261-4fdf-b446-bba9e8e1d8eb"). InnerVolumeSpecName "kube-api-access-v28mf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.653948 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b803ac45-3f40-4188-a0cc-df469b30f1ea-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "b803ac45-3f40-4188-a0cc-df469b30f1ea" (UID: "b803ac45-3f40-4188-a0cc-df469b30f1ea"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.653953 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84169bad-df45-41ac-92ca-356122dec8f3-kube-api-access-52d55" (OuterVolumeSpecName: "kube-api-access-52d55") pod "84169bad-df45-41ac-92ca-356122dec8f3" (UID: "84169bad-df45-41ac-92ca-356122dec8f3"). InnerVolumeSpecName "kube-api-access-52d55". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.654344 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b803ac45-3f40-4188-a0cc-df469b30f1ea-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "b803ac45-3f40-4188-a0cc-df469b30f1ea" (UID: "b803ac45-3f40-4188-a0cc-df469b30f1ea"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.654764 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84169bad-df45-41ac-92ca-356122dec8f3-scripts" (OuterVolumeSpecName: "scripts") pod "84169bad-df45-41ac-92ca-356122dec8f3" (UID: "84169bad-df45-41ac-92ca-356122dec8f3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.655851 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b803ac45-3f40-4188-a0cc-df469b30f1ea-kube-api-access-rp6l8" (OuterVolumeSpecName: "kube-api-access-rp6l8") pod "b803ac45-3f40-4188-a0cc-df469b30f1ea" (UID: "b803ac45-3f40-4188-a0cc-df469b30f1ea"). InnerVolumeSpecName "kube-api-access-rp6l8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.674883 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b803ac45-3f40-4188-a0cc-df469b30f1ea-config-data" (OuterVolumeSpecName: "config-data") pod "b803ac45-3f40-4188-a0cc-df469b30f1ea" (UID: "b803ac45-3f40-4188-a0cc-df469b30f1ea"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.675370 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b803ac45-3f40-4188-a0cc-df469b30f1ea-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b803ac45-3f40-4188-a0cc-df469b30f1ea" (UID: "b803ac45-3f40-4188-a0cc-df469b30f1ea"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.696675 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df163dc8-8261-4fdf-b446-bba9e8e1d8eb-config" (OuterVolumeSpecName: "config") pod "df163dc8-8261-4fdf-b446-bba9e8e1d8eb" (UID: "df163dc8-8261-4fdf-b446-bba9e8e1d8eb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.697950 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df163dc8-8261-4fdf-b446-bba9e8e1d8eb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "df163dc8-8261-4fdf-b446-bba9e8e1d8eb" (UID: "df163dc8-8261-4fdf-b446-bba9e8e1d8eb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.698399 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df163dc8-8261-4fdf-b446-bba9e8e1d8eb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "df163dc8-8261-4fdf-b446-bba9e8e1d8eb" (UID: "df163dc8-8261-4fdf-b446-bba9e8e1d8eb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.701229 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df163dc8-8261-4fdf-b446-bba9e8e1d8eb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "df163dc8-8261-4fdf-b446-bba9e8e1d8eb" (UID: "df163dc8-8261-4fdf-b446-bba9e8e1d8eb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.749941 4890 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/df163dc8-8261-4fdf-b446-bba9e8e1d8eb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.749978 4890 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b803ac45-3f40-4188-a0cc-df469b30f1ea-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.749990 4890 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b803ac45-3f40-4188-a0cc-df469b30f1ea-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.750002 4890 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/84169bad-df45-41ac-92ca-356122dec8f3-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.750015 4890 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/df163dc8-8261-4fdf-b446-bba9e8e1d8eb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.750026 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-52d55\" (UniqueName: \"kubernetes.io/projected/84169bad-df45-41ac-92ca-356122dec8f3-kube-api-access-52d55\") on node \"crc\" DevicePath \"\"" Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.750037 4890 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b803ac45-3f40-4188-a0cc-df469b30f1ea-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.750046 4890 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/df163dc8-8261-4fdf-b446-bba9e8e1d8eb-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.750054 4890 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b803ac45-3f40-4188-a0cc-df469b30f1ea-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.750061 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rp6l8\" (UniqueName: \"kubernetes.io/projected/b803ac45-3f40-4188-a0cc-df469b30f1ea-kube-api-access-rp6l8\") on node \"crc\" DevicePath \"\"" Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.750069 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v28mf\" (UniqueName: \"kubernetes.io/projected/df163dc8-8261-4fdf-b446-bba9e8e1d8eb-kube-api-access-v28mf\") on node \"crc\" DevicePath \"\"" Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.750077 4890 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df163dc8-8261-4fdf-b446-bba9e8e1d8eb-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:20:25 crc kubenswrapper[4890]: I1125 15:20:25.750084 4890 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b803ac45-3f40-4188-a0cc-df469b30f1ea-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 25 15:20:26 crc kubenswrapper[4890]: I1125 15:20:26.272415 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-nrbtc-config-f5xmh" event={"ID":"84169bad-df45-41ac-92ca-356122dec8f3","Type":"ContainerDied","Data":"76823c9bff48e184abe27bff50d1ec3e5677a49a05a2cd32ea24cf1836802ff0"} Nov 25 15:20:26 crc kubenswrapper[4890]: I1125 15:20:26.272526 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="76823c9bff48e184abe27bff50d1ec3e5677a49a05a2cd32ea24cf1836802ff0" Nov 25 15:20:26 crc kubenswrapper[4890]: I1125 15:20:26.272737 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-nrbtc-config-f5xmh" Nov 25 15:20:26 crc kubenswrapper[4890]: I1125 15:20:26.274474 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dblm6" event={"ID":"b803ac45-3f40-4188-a0cc-df469b30f1ea","Type":"ContainerDied","Data":"9082a1f88dc7b7666d50eccc3dd2b60d2075e9f08cec6a5df2fda5a7c6090379"} Nov 25 15:20:26 crc kubenswrapper[4890]: I1125 15:20:26.274586 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9082a1f88dc7b7666d50eccc3dd2b60d2075e9f08cec6a5df2fda5a7c6090379" Nov 25 15:20:26 crc kubenswrapper[4890]: I1125 15:20:26.274529 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dblm6" Nov 25 15:20:26 crc kubenswrapper[4890]: I1125 15:20:26.276196 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-lgp8l" event={"ID":"df163dc8-8261-4fdf-b446-bba9e8e1d8eb","Type":"ContainerDied","Data":"d5aeafc5e22654c2b144daa5cac397a3e4424c28b802dcf6c6bfa7f95aa6ad8f"} Nov 25 15:20:26 crc kubenswrapper[4890]: I1125 15:20:26.276415 4890 scope.go:117] "RemoveContainer" containerID="6a49096a7d81634a7d99776fad685186d59a98d07acb82c3cd930e0645292325" Nov 25 15:20:26 crc kubenswrapper[4890]: I1125 15:20:26.276379 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-lgp8l" Nov 25 15:20:26 crc kubenswrapper[4890]: I1125 15:20:26.307672 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-lgp8l"] Nov 25 15:20:26 crc kubenswrapper[4890]: I1125 15:20:26.313972 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-lgp8l"] Nov 25 15:20:26 crc kubenswrapper[4890]: I1125 15:20:26.447912 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 15:20:26 crc kubenswrapper[4890]: I1125 15:20:26.447997 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 15:20:26 crc kubenswrapper[4890]: I1125 15:20:26.598361 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-nrbtc-config-f5xmh"] Nov 25 15:20:26 crc kubenswrapper[4890]: I1125 15:20:26.604143 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-nrbtc-config-f5xmh"] Nov 25 15:20:26 crc kubenswrapper[4890]: I1125 15:20:26.622298 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-dblm6"] Nov 25 15:20:26 crc kubenswrapper[4890]: I1125 15:20:26.628670 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-dblm6"] Nov 25 15:20:26 crc kubenswrapper[4890]: I1125 15:20:26.730662 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-2dtqm"] Nov 25 15:20:26 crc kubenswrapper[4890]: E1125 15:20:26.731119 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84169bad-df45-41ac-92ca-356122dec8f3" containerName="ovn-config" Nov 25 15:20:26 crc kubenswrapper[4890]: I1125 15:20:26.731137 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="84169bad-df45-41ac-92ca-356122dec8f3" containerName="ovn-config" Nov 25 15:20:26 crc kubenswrapper[4890]: E1125 15:20:26.731176 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df163dc8-8261-4fdf-b446-bba9e8e1d8eb" containerName="init" Nov 25 15:20:26 crc kubenswrapper[4890]: I1125 15:20:26.731184 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="df163dc8-8261-4fdf-b446-bba9e8e1d8eb" containerName="init" Nov 25 15:20:26 crc kubenswrapper[4890]: E1125 15:20:26.731199 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df163dc8-8261-4fdf-b446-bba9e8e1d8eb" containerName="dnsmasq-dns" Nov 25 15:20:26 crc kubenswrapper[4890]: I1125 15:20:26.731208 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="df163dc8-8261-4fdf-b446-bba9e8e1d8eb" containerName="dnsmasq-dns" Nov 25 15:20:26 crc kubenswrapper[4890]: E1125 15:20:26.731219 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b803ac45-3f40-4188-a0cc-df469b30f1ea" containerName="keystone-bootstrap" Nov 25 15:20:26 crc kubenswrapper[4890]: I1125 15:20:26.731225 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="b803ac45-3f40-4188-a0cc-df469b30f1ea" containerName="keystone-bootstrap" Nov 25 15:20:26 crc kubenswrapper[4890]: I1125 15:20:26.731503 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="84169bad-df45-41ac-92ca-356122dec8f3" containerName="ovn-config" Nov 25 15:20:26 crc kubenswrapper[4890]: I1125 15:20:26.731523 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="df163dc8-8261-4fdf-b446-bba9e8e1d8eb" containerName="dnsmasq-dns" Nov 25 15:20:26 crc kubenswrapper[4890]: I1125 15:20:26.731578 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="b803ac45-3f40-4188-a0cc-df469b30f1ea" containerName="keystone-bootstrap" Nov 25 15:20:26 crc kubenswrapper[4890]: I1125 15:20:26.732363 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2dtqm" Nov 25 15:20:26 crc kubenswrapper[4890]: I1125 15:20:26.734192 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 25 15:20:26 crc kubenswrapper[4890]: I1125 15:20:26.734689 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 25 15:20:26 crc kubenswrapper[4890]: I1125 15:20:26.734716 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-t8zql" Nov 25 15:20:26 crc kubenswrapper[4890]: I1125 15:20:26.734736 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 25 15:20:26 crc kubenswrapper[4890]: I1125 15:20:26.738592 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 25 15:20:26 crc kubenswrapper[4890]: I1125 15:20:26.745454 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-2dtqm"] Nov 25 15:20:26 crc kubenswrapper[4890]: I1125 15:20:26.881231 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7354c28-0b1f-4dda-9b1b-42dcd2657a7a-scripts\") pod \"keystone-bootstrap-2dtqm\" (UID: \"c7354c28-0b1f-4dda-9b1b-42dcd2657a7a\") " pod="openstack/keystone-bootstrap-2dtqm" Nov 25 15:20:26 crc kubenswrapper[4890]: I1125 15:20:26.881309 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c7354c28-0b1f-4dda-9b1b-42dcd2657a7a-credential-keys\") pod \"keystone-bootstrap-2dtqm\" (UID: \"c7354c28-0b1f-4dda-9b1b-42dcd2657a7a\") " pod="openstack/keystone-bootstrap-2dtqm" Nov 25 15:20:26 crc kubenswrapper[4890]: I1125 15:20:26.881331 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7354c28-0b1f-4dda-9b1b-42dcd2657a7a-config-data\") pod \"keystone-bootstrap-2dtqm\" (UID: \"c7354c28-0b1f-4dda-9b1b-42dcd2657a7a\") " pod="openstack/keystone-bootstrap-2dtqm" Nov 25 15:20:26 crc kubenswrapper[4890]: I1125 15:20:26.881471 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7354c28-0b1f-4dda-9b1b-42dcd2657a7a-combined-ca-bundle\") pod \"keystone-bootstrap-2dtqm\" (UID: \"c7354c28-0b1f-4dda-9b1b-42dcd2657a7a\") " pod="openstack/keystone-bootstrap-2dtqm" Nov 25 15:20:26 crc kubenswrapper[4890]: I1125 15:20:26.881553 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9rbh\" (UniqueName: \"kubernetes.io/projected/c7354c28-0b1f-4dda-9b1b-42dcd2657a7a-kube-api-access-d9rbh\") pod \"keystone-bootstrap-2dtqm\" (UID: \"c7354c28-0b1f-4dda-9b1b-42dcd2657a7a\") " pod="openstack/keystone-bootstrap-2dtqm" Nov 25 15:20:26 crc kubenswrapper[4890]: I1125 15:20:26.882228 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c7354c28-0b1f-4dda-9b1b-42dcd2657a7a-fernet-keys\") pod \"keystone-bootstrap-2dtqm\" (UID: \"c7354c28-0b1f-4dda-9b1b-42dcd2657a7a\") " pod="openstack/keystone-bootstrap-2dtqm" Nov 25 15:20:26 crc kubenswrapper[4890]: I1125 15:20:26.983494 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c7354c28-0b1f-4dda-9b1b-42dcd2657a7a-fernet-keys\") pod \"keystone-bootstrap-2dtqm\" (UID: \"c7354c28-0b1f-4dda-9b1b-42dcd2657a7a\") " pod="openstack/keystone-bootstrap-2dtqm" Nov 25 15:20:26 crc kubenswrapper[4890]: I1125 15:20:26.983568 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7354c28-0b1f-4dda-9b1b-42dcd2657a7a-scripts\") pod \"keystone-bootstrap-2dtqm\" (UID: \"c7354c28-0b1f-4dda-9b1b-42dcd2657a7a\") " pod="openstack/keystone-bootstrap-2dtqm" Nov 25 15:20:26 crc kubenswrapper[4890]: I1125 15:20:26.983618 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c7354c28-0b1f-4dda-9b1b-42dcd2657a7a-credential-keys\") pod \"keystone-bootstrap-2dtqm\" (UID: \"c7354c28-0b1f-4dda-9b1b-42dcd2657a7a\") " pod="openstack/keystone-bootstrap-2dtqm" Nov 25 15:20:26 crc kubenswrapper[4890]: I1125 15:20:26.983676 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7354c28-0b1f-4dda-9b1b-42dcd2657a7a-config-data\") pod \"keystone-bootstrap-2dtqm\" (UID: \"c7354c28-0b1f-4dda-9b1b-42dcd2657a7a\") " pod="openstack/keystone-bootstrap-2dtqm" Nov 25 15:20:26 crc kubenswrapper[4890]: I1125 15:20:26.983698 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7354c28-0b1f-4dda-9b1b-42dcd2657a7a-combined-ca-bundle\") pod \"keystone-bootstrap-2dtqm\" (UID: \"c7354c28-0b1f-4dda-9b1b-42dcd2657a7a\") " pod="openstack/keystone-bootstrap-2dtqm" Nov 25 15:20:26 crc kubenswrapper[4890]: I1125 15:20:26.983730 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9rbh\" (UniqueName: \"kubernetes.io/projected/c7354c28-0b1f-4dda-9b1b-42dcd2657a7a-kube-api-access-d9rbh\") pod \"keystone-bootstrap-2dtqm\" (UID: \"c7354c28-0b1f-4dda-9b1b-42dcd2657a7a\") " pod="openstack/keystone-bootstrap-2dtqm" Nov 25 15:20:26 crc kubenswrapper[4890]: I1125 15:20:26.991602 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7354c28-0b1f-4dda-9b1b-42dcd2657a7a-scripts\") pod \"keystone-bootstrap-2dtqm\" (UID: \"c7354c28-0b1f-4dda-9b1b-42dcd2657a7a\") " pod="openstack/keystone-bootstrap-2dtqm" Nov 25 15:20:26 crc kubenswrapper[4890]: I1125 15:20:26.991955 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c7354c28-0b1f-4dda-9b1b-42dcd2657a7a-fernet-keys\") pod \"keystone-bootstrap-2dtqm\" (UID: \"c7354c28-0b1f-4dda-9b1b-42dcd2657a7a\") " pod="openstack/keystone-bootstrap-2dtqm" Nov 25 15:20:26 crc kubenswrapper[4890]: I1125 15:20:26.992640 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7354c28-0b1f-4dda-9b1b-42dcd2657a7a-config-data\") pod \"keystone-bootstrap-2dtqm\" (UID: \"c7354c28-0b1f-4dda-9b1b-42dcd2657a7a\") " pod="openstack/keystone-bootstrap-2dtqm" Nov 25 15:20:26 crc kubenswrapper[4890]: I1125 15:20:26.996716 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7354c28-0b1f-4dda-9b1b-42dcd2657a7a-combined-ca-bundle\") pod \"keystone-bootstrap-2dtqm\" (UID: \"c7354c28-0b1f-4dda-9b1b-42dcd2657a7a\") " pod="openstack/keystone-bootstrap-2dtqm" Nov 25 15:20:26 crc kubenswrapper[4890]: I1125 15:20:26.997862 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c7354c28-0b1f-4dda-9b1b-42dcd2657a7a-credential-keys\") pod \"keystone-bootstrap-2dtqm\" (UID: \"c7354c28-0b1f-4dda-9b1b-42dcd2657a7a\") " pod="openstack/keystone-bootstrap-2dtqm" Nov 25 15:20:27 crc kubenswrapper[4890]: I1125 15:20:27.000593 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9rbh\" (UniqueName: \"kubernetes.io/projected/c7354c28-0b1f-4dda-9b1b-42dcd2657a7a-kube-api-access-d9rbh\") pod \"keystone-bootstrap-2dtqm\" (UID: \"c7354c28-0b1f-4dda-9b1b-42dcd2657a7a\") " pod="openstack/keystone-bootstrap-2dtqm" Nov 25 15:20:27 crc kubenswrapper[4890]: I1125 15:20:27.059007 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2dtqm" Nov 25 15:20:27 crc kubenswrapper[4890]: I1125 15:20:27.727965 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f7df64d4f-k6jfd" Nov 25 15:20:27 crc kubenswrapper[4890]: I1125 15:20:27.798385 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b246a631-f563-4f3f-aef3-caa0d92520c6-config-data\") pod \"b246a631-f563-4f3f-aef3-caa0d92520c6\" (UID: \"b246a631-f563-4f3f-aef3-caa0d92520c6\") " Nov 25 15:20:27 crc kubenswrapper[4890]: I1125 15:20:27.798497 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b246a631-f563-4f3f-aef3-caa0d92520c6-scripts\") pod \"b246a631-f563-4f3f-aef3-caa0d92520c6\" (UID: \"b246a631-f563-4f3f-aef3-caa0d92520c6\") " Nov 25 15:20:27 crc kubenswrapper[4890]: I1125 15:20:27.798547 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b246a631-f563-4f3f-aef3-caa0d92520c6-horizon-secret-key\") pod \"b246a631-f563-4f3f-aef3-caa0d92520c6\" (UID: \"b246a631-f563-4f3f-aef3-caa0d92520c6\") " Nov 25 15:20:27 crc kubenswrapper[4890]: I1125 15:20:27.798671 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b246a631-f563-4f3f-aef3-caa0d92520c6-logs\") pod \"b246a631-f563-4f3f-aef3-caa0d92520c6\" (UID: \"b246a631-f563-4f3f-aef3-caa0d92520c6\") " Nov 25 15:20:27 crc kubenswrapper[4890]: I1125 15:20:27.798800 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-skrsm\" (UniqueName: \"kubernetes.io/projected/b246a631-f563-4f3f-aef3-caa0d92520c6-kube-api-access-skrsm\") pod \"b246a631-f563-4f3f-aef3-caa0d92520c6\" (UID: \"b246a631-f563-4f3f-aef3-caa0d92520c6\") " Nov 25 15:20:27 crc kubenswrapper[4890]: I1125 15:20:27.799189 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b246a631-f563-4f3f-aef3-caa0d92520c6-scripts" (OuterVolumeSpecName: "scripts") pod "b246a631-f563-4f3f-aef3-caa0d92520c6" (UID: "b246a631-f563-4f3f-aef3-caa0d92520c6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:20:27 crc kubenswrapper[4890]: I1125 15:20:27.799534 4890 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b246a631-f563-4f3f-aef3-caa0d92520c6-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 15:20:27 crc kubenswrapper[4890]: I1125 15:20:27.799589 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b246a631-f563-4f3f-aef3-caa0d92520c6-logs" (OuterVolumeSpecName: "logs") pod "b246a631-f563-4f3f-aef3-caa0d92520c6" (UID: "b246a631-f563-4f3f-aef3-caa0d92520c6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:20:27 crc kubenswrapper[4890]: I1125 15:20:27.801664 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b246a631-f563-4f3f-aef3-caa0d92520c6-config-data" (OuterVolumeSpecName: "config-data") pod "b246a631-f563-4f3f-aef3-caa0d92520c6" (UID: "b246a631-f563-4f3f-aef3-caa0d92520c6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:20:27 crc kubenswrapper[4890]: I1125 15:20:27.810370 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b246a631-f563-4f3f-aef3-caa0d92520c6-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "b246a631-f563-4f3f-aef3-caa0d92520c6" (UID: "b246a631-f563-4f3f-aef3-caa0d92520c6"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:20:27 crc kubenswrapper[4890]: I1125 15:20:27.818383 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b246a631-f563-4f3f-aef3-caa0d92520c6-kube-api-access-skrsm" (OuterVolumeSpecName: "kube-api-access-skrsm") pod "b246a631-f563-4f3f-aef3-caa0d92520c6" (UID: "b246a631-f563-4f3f-aef3-caa0d92520c6"). InnerVolumeSpecName "kube-api-access-skrsm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:20:27 crc kubenswrapper[4890]: I1125 15:20:27.900803 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-skrsm\" (UniqueName: \"kubernetes.io/projected/b246a631-f563-4f3f-aef3-caa0d92520c6-kube-api-access-skrsm\") on node \"crc\" DevicePath \"\"" Nov 25 15:20:27 crc kubenswrapper[4890]: I1125 15:20:27.900840 4890 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b246a631-f563-4f3f-aef3-caa0d92520c6-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 15:20:27 crc kubenswrapper[4890]: I1125 15:20:27.900852 4890 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b246a631-f563-4f3f-aef3-caa0d92520c6-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 25 15:20:27 crc kubenswrapper[4890]: I1125 15:20:27.900861 4890 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b246a631-f563-4f3f-aef3-caa0d92520c6-logs\") on node \"crc\" DevicePath \"\"" Nov 25 15:20:28 crc kubenswrapper[4890]: I1125 15:20:28.095280 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-556858788d-s2ppl"] Nov 25 15:20:28 crc kubenswrapper[4890]: E1125 15:20:28.126362 4890 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-swift-container:current-podified" Nov 25 15:20:28 crc kubenswrapper[4890]: E1125 15:20:28.126538 4890 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:container-server,Image:quay.io/podified-antelope-centos9/openstack-swift-container:current-podified,Command:[/usr/bin/swift-container-server /etc/swift/container-server.conf.d -v],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:container,HostPort:0,ContainerPort:6201,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5b7h56h9dh94h67bh697h95h55hbh555h556h675h5fdh57dh579h5fbh64fh5c9h687hb6h678h5d4h549h54h98h8ch564h5bh5bch55dhc8hf8q,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:swift,ReadOnly:false,MountPath:/srv/node/pv,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:etc-swift,ReadOnly:false,MountPath:/etc/swift,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cache,ReadOnly:false,MountPath:/var/cache/swift,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:lock,ReadOnly:false,MountPath:/var/lock,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8lp76,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42445,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-storage-0_openstack(6517688d-0723-4809-ba51-bcaf48524157): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 15:20:28 crc kubenswrapper[4890]: I1125 15:20:28.183017 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84169bad-df45-41ac-92ca-356122dec8f3" path="/var/lib/kubelet/pods/84169bad-df45-41ac-92ca-356122dec8f3/volumes" Nov 25 15:20:28 crc kubenswrapper[4890]: I1125 15:20:28.184103 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b803ac45-3f40-4188-a0cc-df469b30f1ea" path="/var/lib/kubelet/pods/b803ac45-3f40-4188-a0cc-df469b30f1ea/volumes" Nov 25 15:20:28 crc kubenswrapper[4890]: I1125 15:20:28.185025 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df163dc8-8261-4fdf-b446-bba9e8e1d8eb" path="/var/lib/kubelet/pods/df163dc8-8261-4fdf-b446-bba9e8e1d8eb/volumes" Nov 25 15:20:28 crc kubenswrapper[4890]: I1125 15:20:28.297414 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f7df64d4f-k6jfd" event={"ID":"b246a631-f563-4f3f-aef3-caa0d92520c6","Type":"ContainerDied","Data":"ed9e23083e7e95e1e0c845406cfae1a4059af7e8375cd11d1c0879b29033bb3f"} Nov 25 15:20:28 crc kubenswrapper[4890]: I1125 15:20:28.297568 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f7df64d4f-k6jfd" Nov 25 15:20:28 crc kubenswrapper[4890]: I1125 15:20:28.350692 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7f7df64d4f-k6jfd"] Nov 25 15:20:28 crc kubenswrapper[4890]: I1125 15:20:28.359215 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7f7df64d4f-k6jfd"] Nov 25 15:20:29 crc kubenswrapper[4890]: I1125 15:20:29.086005 4890 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-lgp8l" podUID="df163dc8-8261-4fdf-b446-bba9e8e1d8eb" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.116:5353: i/o timeout" Nov 25 15:20:29 crc kubenswrapper[4890]: I1125 15:20:29.205056 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-bf56cc7bf-tn5lx" Nov 25 15:20:29 crc kubenswrapper[4890]: I1125 15:20:29.213371 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-568c5cd49-gz9n4" Nov 25 15:20:29 crc kubenswrapper[4890]: I1125 15:20:29.306960 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-bf56cc7bf-tn5lx" event={"ID":"d64cf59f-da15-4183-be46-13510f3f6efb","Type":"ContainerDied","Data":"14b8c029bcac1c3a368275ab4b80a168117208e12013f7f2dafc0bad80497751"} Nov 25 15:20:29 crc kubenswrapper[4890]: I1125 15:20:29.306996 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-bf56cc7bf-tn5lx" Nov 25 15:20:29 crc kubenswrapper[4890]: I1125 15:20:29.308427 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-568c5cd49-gz9n4" event={"ID":"e40cb582-ce6c-4064-8c1b-16fb4f7f3e62","Type":"ContainerDied","Data":"85ed31693c35ff51013b13e0efe5e290540726e59a59b4df9cd1eee9522cfbcd"} Nov 25 15:20:29 crc kubenswrapper[4890]: I1125 15:20:29.308523 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-568c5cd49-gz9n4" Nov 25 15:20:29 crc kubenswrapper[4890]: I1125 15:20:29.324124 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5q48k\" (UniqueName: \"kubernetes.io/projected/e40cb582-ce6c-4064-8c1b-16fb4f7f3e62-kube-api-access-5q48k\") pod \"e40cb582-ce6c-4064-8c1b-16fb4f7f3e62\" (UID: \"e40cb582-ce6c-4064-8c1b-16fb4f7f3e62\") " Nov 25 15:20:29 crc kubenswrapper[4890]: I1125 15:20:29.324199 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d64cf59f-da15-4183-be46-13510f3f6efb-config-data\") pod \"d64cf59f-da15-4183-be46-13510f3f6efb\" (UID: \"d64cf59f-da15-4183-be46-13510f3f6efb\") " Nov 25 15:20:29 crc kubenswrapper[4890]: I1125 15:20:29.324224 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d64cf59f-da15-4183-be46-13510f3f6efb-logs\") pod \"d64cf59f-da15-4183-be46-13510f3f6efb\" (UID: \"d64cf59f-da15-4183-be46-13510f3f6efb\") " Nov 25 15:20:29 crc kubenswrapper[4890]: I1125 15:20:29.324289 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d64cf59f-da15-4183-be46-13510f3f6efb-horizon-secret-key\") pod \"d64cf59f-da15-4183-be46-13510f3f6efb\" (UID: \"d64cf59f-da15-4183-be46-13510f3f6efb\") " Nov 25 15:20:29 crc kubenswrapper[4890]: I1125 15:20:29.324342 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e40cb582-ce6c-4064-8c1b-16fb4f7f3e62-scripts\") pod \"e40cb582-ce6c-4064-8c1b-16fb4f7f3e62\" (UID: \"e40cb582-ce6c-4064-8c1b-16fb4f7f3e62\") " Nov 25 15:20:29 crc kubenswrapper[4890]: I1125 15:20:29.324384 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tqntl\" (UniqueName: \"kubernetes.io/projected/d64cf59f-da15-4183-be46-13510f3f6efb-kube-api-access-tqntl\") pod \"d64cf59f-da15-4183-be46-13510f3f6efb\" (UID: \"d64cf59f-da15-4183-be46-13510f3f6efb\") " Nov 25 15:20:29 crc kubenswrapper[4890]: I1125 15:20:29.324409 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d64cf59f-da15-4183-be46-13510f3f6efb-scripts\") pod \"d64cf59f-da15-4183-be46-13510f3f6efb\" (UID: \"d64cf59f-da15-4183-be46-13510f3f6efb\") " Nov 25 15:20:29 crc kubenswrapper[4890]: I1125 15:20:29.324705 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d64cf59f-da15-4183-be46-13510f3f6efb-logs" (OuterVolumeSpecName: "logs") pod "d64cf59f-da15-4183-be46-13510f3f6efb" (UID: "d64cf59f-da15-4183-be46-13510f3f6efb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:20:29 crc kubenswrapper[4890]: I1125 15:20:29.324866 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d64cf59f-da15-4183-be46-13510f3f6efb-scripts" (OuterVolumeSpecName: "scripts") pod "d64cf59f-da15-4183-be46-13510f3f6efb" (UID: "d64cf59f-da15-4183-be46-13510f3f6efb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:20:29 crc kubenswrapper[4890]: I1125 15:20:29.324911 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e40cb582-ce6c-4064-8c1b-16fb4f7f3e62-scripts" (OuterVolumeSpecName: "scripts") pod "e40cb582-ce6c-4064-8c1b-16fb4f7f3e62" (UID: "e40cb582-ce6c-4064-8c1b-16fb4f7f3e62"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:20:29 crc kubenswrapper[4890]: I1125 15:20:29.325096 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e40cb582-ce6c-4064-8c1b-16fb4f7f3e62-logs\") pod \"e40cb582-ce6c-4064-8c1b-16fb4f7f3e62\" (UID: \"e40cb582-ce6c-4064-8c1b-16fb4f7f3e62\") " Nov 25 15:20:29 crc kubenswrapper[4890]: I1125 15:20:29.325222 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e40cb582-ce6c-4064-8c1b-16fb4f7f3e62-horizon-secret-key\") pod \"e40cb582-ce6c-4064-8c1b-16fb4f7f3e62\" (UID: \"e40cb582-ce6c-4064-8c1b-16fb4f7f3e62\") " Nov 25 15:20:29 crc kubenswrapper[4890]: I1125 15:20:29.325339 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e40cb582-ce6c-4064-8c1b-16fb4f7f3e62-config-data\") pod \"e40cb582-ce6c-4064-8c1b-16fb4f7f3e62\" (UID: \"e40cb582-ce6c-4064-8c1b-16fb4f7f3e62\") " Nov 25 15:20:29 crc kubenswrapper[4890]: I1125 15:20:29.325549 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e40cb582-ce6c-4064-8c1b-16fb4f7f3e62-logs" (OuterVolumeSpecName: "logs") pod "e40cb582-ce6c-4064-8c1b-16fb4f7f3e62" (UID: "e40cb582-ce6c-4064-8c1b-16fb4f7f3e62"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:20:29 crc kubenswrapper[4890]: I1125 15:20:29.325881 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d64cf59f-da15-4183-be46-13510f3f6efb-config-data" (OuterVolumeSpecName: "config-data") pod "d64cf59f-da15-4183-be46-13510f3f6efb" (UID: "d64cf59f-da15-4183-be46-13510f3f6efb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:20:29 crc kubenswrapper[4890]: I1125 15:20:29.325926 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e40cb582-ce6c-4064-8c1b-16fb4f7f3e62-config-data" (OuterVolumeSpecName: "config-data") pod "e40cb582-ce6c-4064-8c1b-16fb4f7f3e62" (UID: "e40cb582-ce6c-4064-8c1b-16fb4f7f3e62"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:20:29 crc kubenswrapper[4890]: I1125 15:20:29.326690 4890 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e40cb582-ce6c-4064-8c1b-16fb4f7f3e62-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 15:20:29 crc kubenswrapper[4890]: I1125 15:20:29.326728 4890 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d64cf59f-da15-4183-be46-13510f3f6efb-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 15:20:29 crc kubenswrapper[4890]: I1125 15:20:29.326744 4890 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e40cb582-ce6c-4064-8c1b-16fb4f7f3e62-logs\") on node \"crc\" DevicePath \"\"" Nov 25 15:20:29 crc kubenswrapper[4890]: I1125 15:20:29.326757 4890 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e40cb582-ce6c-4064-8c1b-16fb4f7f3e62-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 15:20:29 crc kubenswrapper[4890]: I1125 15:20:29.326768 4890 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d64cf59f-da15-4183-be46-13510f3f6efb-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 15:20:29 crc kubenswrapper[4890]: I1125 15:20:29.326789 4890 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d64cf59f-da15-4183-be46-13510f3f6efb-logs\") on node \"crc\" DevicePath \"\"" Nov 25 15:20:29 crc kubenswrapper[4890]: I1125 15:20:29.329561 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e40cb582-ce6c-4064-8c1b-16fb4f7f3e62-kube-api-access-5q48k" (OuterVolumeSpecName: "kube-api-access-5q48k") pod "e40cb582-ce6c-4064-8c1b-16fb4f7f3e62" (UID: "e40cb582-ce6c-4064-8c1b-16fb4f7f3e62"). InnerVolumeSpecName "kube-api-access-5q48k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:20:29 crc kubenswrapper[4890]: I1125 15:20:29.329912 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d64cf59f-da15-4183-be46-13510f3f6efb-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "d64cf59f-da15-4183-be46-13510f3f6efb" (UID: "d64cf59f-da15-4183-be46-13510f3f6efb"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:20:29 crc kubenswrapper[4890]: I1125 15:20:29.329928 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e40cb582-ce6c-4064-8c1b-16fb4f7f3e62-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "e40cb582-ce6c-4064-8c1b-16fb4f7f3e62" (UID: "e40cb582-ce6c-4064-8c1b-16fb4f7f3e62"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:20:29 crc kubenswrapper[4890]: I1125 15:20:29.330296 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d64cf59f-da15-4183-be46-13510f3f6efb-kube-api-access-tqntl" (OuterVolumeSpecName: "kube-api-access-tqntl") pod "d64cf59f-da15-4183-be46-13510f3f6efb" (UID: "d64cf59f-da15-4183-be46-13510f3f6efb"). InnerVolumeSpecName "kube-api-access-tqntl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:20:29 crc kubenswrapper[4890]: I1125 15:20:29.427896 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tqntl\" (UniqueName: \"kubernetes.io/projected/d64cf59f-da15-4183-be46-13510f3f6efb-kube-api-access-tqntl\") on node \"crc\" DevicePath \"\"" Nov 25 15:20:29 crc kubenswrapper[4890]: I1125 15:20:29.427931 4890 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e40cb582-ce6c-4064-8c1b-16fb4f7f3e62-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 25 15:20:29 crc kubenswrapper[4890]: I1125 15:20:29.427942 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5q48k\" (UniqueName: \"kubernetes.io/projected/e40cb582-ce6c-4064-8c1b-16fb4f7f3e62-kube-api-access-5q48k\") on node \"crc\" DevicePath \"\"" Nov 25 15:20:29 crc kubenswrapper[4890]: I1125 15:20:29.427954 4890 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d64cf59f-da15-4183-be46-13510f3f6efb-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 25 15:20:29 crc kubenswrapper[4890]: I1125 15:20:29.670409 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-568c5cd49-gz9n4"] Nov 25 15:20:29 crc kubenswrapper[4890]: I1125 15:20:29.670456 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-568c5cd49-gz9n4"] Nov 25 15:20:29 crc kubenswrapper[4890]: I1125 15:20:29.754154 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-bf56cc7bf-tn5lx"] Nov 25 15:20:29 crc kubenswrapper[4890]: I1125 15:20:29.761458 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-bf56cc7bf-tn5lx"] Nov 25 15:20:30 crc kubenswrapper[4890]: I1125 15:20:30.186375 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b246a631-f563-4f3f-aef3-caa0d92520c6" path="/var/lib/kubelet/pods/b246a631-f563-4f3f-aef3-caa0d92520c6/volumes" Nov 25 15:20:30 crc kubenswrapper[4890]: I1125 15:20:30.187273 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d64cf59f-da15-4183-be46-13510f3f6efb" path="/var/lib/kubelet/pods/d64cf59f-da15-4183-be46-13510f3f6efb/volumes" Nov 25 15:20:30 crc kubenswrapper[4890]: I1125 15:20:30.187823 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e40cb582-ce6c-4064-8c1b-16fb4f7f3e62" path="/var/lib/kubelet/pods/e40cb582-ce6c-4064-8c1b-16fb4f7f3e62/volumes" Nov 25 15:20:30 crc kubenswrapper[4890]: E1125 15:20:30.373753 4890 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Nov 25 15:20:30 crc kubenswrapper[4890]: E1125 15:20:30.373919 4890 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n8dh74h5b9h56dh55dhb9hd8h658h64dh584h599hcbh668hc9h6h694h57bh7bh654hbfh654h694h74h569h676h646h5d4hc5h576h688h66fhc9q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rk2gl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(1a886f74-d737-46b2-9647-232deec47e87): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 15:20:31 crc kubenswrapper[4890]: I1125 15:20:31.485810 4890 scope.go:117] "RemoveContainer" containerID="f8776797ac98bfd2201c355adc063e4b947dcb3f65103a6110c016067ba28acd" Nov 25 15:20:31 crc kubenswrapper[4890]: E1125 15:20:31.494310 4890 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Nov 25 15:20:31 crc kubenswrapper[4890]: E1125 15:20:31.494495 4890 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-q98k8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-wlvxc_openstack(449c7dd2-e3c8-48c2-8770-c1d6a77fce7b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 15:20:31 crc kubenswrapper[4890]: E1125 15:20:31.495716 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-wlvxc" podUID="449c7dd2-e3c8-48c2-8770-c1d6a77fce7b" Nov 25 15:20:32 crc kubenswrapper[4890]: I1125 15:20:32.120058 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-2dtqm"] Nov 25 15:20:32 crc kubenswrapper[4890]: I1125 15:20:32.126772 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7d75bbdb6d-5dvcn"] Nov 25 15:20:32 crc kubenswrapper[4890]: W1125 15:20:32.176228 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2bf94c7c_4031_4868_bb7a_39c84abc96ea.slice/crio-3cdfb10313f59ba0200e0a9911b7eed3a4a6e967bd8a27b2c5a7a1a6febbca43 WatchSource:0}: Error finding container 3cdfb10313f59ba0200e0a9911b7eed3a4a6e967bd8a27b2c5a7a1a6febbca43: Status 404 returned error can't find the container with id 3cdfb10313f59ba0200e0a9911b7eed3a4a6e967bd8a27b2c5a7a1a6febbca43 Nov 25 15:20:32 crc kubenswrapper[4890]: I1125 15:20:32.332616 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2dtqm" event={"ID":"c7354c28-0b1f-4dda-9b1b-42dcd2657a7a","Type":"ContainerStarted","Data":"abfb197a2fff9f7341ba2aabdfede31737a3253efcd021058d81946377933b60"} Nov 25 15:20:32 crc kubenswrapper[4890]: I1125 15:20:32.333992 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-556858788d-s2ppl" event={"ID":"1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f","Type":"ContainerStarted","Data":"fa469e382145d10056c39ef4297c66072335b5c0cd9e1036a451051c60370bcb"} Nov 25 15:20:32 crc kubenswrapper[4890]: I1125 15:20:32.335553 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d75bbdb6d-5dvcn" event={"ID":"2bf94c7c-4031-4868-bb7a-39c84abc96ea","Type":"ContainerStarted","Data":"3cdfb10313f59ba0200e0a9911b7eed3a4a6e967bd8a27b2c5a7a1a6febbca43"} Nov 25 15:20:32 crc kubenswrapper[4890]: E1125 15:20:32.339596 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-wlvxc" podUID="449c7dd2-e3c8-48c2-8770-c1d6a77fce7b" Nov 25 15:20:33 crc kubenswrapper[4890]: I1125 15:20:33.350003 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-j7w9k" event={"ID":"997a078a-52bd-4eb2-8fde-f80d05d9b18c","Type":"ContainerStarted","Data":"ee352510e5cc498c734f0e0cdd27fdfe9c81cc7a997ed487be227c10e1a3569c"} Nov 25 15:20:33 crc kubenswrapper[4890]: I1125 15:20:33.353225 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2dtqm" event={"ID":"c7354c28-0b1f-4dda-9b1b-42dcd2657a7a","Type":"ContainerStarted","Data":"cc5632f90b85f3b1e3a9f74ec6b9c2852942cfe570e5db273119a852e7086cb4"} Nov 25 15:20:33 crc kubenswrapper[4890]: I1125 15:20:33.354933 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-f8ctf" event={"ID":"fdcacd06-b857-46cf-8516-345e271cf882","Type":"ContainerStarted","Data":"040c842a972c23a9692699ce5354d20d5e885f515d047301cf76e6ca7da7c2c7"} Nov 25 15:20:33 crc kubenswrapper[4890]: I1125 15:20:33.357692 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-556858788d-s2ppl" event={"ID":"1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f","Type":"ContainerStarted","Data":"97df7bc09544ad16b4be8d113257f79ac856eb0352b81415283734caeb5030bb"} Nov 25 15:20:33 crc kubenswrapper[4890]: I1125 15:20:33.357755 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-556858788d-s2ppl" event={"ID":"1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f","Type":"ContainerStarted","Data":"d1b1f601af90078f6742d375b3a9c3a0ce2fefdf39c2cde7a1057b3e2a3f8950"} Nov 25 15:20:33 crc kubenswrapper[4890]: I1125 15:20:33.371347 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-j7w9k" podStartSLOduration=3.047208677 podStartE2EDuration="54.371323535s" podCreationTimestamp="2025-11-25 15:19:39 +0000 UTC" firstStartedPulling="2025-11-25 15:19:40.871371552 +0000 UTC m=+1039.313834162" lastFinishedPulling="2025-11-25 15:20:32.19548641 +0000 UTC m=+1090.637949020" observedRunningTime="2025-11-25 15:20:33.370874764 +0000 UTC m=+1091.813337374" watchObservedRunningTime="2025-11-25 15:20:33.371323535 +0000 UTC m=+1091.813786145" Nov 25 15:20:33 crc kubenswrapper[4890]: I1125 15:20:33.398983 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-2dtqm" podStartSLOduration=7.398959251 podStartE2EDuration="7.398959251s" podCreationTimestamp="2025-11-25 15:20:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:20:33.391593366 +0000 UTC m=+1091.834055996" watchObservedRunningTime="2025-11-25 15:20:33.398959251 +0000 UTC m=+1091.841421861" Nov 25 15:20:33 crc kubenswrapper[4890]: I1125 15:20:33.415895 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-556858788d-s2ppl" podStartSLOduration=44.614874967 podStartE2EDuration="45.415872017s" podCreationTimestamp="2025-11-25 15:19:48 +0000 UTC" firstStartedPulling="2025-11-25 15:20:31.467594432 +0000 UTC m=+1089.910057042" lastFinishedPulling="2025-11-25 15:20:32.268591492 +0000 UTC m=+1090.711054092" observedRunningTime="2025-11-25 15:20:33.41279047 +0000 UTC m=+1091.855253100" watchObservedRunningTime="2025-11-25 15:20:33.415872017 +0000 UTC m=+1091.858334627" Nov 25 15:20:33 crc kubenswrapper[4890]: I1125 15:20:33.441625 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-f8ctf" podStartSLOduration=3.847970493 podStartE2EDuration="54.441597805s" podCreationTimestamp="2025-11-25 15:19:39 +0000 UTC" firstStartedPulling="2025-11-25 15:19:41.6055202 +0000 UTC m=+1040.047982810" lastFinishedPulling="2025-11-25 15:20:32.199147512 +0000 UTC m=+1090.641610122" observedRunningTime="2025-11-25 15:20:33.43782884 +0000 UTC m=+1091.880291450" watchObservedRunningTime="2025-11-25 15:20:33.441597805 +0000 UTC m=+1091.884060425" Nov 25 15:20:36 crc kubenswrapper[4890]: I1125 15:20:36.394452 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6517688d-0723-4809-ba51-bcaf48524157","Type":"ContainerStarted","Data":"991693f8c1376bf61beb8011ad6ffc45c95a03b68cf16aecbf073b0e7471c6c8"} Nov 25 15:20:36 crc kubenswrapper[4890]: I1125 15:20:36.395447 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6517688d-0723-4809-ba51-bcaf48524157","Type":"ContainerStarted","Data":"863ed2acb1632df5aa0b48e110f44f3bb6f794da4dac93c7001300c6772b533c"} Nov 25 15:20:36 crc kubenswrapper[4890]: I1125 15:20:36.395464 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6517688d-0723-4809-ba51-bcaf48524157","Type":"ContainerStarted","Data":"f629b3e878730975419fdcce5b764cdc1ac1454fb3a5ad4a41575de5e357a663"} Nov 25 15:20:36 crc kubenswrapper[4890]: I1125 15:20:36.396433 4890 generic.go:334] "Generic (PLEG): container finished" podID="c7354c28-0b1f-4dda-9b1b-42dcd2657a7a" containerID="cc5632f90b85f3b1e3a9f74ec6b9c2852942cfe570e5db273119a852e7086cb4" exitCode=0 Nov 25 15:20:36 crc kubenswrapper[4890]: I1125 15:20:36.396494 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2dtqm" event={"ID":"c7354c28-0b1f-4dda-9b1b-42dcd2657a7a","Type":"ContainerDied","Data":"cc5632f90b85f3b1e3a9f74ec6b9c2852942cfe570e5db273119a852e7086cb4"} Nov 25 15:20:36 crc kubenswrapper[4890]: I1125 15:20:36.402948 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d75bbdb6d-5dvcn" event={"ID":"2bf94c7c-4031-4868-bb7a-39c84abc96ea","Type":"ContainerStarted","Data":"a3584a13e591d3c86ea2f545691d1c9f7b0bc94e1f637d66290c07874f8af1f9"} Nov 25 15:20:36 crc kubenswrapper[4890]: I1125 15:20:36.403611 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d75bbdb6d-5dvcn" event={"ID":"2bf94c7c-4031-4868-bb7a-39c84abc96ea","Type":"ContainerStarted","Data":"ceb406d5306a6aea5eb0e70d339ac8d6712aa7dca63d2b42fa254863d2d0cdbf"} Nov 25 15:20:36 crc kubenswrapper[4890]: I1125 15:20:36.405098 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a886f74-d737-46b2-9647-232deec47e87","Type":"ContainerStarted","Data":"ebee09cd8dc4bb3bdcf5d00c763735682f030f46aedd33153df2043de970ac49"} Nov 25 15:20:37 crc kubenswrapper[4890]: I1125 15:20:37.422525 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6517688d-0723-4809-ba51-bcaf48524157","Type":"ContainerStarted","Data":"01b79716938e1599f6acbf7f736a63ba77ba32e6c85e9d34b9b2ad8f7c864db4"} Nov 25 15:20:37 crc kubenswrapper[4890]: I1125 15:20:37.422571 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6517688d-0723-4809-ba51-bcaf48524157","Type":"ContainerStarted","Data":"bcb127ff1df57fb6601ead8f3209ea14f704532746a8a2f6cf3dc7f34f74c4e9"} Nov 25 15:20:38 crc kubenswrapper[4890]: I1125 15:20:38.460543 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-556858788d-s2ppl" Nov 25 15:20:38 crc kubenswrapper[4890]: I1125 15:20:38.461737 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-556858788d-s2ppl" Nov 25 15:20:38 crc kubenswrapper[4890]: I1125 15:20:38.604409 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7d75bbdb6d-5dvcn" Nov 25 15:20:38 crc kubenswrapper[4890]: I1125 15:20:38.604464 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7d75bbdb6d-5dvcn" Nov 25 15:20:39 crc kubenswrapper[4890]: I1125 15:20:39.690949 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2dtqm" Nov 25 15:20:39 crc kubenswrapper[4890]: I1125 15:20:39.710213 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7d75bbdb6d-5dvcn" podStartSLOduration=50.572166841 podStartE2EDuration="51.710193224s" podCreationTimestamp="2025-11-25 15:19:48 +0000 UTC" firstStartedPulling="2025-11-25 15:20:32.193193652 +0000 UTC m=+1090.635656262" lastFinishedPulling="2025-11-25 15:20:33.331220035 +0000 UTC m=+1091.773682645" observedRunningTime="2025-11-25 15:20:36.485478212 +0000 UTC m=+1094.927940822" watchObservedRunningTime="2025-11-25 15:20:39.710193224 +0000 UTC m=+1098.152655854" Nov 25 15:20:39 crc kubenswrapper[4890]: I1125 15:20:39.833148 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c7354c28-0b1f-4dda-9b1b-42dcd2657a7a-fernet-keys\") pod \"c7354c28-0b1f-4dda-9b1b-42dcd2657a7a\" (UID: \"c7354c28-0b1f-4dda-9b1b-42dcd2657a7a\") " Nov 25 15:20:39 crc kubenswrapper[4890]: I1125 15:20:39.833370 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7354c28-0b1f-4dda-9b1b-42dcd2657a7a-scripts\") pod \"c7354c28-0b1f-4dda-9b1b-42dcd2657a7a\" (UID: \"c7354c28-0b1f-4dda-9b1b-42dcd2657a7a\") " Nov 25 15:20:39 crc kubenswrapper[4890]: I1125 15:20:39.833392 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d9rbh\" (UniqueName: \"kubernetes.io/projected/c7354c28-0b1f-4dda-9b1b-42dcd2657a7a-kube-api-access-d9rbh\") pod \"c7354c28-0b1f-4dda-9b1b-42dcd2657a7a\" (UID: \"c7354c28-0b1f-4dda-9b1b-42dcd2657a7a\") " Nov 25 15:20:39 crc kubenswrapper[4890]: I1125 15:20:39.833449 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7354c28-0b1f-4dda-9b1b-42dcd2657a7a-config-data\") pod \"c7354c28-0b1f-4dda-9b1b-42dcd2657a7a\" (UID: \"c7354c28-0b1f-4dda-9b1b-42dcd2657a7a\") " Nov 25 15:20:39 crc kubenswrapper[4890]: I1125 15:20:39.833518 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c7354c28-0b1f-4dda-9b1b-42dcd2657a7a-credential-keys\") pod \"c7354c28-0b1f-4dda-9b1b-42dcd2657a7a\" (UID: \"c7354c28-0b1f-4dda-9b1b-42dcd2657a7a\") " Nov 25 15:20:39 crc kubenswrapper[4890]: I1125 15:20:39.833569 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7354c28-0b1f-4dda-9b1b-42dcd2657a7a-combined-ca-bundle\") pod \"c7354c28-0b1f-4dda-9b1b-42dcd2657a7a\" (UID: \"c7354c28-0b1f-4dda-9b1b-42dcd2657a7a\") " Nov 25 15:20:39 crc kubenswrapper[4890]: I1125 15:20:39.840513 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7354c28-0b1f-4dda-9b1b-42dcd2657a7a-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "c7354c28-0b1f-4dda-9b1b-42dcd2657a7a" (UID: "c7354c28-0b1f-4dda-9b1b-42dcd2657a7a"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:20:39 crc kubenswrapper[4890]: I1125 15:20:39.840626 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7354c28-0b1f-4dda-9b1b-42dcd2657a7a-kube-api-access-d9rbh" (OuterVolumeSpecName: "kube-api-access-d9rbh") pod "c7354c28-0b1f-4dda-9b1b-42dcd2657a7a" (UID: "c7354c28-0b1f-4dda-9b1b-42dcd2657a7a"). InnerVolumeSpecName "kube-api-access-d9rbh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:20:39 crc kubenswrapper[4890]: I1125 15:20:39.856094 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7354c28-0b1f-4dda-9b1b-42dcd2657a7a-scripts" (OuterVolumeSpecName: "scripts") pod "c7354c28-0b1f-4dda-9b1b-42dcd2657a7a" (UID: "c7354c28-0b1f-4dda-9b1b-42dcd2657a7a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:20:39 crc kubenswrapper[4890]: I1125 15:20:39.856152 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7354c28-0b1f-4dda-9b1b-42dcd2657a7a-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "c7354c28-0b1f-4dda-9b1b-42dcd2657a7a" (UID: "c7354c28-0b1f-4dda-9b1b-42dcd2657a7a"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:20:39 crc kubenswrapper[4890]: I1125 15:20:39.862846 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7354c28-0b1f-4dda-9b1b-42dcd2657a7a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c7354c28-0b1f-4dda-9b1b-42dcd2657a7a" (UID: "c7354c28-0b1f-4dda-9b1b-42dcd2657a7a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:20:39 crc kubenswrapper[4890]: I1125 15:20:39.879085 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7354c28-0b1f-4dda-9b1b-42dcd2657a7a-config-data" (OuterVolumeSpecName: "config-data") pod "c7354c28-0b1f-4dda-9b1b-42dcd2657a7a" (UID: "c7354c28-0b1f-4dda-9b1b-42dcd2657a7a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:20:39 crc kubenswrapper[4890]: I1125 15:20:39.935878 4890 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7354c28-0b1f-4dda-9b1b-42dcd2657a7a-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 15:20:39 crc kubenswrapper[4890]: I1125 15:20:39.935918 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d9rbh\" (UniqueName: \"kubernetes.io/projected/c7354c28-0b1f-4dda-9b1b-42dcd2657a7a-kube-api-access-d9rbh\") on node \"crc\" DevicePath \"\"" Nov 25 15:20:39 crc kubenswrapper[4890]: I1125 15:20:39.935933 4890 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7354c28-0b1f-4dda-9b1b-42dcd2657a7a-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 15:20:39 crc kubenswrapper[4890]: I1125 15:20:39.935945 4890 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c7354c28-0b1f-4dda-9b1b-42dcd2657a7a-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 25 15:20:39 crc kubenswrapper[4890]: I1125 15:20:39.935956 4890 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7354c28-0b1f-4dda-9b1b-42dcd2657a7a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:20:39 crc kubenswrapper[4890]: I1125 15:20:39.935966 4890 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c7354c28-0b1f-4dda-9b1b-42dcd2657a7a-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 25 15:20:40 crc kubenswrapper[4890]: I1125 15:20:40.465574 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6517688d-0723-4809-ba51-bcaf48524157","Type":"ContainerStarted","Data":"7b1bc50fcbda8a18716a1e9e65010693779737d96a317f1d3c2e9c05ff6d80b3"} Nov 25 15:20:40 crc kubenswrapper[4890]: I1125 15:20:40.467589 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2dtqm" event={"ID":"c7354c28-0b1f-4dda-9b1b-42dcd2657a7a","Type":"ContainerDied","Data":"abfb197a2fff9f7341ba2aabdfede31737a3253efcd021058d81946377933b60"} Nov 25 15:20:40 crc kubenswrapper[4890]: I1125 15:20:40.467636 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="abfb197a2fff9f7341ba2aabdfede31737a3253efcd021058d81946377933b60" Nov 25 15:20:40 crc kubenswrapper[4890]: I1125 15:20:40.467700 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2dtqm" Nov 25 15:20:40 crc kubenswrapper[4890]: I1125 15:20:40.889891 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-64fdb55f6-22c8w"] Nov 25 15:20:40 crc kubenswrapper[4890]: E1125 15:20:40.891204 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7354c28-0b1f-4dda-9b1b-42dcd2657a7a" containerName="keystone-bootstrap" Nov 25 15:20:40 crc kubenswrapper[4890]: I1125 15:20:40.891229 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7354c28-0b1f-4dda-9b1b-42dcd2657a7a" containerName="keystone-bootstrap" Nov 25 15:20:40 crc kubenswrapper[4890]: I1125 15:20:40.891501 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7354c28-0b1f-4dda-9b1b-42dcd2657a7a" containerName="keystone-bootstrap" Nov 25 15:20:40 crc kubenswrapper[4890]: I1125 15:20:40.892467 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-64fdb55f6-22c8w" Nov 25 15:20:40 crc kubenswrapper[4890]: I1125 15:20:40.898042 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-t8zql" Nov 25 15:20:40 crc kubenswrapper[4890]: I1125 15:20:40.898091 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 25 15:20:40 crc kubenswrapper[4890]: I1125 15:20:40.898059 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 25 15:20:40 crc kubenswrapper[4890]: I1125 15:20:40.898462 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 25 15:20:40 crc kubenswrapper[4890]: I1125 15:20:40.898596 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Nov 25 15:20:40 crc kubenswrapper[4890]: I1125 15:20:40.904462 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-64fdb55f6-22c8w"] Nov 25 15:20:40 crc kubenswrapper[4890]: I1125 15:20:40.905574 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Nov 25 15:20:40 crc kubenswrapper[4890]: I1125 15:20:40.954479 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e127320b-04a5-48f0-b434-ba6bfe1c2696-internal-tls-certs\") pod \"keystone-64fdb55f6-22c8w\" (UID: \"e127320b-04a5-48f0-b434-ba6bfe1c2696\") " pod="openstack/keystone-64fdb55f6-22c8w" Nov 25 15:20:40 crc kubenswrapper[4890]: I1125 15:20:40.955290 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e127320b-04a5-48f0-b434-ba6bfe1c2696-config-data\") pod \"keystone-64fdb55f6-22c8w\" (UID: \"e127320b-04a5-48f0-b434-ba6bfe1c2696\") " pod="openstack/keystone-64fdb55f6-22c8w" Nov 25 15:20:40 crc kubenswrapper[4890]: I1125 15:20:40.955383 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e127320b-04a5-48f0-b434-ba6bfe1c2696-scripts\") pod \"keystone-64fdb55f6-22c8w\" (UID: \"e127320b-04a5-48f0-b434-ba6bfe1c2696\") " pod="openstack/keystone-64fdb55f6-22c8w" Nov 25 15:20:40 crc kubenswrapper[4890]: I1125 15:20:40.955440 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e127320b-04a5-48f0-b434-ba6bfe1c2696-fernet-keys\") pod \"keystone-64fdb55f6-22c8w\" (UID: \"e127320b-04a5-48f0-b434-ba6bfe1c2696\") " pod="openstack/keystone-64fdb55f6-22c8w" Nov 25 15:20:40 crc kubenswrapper[4890]: I1125 15:20:40.955526 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e127320b-04a5-48f0-b434-ba6bfe1c2696-combined-ca-bundle\") pod \"keystone-64fdb55f6-22c8w\" (UID: \"e127320b-04a5-48f0-b434-ba6bfe1c2696\") " pod="openstack/keystone-64fdb55f6-22c8w" Nov 25 15:20:40 crc kubenswrapper[4890]: I1125 15:20:40.955564 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p597q\" (UniqueName: \"kubernetes.io/projected/e127320b-04a5-48f0-b434-ba6bfe1c2696-kube-api-access-p597q\") pod \"keystone-64fdb55f6-22c8w\" (UID: \"e127320b-04a5-48f0-b434-ba6bfe1c2696\") " pod="openstack/keystone-64fdb55f6-22c8w" Nov 25 15:20:40 crc kubenswrapper[4890]: I1125 15:20:40.955700 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e127320b-04a5-48f0-b434-ba6bfe1c2696-public-tls-certs\") pod \"keystone-64fdb55f6-22c8w\" (UID: \"e127320b-04a5-48f0-b434-ba6bfe1c2696\") " pod="openstack/keystone-64fdb55f6-22c8w" Nov 25 15:20:40 crc kubenswrapper[4890]: I1125 15:20:40.955756 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e127320b-04a5-48f0-b434-ba6bfe1c2696-credential-keys\") pod \"keystone-64fdb55f6-22c8w\" (UID: \"e127320b-04a5-48f0-b434-ba6bfe1c2696\") " pod="openstack/keystone-64fdb55f6-22c8w" Nov 25 15:20:41 crc kubenswrapper[4890]: I1125 15:20:41.056297 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e127320b-04a5-48f0-b434-ba6bfe1c2696-public-tls-certs\") pod \"keystone-64fdb55f6-22c8w\" (UID: \"e127320b-04a5-48f0-b434-ba6bfe1c2696\") " pod="openstack/keystone-64fdb55f6-22c8w" Nov 25 15:20:41 crc kubenswrapper[4890]: I1125 15:20:41.056350 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e127320b-04a5-48f0-b434-ba6bfe1c2696-credential-keys\") pod \"keystone-64fdb55f6-22c8w\" (UID: \"e127320b-04a5-48f0-b434-ba6bfe1c2696\") " pod="openstack/keystone-64fdb55f6-22c8w" Nov 25 15:20:41 crc kubenswrapper[4890]: I1125 15:20:41.056385 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e127320b-04a5-48f0-b434-ba6bfe1c2696-internal-tls-certs\") pod \"keystone-64fdb55f6-22c8w\" (UID: \"e127320b-04a5-48f0-b434-ba6bfe1c2696\") " pod="openstack/keystone-64fdb55f6-22c8w" Nov 25 15:20:41 crc kubenswrapper[4890]: I1125 15:20:41.056456 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e127320b-04a5-48f0-b434-ba6bfe1c2696-config-data\") pod \"keystone-64fdb55f6-22c8w\" (UID: \"e127320b-04a5-48f0-b434-ba6bfe1c2696\") " pod="openstack/keystone-64fdb55f6-22c8w" Nov 25 15:20:41 crc kubenswrapper[4890]: I1125 15:20:41.056492 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e127320b-04a5-48f0-b434-ba6bfe1c2696-scripts\") pod \"keystone-64fdb55f6-22c8w\" (UID: \"e127320b-04a5-48f0-b434-ba6bfe1c2696\") " pod="openstack/keystone-64fdb55f6-22c8w" Nov 25 15:20:41 crc kubenswrapper[4890]: I1125 15:20:41.056521 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e127320b-04a5-48f0-b434-ba6bfe1c2696-fernet-keys\") pod \"keystone-64fdb55f6-22c8w\" (UID: \"e127320b-04a5-48f0-b434-ba6bfe1c2696\") " pod="openstack/keystone-64fdb55f6-22c8w" Nov 25 15:20:41 crc kubenswrapper[4890]: I1125 15:20:41.056556 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e127320b-04a5-48f0-b434-ba6bfe1c2696-combined-ca-bundle\") pod \"keystone-64fdb55f6-22c8w\" (UID: \"e127320b-04a5-48f0-b434-ba6bfe1c2696\") " pod="openstack/keystone-64fdb55f6-22c8w" Nov 25 15:20:41 crc kubenswrapper[4890]: I1125 15:20:41.056576 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p597q\" (UniqueName: \"kubernetes.io/projected/e127320b-04a5-48f0-b434-ba6bfe1c2696-kube-api-access-p597q\") pod \"keystone-64fdb55f6-22c8w\" (UID: \"e127320b-04a5-48f0-b434-ba6bfe1c2696\") " pod="openstack/keystone-64fdb55f6-22c8w" Nov 25 15:20:41 crc kubenswrapper[4890]: I1125 15:20:41.061530 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e127320b-04a5-48f0-b434-ba6bfe1c2696-credential-keys\") pod \"keystone-64fdb55f6-22c8w\" (UID: \"e127320b-04a5-48f0-b434-ba6bfe1c2696\") " pod="openstack/keystone-64fdb55f6-22c8w" Nov 25 15:20:41 crc kubenswrapper[4890]: I1125 15:20:41.061576 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e127320b-04a5-48f0-b434-ba6bfe1c2696-public-tls-certs\") pod \"keystone-64fdb55f6-22c8w\" (UID: \"e127320b-04a5-48f0-b434-ba6bfe1c2696\") " pod="openstack/keystone-64fdb55f6-22c8w" Nov 25 15:20:41 crc kubenswrapper[4890]: I1125 15:20:41.061952 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e127320b-04a5-48f0-b434-ba6bfe1c2696-internal-tls-certs\") pod \"keystone-64fdb55f6-22c8w\" (UID: \"e127320b-04a5-48f0-b434-ba6bfe1c2696\") " pod="openstack/keystone-64fdb55f6-22c8w" Nov 25 15:20:41 crc kubenswrapper[4890]: I1125 15:20:41.062402 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e127320b-04a5-48f0-b434-ba6bfe1c2696-scripts\") pod \"keystone-64fdb55f6-22c8w\" (UID: \"e127320b-04a5-48f0-b434-ba6bfe1c2696\") " pod="openstack/keystone-64fdb55f6-22c8w" Nov 25 15:20:41 crc kubenswrapper[4890]: I1125 15:20:41.063263 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e127320b-04a5-48f0-b434-ba6bfe1c2696-fernet-keys\") pod \"keystone-64fdb55f6-22c8w\" (UID: \"e127320b-04a5-48f0-b434-ba6bfe1c2696\") " pod="openstack/keystone-64fdb55f6-22c8w" Nov 25 15:20:41 crc kubenswrapper[4890]: I1125 15:20:41.071106 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e127320b-04a5-48f0-b434-ba6bfe1c2696-combined-ca-bundle\") pod \"keystone-64fdb55f6-22c8w\" (UID: \"e127320b-04a5-48f0-b434-ba6bfe1c2696\") " pod="openstack/keystone-64fdb55f6-22c8w" Nov 25 15:20:41 crc kubenswrapper[4890]: I1125 15:20:41.074031 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e127320b-04a5-48f0-b434-ba6bfe1c2696-config-data\") pod \"keystone-64fdb55f6-22c8w\" (UID: \"e127320b-04a5-48f0-b434-ba6bfe1c2696\") " pod="openstack/keystone-64fdb55f6-22c8w" Nov 25 15:20:41 crc kubenswrapper[4890]: I1125 15:20:41.075960 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p597q\" (UniqueName: \"kubernetes.io/projected/e127320b-04a5-48f0-b434-ba6bfe1c2696-kube-api-access-p597q\") pod \"keystone-64fdb55f6-22c8w\" (UID: \"e127320b-04a5-48f0-b434-ba6bfe1c2696\") " pod="openstack/keystone-64fdb55f6-22c8w" Nov 25 15:20:41 crc kubenswrapper[4890]: I1125 15:20:41.225825 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-64fdb55f6-22c8w" Nov 25 15:20:41 crc kubenswrapper[4890]: I1125 15:20:41.671227 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-64fdb55f6-22c8w"] Nov 25 15:20:41 crc kubenswrapper[4890]: W1125 15:20:41.674217 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode127320b_04a5_48f0_b434_ba6bfe1c2696.slice/crio-26a2839d4b874d6cc1a4dbf607c1134b305251929661c31892d6c4431f1ddcd5 WatchSource:0}: Error finding container 26a2839d4b874d6cc1a4dbf607c1134b305251929661c31892d6c4431f1ddcd5: Status 404 returned error can't find the container with id 26a2839d4b874d6cc1a4dbf607c1134b305251929661c31892d6c4431f1ddcd5 Nov 25 15:20:42 crc kubenswrapper[4890]: I1125 15:20:42.489464 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6517688d-0723-4809-ba51-bcaf48524157","Type":"ContainerStarted","Data":"9c01200278faee072bb4bc5a358b5cbfa34569fb1840c75c26451074e44fefd5"} Nov 25 15:20:42 crc kubenswrapper[4890]: I1125 15:20:42.490752 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-64fdb55f6-22c8w" event={"ID":"e127320b-04a5-48f0-b434-ba6bfe1c2696","Type":"ContainerStarted","Data":"26a2839d4b874d6cc1a4dbf607c1134b305251929661c31892d6c4431f1ddcd5"} Nov 25 15:20:43 crc kubenswrapper[4890]: I1125 15:20:43.500030 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-64fdb55f6-22c8w" event={"ID":"e127320b-04a5-48f0-b434-ba6bfe1c2696","Type":"ContainerStarted","Data":"6c4df70a9b6f7340e2e2b9138b7fff55e2b8ac2e362bd5e5b67c19d6c1fa8a8c"} Nov 25 15:20:43 crc kubenswrapper[4890]: E1125 15:20:43.519966 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"container-server\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"container-replicator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-swift-container:current-podified\\\"\", failed to \"StartContainer\" for \"container-auditor\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-swift-container:current-podified\\\"\", failed to \"StartContainer\" for \"container-updater\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-swift-container:current-podified\\\"\"]" pod="openstack/swift-storage-0" podUID="6517688d-0723-4809-ba51-bcaf48524157" Nov 25 15:20:44 crc kubenswrapper[4890]: I1125 15:20:44.515315 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-64fdb55f6-22c8w" Nov 25 15:20:44 crc kubenswrapper[4890]: I1125 15:20:44.578010 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-64fdb55f6-22c8w" podStartSLOduration=4.577991172 podStartE2EDuration="4.577991172s" podCreationTimestamp="2025-11-25 15:20:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:20:44.57549784 +0000 UTC m=+1103.017960480" watchObservedRunningTime="2025-11-25 15:20:44.577991172 +0000 UTC m=+1103.020453792" Nov 25 15:20:48 crc kubenswrapper[4890]: I1125 15:20:48.462799 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-556858788d-s2ppl" podUID="1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.145:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.145:8443: connect: connection refused" Nov 25 15:20:48 crc kubenswrapper[4890]: I1125 15:20:48.605817 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7d75bbdb6d-5dvcn" podUID="2bf94c7c-4031-4868-bb7a-39c84abc96ea" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Nov 25 15:20:49 crc kubenswrapper[4890]: I1125 15:20:49.575207 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a886f74-d737-46b2-9647-232deec47e87","Type":"ContainerStarted","Data":"86359cd1725531e6f39368d66c19f7fbf05633d7f16fcb20f6cab85de7b914cc"} Nov 25 15:20:49 crc kubenswrapper[4890]: I1125 15:20:49.588716 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6517688d-0723-4809-ba51-bcaf48524157","Type":"ContainerStarted","Data":"ca715c4d03a178be76385b05a8f8e82c01f0151f2bfb1e1338bbb5a8a40bec51"} Nov 25 15:20:50 crc kubenswrapper[4890]: I1125 15:20:50.623767 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6517688d-0723-4809-ba51-bcaf48524157","Type":"ContainerStarted","Data":"c549f530d51242a8288dd637537dd647ad71efec275a6043034e937f7df945e4"} Nov 25 15:20:50 crc kubenswrapper[4890]: I1125 15:20:50.624281 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6517688d-0723-4809-ba51-bcaf48524157","Type":"ContainerStarted","Data":"5da5609243a482155c9b485c8e84b5d374940bb1d7bee1c4fe1002d900f81391"} Nov 25 15:20:50 crc kubenswrapper[4890]: I1125 15:20:50.637974 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-wlvxc" event={"ID":"449c7dd2-e3c8-48c2-8770-c1d6a77fce7b","Type":"ContainerStarted","Data":"92f0c8a918d324a2aabee016770d39884c7dbbe53d97c21e28654daf6eef0a8a"} Nov 25 15:20:50 crc kubenswrapper[4890]: I1125 15:20:50.678980 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-wlvxc" podStartSLOduration=3.982276378 podStartE2EDuration="1m12.678962518s" podCreationTimestamp="2025-11-25 15:19:38 +0000 UTC" firstStartedPulling="2025-11-25 15:19:40.53837074 +0000 UTC m=+1038.980833350" lastFinishedPulling="2025-11-25 15:20:49.23505688 +0000 UTC m=+1107.677519490" observedRunningTime="2025-11-25 15:20:50.677477241 +0000 UTC m=+1109.119939861" watchObservedRunningTime="2025-11-25 15:20:50.678962518 +0000 UTC m=+1109.121425128" Nov 25 15:20:51 crc kubenswrapper[4890]: I1125 15:20:51.653064 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6517688d-0723-4809-ba51-bcaf48524157","Type":"ContainerStarted","Data":"e5aa9528268ec02cbd87a9591796c7b5c6ea0ded70c3c44d54e6fb23e79bbc04"} Nov 25 15:20:51 crc kubenswrapper[4890]: I1125 15:20:51.703548 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=39.67021387 podStartE2EDuration="1m50.703529381s" podCreationTimestamp="2025-11-25 15:19:01 +0000 UTC" firstStartedPulling="2025-11-25 15:19:38.207140385 +0000 UTC m=+1036.649602995" lastFinishedPulling="2025-11-25 15:20:49.240455896 +0000 UTC m=+1107.682918506" observedRunningTime="2025-11-25 15:20:51.688674037 +0000 UTC m=+1110.131136667" watchObservedRunningTime="2025-11-25 15:20:51.703529381 +0000 UTC m=+1110.145991991" Nov 25 15:20:52 crc kubenswrapper[4890]: I1125 15:20:52.084840 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-2hjbr"] Nov 25 15:20:52 crc kubenswrapper[4890]: I1125 15:20:52.087633 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dd9ff6bc-2hjbr" Nov 25 15:20:52 crc kubenswrapper[4890]: I1125 15:20:52.101269 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Nov 25 15:20:52 crc kubenswrapper[4890]: I1125 15:20:52.114750 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-2hjbr"] Nov 25 15:20:52 crc kubenswrapper[4890]: I1125 15:20:52.270024 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e5750e4c-060e-4dfa-9592-9295e1e44359-ovsdbserver-sb\") pod \"dnsmasq-dns-58dd9ff6bc-2hjbr\" (UID: \"e5750e4c-060e-4dfa-9592-9295e1e44359\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-2hjbr" Nov 25 15:20:52 crc kubenswrapper[4890]: I1125 15:20:52.270141 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e5750e4c-060e-4dfa-9592-9295e1e44359-dns-svc\") pod \"dnsmasq-dns-58dd9ff6bc-2hjbr\" (UID: \"e5750e4c-060e-4dfa-9592-9295e1e44359\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-2hjbr" Nov 25 15:20:52 crc kubenswrapper[4890]: I1125 15:20:52.270205 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e5750e4c-060e-4dfa-9592-9295e1e44359-ovsdbserver-nb\") pod \"dnsmasq-dns-58dd9ff6bc-2hjbr\" (UID: \"e5750e4c-060e-4dfa-9592-9295e1e44359\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-2hjbr" Nov 25 15:20:52 crc kubenswrapper[4890]: I1125 15:20:52.270698 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mctq\" (UniqueName: \"kubernetes.io/projected/e5750e4c-060e-4dfa-9592-9295e1e44359-kube-api-access-6mctq\") pod \"dnsmasq-dns-58dd9ff6bc-2hjbr\" (UID: \"e5750e4c-060e-4dfa-9592-9295e1e44359\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-2hjbr" Nov 25 15:20:52 crc kubenswrapper[4890]: I1125 15:20:52.270746 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e5750e4c-060e-4dfa-9592-9295e1e44359-dns-swift-storage-0\") pod \"dnsmasq-dns-58dd9ff6bc-2hjbr\" (UID: \"e5750e4c-060e-4dfa-9592-9295e1e44359\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-2hjbr" Nov 25 15:20:52 crc kubenswrapper[4890]: I1125 15:20:52.270910 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5750e4c-060e-4dfa-9592-9295e1e44359-config\") pod \"dnsmasq-dns-58dd9ff6bc-2hjbr\" (UID: \"e5750e4c-060e-4dfa-9592-9295e1e44359\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-2hjbr" Nov 25 15:20:52 crc kubenswrapper[4890]: I1125 15:20:52.371807 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e5750e4c-060e-4dfa-9592-9295e1e44359-ovsdbserver-nb\") pod \"dnsmasq-dns-58dd9ff6bc-2hjbr\" (UID: \"e5750e4c-060e-4dfa-9592-9295e1e44359\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-2hjbr" Nov 25 15:20:52 crc kubenswrapper[4890]: I1125 15:20:52.371894 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mctq\" (UniqueName: \"kubernetes.io/projected/e5750e4c-060e-4dfa-9592-9295e1e44359-kube-api-access-6mctq\") pod \"dnsmasq-dns-58dd9ff6bc-2hjbr\" (UID: \"e5750e4c-060e-4dfa-9592-9295e1e44359\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-2hjbr" Nov 25 15:20:52 crc kubenswrapper[4890]: I1125 15:20:52.371912 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e5750e4c-060e-4dfa-9592-9295e1e44359-dns-swift-storage-0\") pod \"dnsmasq-dns-58dd9ff6bc-2hjbr\" (UID: \"e5750e4c-060e-4dfa-9592-9295e1e44359\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-2hjbr" Nov 25 15:20:52 crc kubenswrapper[4890]: I1125 15:20:52.371952 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5750e4c-060e-4dfa-9592-9295e1e44359-config\") pod \"dnsmasq-dns-58dd9ff6bc-2hjbr\" (UID: \"e5750e4c-060e-4dfa-9592-9295e1e44359\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-2hjbr" Nov 25 15:20:52 crc kubenswrapper[4890]: I1125 15:20:52.372002 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e5750e4c-060e-4dfa-9592-9295e1e44359-ovsdbserver-sb\") pod \"dnsmasq-dns-58dd9ff6bc-2hjbr\" (UID: \"e5750e4c-060e-4dfa-9592-9295e1e44359\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-2hjbr" Nov 25 15:20:52 crc kubenswrapper[4890]: I1125 15:20:52.372028 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e5750e4c-060e-4dfa-9592-9295e1e44359-dns-svc\") pod \"dnsmasq-dns-58dd9ff6bc-2hjbr\" (UID: \"e5750e4c-060e-4dfa-9592-9295e1e44359\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-2hjbr" Nov 25 15:20:52 crc kubenswrapper[4890]: I1125 15:20:52.373280 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e5750e4c-060e-4dfa-9592-9295e1e44359-dns-swift-storage-0\") pod \"dnsmasq-dns-58dd9ff6bc-2hjbr\" (UID: \"e5750e4c-060e-4dfa-9592-9295e1e44359\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-2hjbr" Nov 25 15:20:52 crc kubenswrapper[4890]: I1125 15:20:52.373306 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e5750e4c-060e-4dfa-9592-9295e1e44359-dns-svc\") pod \"dnsmasq-dns-58dd9ff6bc-2hjbr\" (UID: \"e5750e4c-060e-4dfa-9592-9295e1e44359\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-2hjbr" Nov 25 15:20:52 crc kubenswrapper[4890]: I1125 15:20:52.373574 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e5750e4c-060e-4dfa-9592-9295e1e44359-ovsdbserver-sb\") pod \"dnsmasq-dns-58dd9ff6bc-2hjbr\" (UID: \"e5750e4c-060e-4dfa-9592-9295e1e44359\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-2hjbr" Nov 25 15:20:52 crc kubenswrapper[4890]: I1125 15:20:52.374421 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e5750e4c-060e-4dfa-9592-9295e1e44359-ovsdbserver-nb\") pod \"dnsmasq-dns-58dd9ff6bc-2hjbr\" (UID: \"e5750e4c-060e-4dfa-9592-9295e1e44359\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-2hjbr" Nov 25 15:20:52 crc kubenswrapper[4890]: I1125 15:20:52.374571 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5750e4c-060e-4dfa-9592-9295e1e44359-config\") pod \"dnsmasq-dns-58dd9ff6bc-2hjbr\" (UID: \"e5750e4c-060e-4dfa-9592-9295e1e44359\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-2hjbr" Nov 25 15:20:52 crc kubenswrapper[4890]: I1125 15:20:52.407103 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mctq\" (UniqueName: \"kubernetes.io/projected/e5750e4c-060e-4dfa-9592-9295e1e44359-kube-api-access-6mctq\") pod \"dnsmasq-dns-58dd9ff6bc-2hjbr\" (UID: \"e5750e4c-060e-4dfa-9592-9295e1e44359\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-2hjbr" Nov 25 15:20:52 crc kubenswrapper[4890]: I1125 15:20:52.408459 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dd9ff6bc-2hjbr" Nov 25 15:20:52 crc kubenswrapper[4890]: I1125 15:20:52.904782 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-2hjbr"] Nov 25 15:20:52 crc kubenswrapper[4890]: W1125 15:20:52.921379 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5750e4c_060e_4dfa_9592_9295e1e44359.slice/crio-630c580be58b970646e55355182892687cd7347528ab0ef1a920711906b3cdce WatchSource:0}: Error finding container 630c580be58b970646e55355182892687cd7347528ab0ef1a920711906b3cdce: Status 404 returned error can't find the container with id 630c580be58b970646e55355182892687cd7347528ab0ef1a920711906b3cdce Nov 25 15:20:53 crc kubenswrapper[4890]: I1125 15:20:53.693811 4890 generic.go:334] "Generic (PLEG): container finished" podID="e5750e4c-060e-4dfa-9592-9295e1e44359" containerID="7d04dec3867ee45102cae2108850a714091b4e31593cb0da520c5d67c49c75d0" exitCode=0 Nov 25 15:20:53 crc kubenswrapper[4890]: I1125 15:20:53.693890 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-2hjbr" event={"ID":"e5750e4c-060e-4dfa-9592-9295e1e44359","Type":"ContainerDied","Data":"7d04dec3867ee45102cae2108850a714091b4e31593cb0da520c5d67c49c75d0"} Nov 25 15:20:53 crc kubenswrapper[4890]: I1125 15:20:53.694221 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-2hjbr" event={"ID":"e5750e4c-060e-4dfa-9592-9295e1e44359","Type":"ContainerStarted","Data":"630c580be58b970646e55355182892687cd7347528ab0ef1a920711906b3cdce"} Nov 25 15:20:54 crc kubenswrapper[4890]: I1125 15:20:54.703583 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-2hjbr" event={"ID":"e5750e4c-060e-4dfa-9592-9295e1e44359","Type":"ContainerStarted","Data":"098c37f5179061c75b6eebf65637f3531baf46cd67545d729b0204eef4197600"} Nov 25 15:20:54 crc kubenswrapper[4890]: I1125 15:20:54.703929 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-58dd9ff6bc-2hjbr" Nov 25 15:20:54 crc kubenswrapper[4890]: I1125 15:20:54.726585 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-58dd9ff6bc-2hjbr" podStartSLOduration=2.726567522 podStartE2EDuration="2.726567522s" podCreationTimestamp="2025-11-25 15:20:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:20:54.722466659 +0000 UTC m=+1113.164929289" watchObservedRunningTime="2025-11-25 15:20:54.726567522 +0000 UTC m=+1113.169030132" Nov 25 15:20:56 crc kubenswrapper[4890]: I1125 15:20:56.447120 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 15:20:56 crc kubenswrapper[4890]: I1125 15:20:56.447458 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 15:20:58 crc kubenswrapper[4890]: I1125 15:20:58.604309 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7d75bbdb6d-5dvcn" podUID="2bf94c7c-4031-4868-bb7a-39c84abc96ea" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Nov 25 15:20:58 crc kubenswrapper[4890]: E1125 15:20:58.997693 4890 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/system.slice/rpm-ostreed.service\": RecentStats: unable to find data in memory cache]" Nov 25 15:21:00 crc kubenswrapper[4890]: I1125 15:21:00.636875 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-556858788d-s2ppl" Nov 25 15:21:02 crc kubenswrapper[4890]: E1125 15:21:02.126260 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="1a886f74-d737-46b2-9647-232deec47e87" Nov 25 15:21:02 crc kubenswrapper[4890]: I1125 15:21:02.402793 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-556858788d-s2ppl" Nov 25 15:21:02 crc kubenswrapper[4890]: I1125 15:21:02.410637 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-58dd9ff6bc-2hjbr" Nov 25 15:21:02 crc kubenswrapper[4890]: I1125 15:21:02.511466 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68dcc9cf6f-lm2sb"] Nov 25 15:21:02 crc kubenswrapper[4890]: I1125 15:21:02.511695 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-68dcc9cf6f-lm2sb" podUID="d760f8c7-b47e-455f-81c3-0ce064da5c96" containerName="dnsmasq-dns" containerID="cri-o://c2caff9e64a05d375605099a37e8642c11e8fcf517f2776b1cc179c5b8d8d3e5" gracePeriod=10 Nov 25 15:21:02 crc kubenswrapper[4890]: I1125 15:21:02.790719 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a886f74-d737-46b2-9647-232deec47e87","Type":"ContainerStarted","Data":"1e4367f1ecb3a5409dfa48315f7c61d1d4b3843638744d991f42f0429f63358f"} Nov 25 15:21:02 crc kubenswrapper[4890]: I1125 15:21:02.791083 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1a886f74-d737-46b2-9647-232deec47e87" containerName="ceilometer-notification-agent" containerID="cri-o://ebee09cd8dc4bb3bdcf5d00c763735682f030f46aedd33153df2043de970ac49" gracePeriod=30 Nov 25 15:21:02 crc kubenswrapper[4890]: I1125 15:21:02.791120 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 25 15:21:02 crc kubenswrapper[4890]: I1125 15:21:02.791495 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1a886f74-d737-46b2-9647-232deec47e87" containerName="proxy-httpd" containerID="cri-o://1e4367f1ecb3a5409dfa48315f7c61d1d4b3843638744d991f42f0429f63358f" gracePeriod=30 Nov 25 15:21:02 crc kubenswrapper[4890]: I1125 15:21:02.795251 4890 generic.go:334] "Generic (PLEG): container finished" podID="d760f8c7-b47e-455f-81c3-0ce064da5c96" containerID="c2caff9e64a05d375605099a37e8642c11e8fcf517f2776b1cc179c5b8d8d3e5" exitCode=0 Nov 25 15:21:02 crc kubenswrapper[4890]: I1125 15:21:02.795288 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68dcc9cf6f-lm2sb" event={"ID":"d760f8c7-b47e-455f-81c3-0ce064da5c96","Type":"ContainerDied","Data":"c2caff9e64a05d375605099a37e8642c11e8fcf517f2776b1cc179c5b8d8d3e5"} Nov 25 15:21:02 crc kubenswrapper[4890]: I1125 15:21:02.791546 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1a886f74-d737-46b2-9647-232deec47e87" containerName="sg-core" containerID="cri-o://86359cd1725531e6f39368d66c19f7fbf05633d7f16fcb20f6cab85de7b914cc" gracePeriod=30 Nov 25 15:21:03 crc kubenswrapper[4890]: I1125 15:21:03.082850 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68dcc9cf6f-lm2sb" Nov 25 15:21:03 crc kubenswrapper[4890]: I1125 15:21:03.271497 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4q26\" (UniqueName: \"kubernetes.io/projected/d760f8c7-b47e-455f-81c3-0ce064da5c96-kube-api-access-p4q26\") pod \"d760f8c7-b47e-455f-81c3-0ce064da5c96\" (UID: \"d760f8c7-b47e-455f-81c3-0ce064da5c96\") " Nov 25 15:21:03 crc kubenswrapper[4890]: I1125 15:21:03.271609 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d760f8c7-b47e-455f-81c3-0ce064da5c96-dns-svc\") pod \"d760f8c7-b47e-455f-81c3-0ce064da5c96\" (UID: \"d760f8c7-b47e-455f-81c3-0ce064da5c96\") " Nov 25 15:21:03 crc kubenswrapper[4890]: I1125 15:21:03.271644 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d760f8c7-b47e-455f-81c3-0ce064da5c96-ovsdbserver-sb\") pod \"d760f8c7-b47e-455f-81c3-0ce064da5c96\" (UID: \"d760f8c7-b47e-455f-81c3-0ce064da5c96\") " Nov 25 15:21:03 crc kubenswrapper[4890]: I1125 15:21:03.272251 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d760f8c7-b47e-455f-81c3-0ce064da5c96-config\") pod \"d760f8c7-b47e-455f-81c3-0ce064da5c96\" (UID: \"d760f8c7-b47e-455f-81c3-0ce064da5c96\") " Nov 25 15:21:03 crc kubenswrapper[4890]: I1125 15:21:03.272352 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d760f8c7-b47e-455f-81c3-0ce064da5c96-ovsdbserver-nb\") pod \"d760f8c7-b47e-455f-81c3-0ce064da5c96\" (UID: \"d760f8c7-b47e-455f-81c3-0ce064da5c96\") " Nov 25 15:21:03 crc kubenswrapper[4890]: I1125 15:21:03.277009 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d760f8c7-b47e-455f-81c3-0ce064da5c96-kube-api-access-p4q26" (OuterVolumeSpecName: "kube-api-access-p4q26") pod "d760f8c7-b47e-455f-81c3-0ce064da5c96" (UID: "d760f8c7-b47e-455f-81c3-0ce064da5c96"). InnerVolumeSpecName "kube-api-access-p4q26". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:21:03 crc kubenswrapper[4890]: I1125 15:21:03.323681 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d760f8c7-b47e-455f-81c3-0ce064da5c96-config" (OuterVolumeSpecName: "config") pod "d760f8c7-b47e-455f-81c3-0ce064da5c96" (UID: "d760f8c7-b47e-455f-81c3-0ce064da5c96"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:21:03 crc kubenswrapper[4890]: I1125 15:21:03.326554 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d760f8c7-b47e-455f-81c3-0ce064da5c96-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d760f8c7-b47e-455f-81c3-0ce064da5c96" (UID: "d760f8c7-b47e-455f-81c3-0ce064da5c96"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:21:03 crc kubenswrapper[4890]: I1125 15:21:03.326759 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d760f8c7-b47e-455f-81c3-0ce064da5c96-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d760f8c7-b47e-455f-81c3-0ce064da5c96" (UID: "d760f8c7-b47e-455f-81c3-0ce064da5c96"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:21:03 crc kubenswrapper[4890]: I1125 15:21:03.327405 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d760f8c7-b47e-455f-81c3-0ce064da5c96-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d760f8c7-b47e-455f-81c3-0ce064da5c96" (UID: "d760f8c7-b47e-455f-81c3-0ce064da5c96"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:21:03 crc kubenswrapper[4890]: I1125 15:21:03.374287 4890 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d760f8c7-b47e-455f-81c3-0ce064da5c96-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:03 crc kubenswrapper[4890]: I1125 15:21:03.374321 4890 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d760f8c7-b47e-455f-81c3-0ce064da5c96-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:03 crc kubenswrapper[4890]: I1125 15:21:03.374332 4890 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d760f8c7-b47e-455f-81c3-0ce064da5c96-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:03 crc kubenswrapper[4890]: I1125 15:21:03.374340 4890 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d760f8c7-b47e-455f-81c3-0ce064da5c96-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:03 crc kubenswrapper[4890]: I1125 15:21:03.374350 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4q26\" (UniqueName: \"kubernetes.io/projected/d760f8c7-b47e-455f-81c3-0ce064da5c96-kube-api-access-p4q26\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:03 crc kubenswrapper[4890]: I1125 15:21:03.806077 4890 generic.go:334] "Generic (PLEG): container finished" podID="1a886f74-d737-46b2-9647-232deec47e87" containerID="1e4367f1ecb3a5409dfa48315f7c61d1d4b3843638744d991f42f0429f63358f" exitCode=0 Nov 25 15:21:03 crc kubenswrapper[4890]: I1125 15:21:03.806123 4890 generic.go:334] "Generic (PLEG): container finished" podID="1a886f74-d737-46b2-9647-232deec47e87" containerID="86359cd1725531e6f39368d66c19f7fbf05633d7f16fcb20f6cab85de7b914cc" exitCode=2 Nov 25 15:21:03 crc kubenswrapper[4890]: I1125 15:21:03.806152 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a886f74-d737-46b2-9647-232deec47e87","Type":"ContainerDied","Data":"1e4367f1ecb3a5409dfa48315f7c61d1d4b3843638744d991f42f0429f63358f"} Nov 25 15:21:03 crc kubenswrapper[4890]: I1125 15:21:03.806522 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a886f74-d737-46b2-9647-232deec47e87","Type":"ContainerDied","Data":"86359cd1725531e6f39368d66c19f7fbf05633d7f16fcb20f6cab85de7b914cc"} Nov 25 15:21:03 crc kubenswrapper[4890]: I1125 15:21:03.809639 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68dcc9cf6f-lm2sb" event={"ID":"d760f8c7-b47e-455f-81c3-0ce064da5c96","Type":"ContainerDied","Data":"7d98a127dfc030e0dd214a9f59a1f1f03c0fc1a527250179ecd5b339fb28e6a8"} Nov 25 15:21:03 crc kubenswrapper[4890]: I1125 15:21:03.809691 4890 scope.go:117] "RemoveContainer" containerID="c2caff9e64a05d375605099a37e8642c11e8fcf517f2776b1cc179c5b8d8d3e5" Nov 25 15:21:03 crc kubenswrapper[4890]: I1125 15:21:03.809794 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68dcc9cf6f-lm2sb" Nov 25 15:21:03 crc kubenswrapper[4890]: I1125 15:21:03.828936 4890 scope.go:117] "RemoveContainer" containerID="fecd5109953cc4cec007d516b879961539f951212be4adc3272e23b74098f3c3" Nov 25 15:21:03 crc kubenswrapper[4890]: I1125 15:21:03.842938 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68dcc9cf6f-lm2sb"] Nov 25 15:21:03 crc kubenswrapper[4890]: I1125 15:21:03.850005 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-68dcc9cf6f-lm2sb"] Nov 25 15:21:04 crc kubenswrapper[4890]: I1125 15:21:04.184400 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d760f8c7-b47e-455f-81c3-0ce064da5c96" path="/var/lib/kubelet/pods/d760f8c7-b47e-455f-81c3-0ce064da5c96/volumes" Nov 25 15:21:07 crc kubenswrapper[4890]: I1125 15:21:07.848360 4890 generic.go:334] "Generic (PLEG): container finished" podID="1a886f74-d737-46b2-9647-232deec47e87" containerID="ebee09cd8dc4bb3bdcf5d00c763735682f030f46aedd33153df2043de970ac49" exitCode=0 Nov 25 15:21:07 crc kubenswrapper[4890]: I1125 15:21:07.848433 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a886f74-d737-46b2-9647-232deec47e87","Type":"ContainerDied","Data":"ebee09cd8dc4bb3bdcf5d00c763735682f030f46aedd33153df2043de970ac49"} Nov 25 15:21:07 crc kubenswrapper[4890]: I1125 15:21:07.850957 4890 generic.go:334] "Generic (PLEG): container finished" podID="997a078a-52bd-4eb2-8fde-f80d05d9b18c" containerID="ee352510e5cc498c734f0e0cdd27fdfe9c81cc7a997ed487be227c10e1a3569c" exitCode=0 Nov 25 15:21:07 crc kubenswrapper[4890]: I1125 15:21:07.850981 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-j7w9k" event={"ID":"997a078a-52bd-4eb2-8fde-f80d05d9b18c","Type":"ContainerDied","Data":"ee352510e5cc498c734f0e0cdd27fdfe9c81cc7a997ed487be227c10e1a3569c"} Nov 25 15:21:08 crc kubenswrapper[4890]: I1125 15:21:08.383310 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 15:21:08 crc kubenswrapper[4890]: I1125 15:21:08.483634 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a886f74-d737-46b2-9647-232deec47e87-scripts\") pod \"1a886f74-d737-46b2-9647-232deec47e87\" (UID: \"1a886f74-d737-46b2-9647-232deec47e87\") " Nov 25 15:21:08 crc kubenswrapper[4890]: I1125 15:21:08.483772 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a886f74-d737-46b2-9647-232deec47e87-log-httpd\") pod \"1a886f74-d737-46b2-9647-232deec47e87\" (UID: \"1a886f74-d737-46b2-9647-232deec47e87\") " Nov 25 15:21:08 crc kubenswrapper[4890]: I1125 15:21:08.483810 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1a886f74-d737-46b2-9647-232deec47e87-sg-core-conf-yaml\") pod \"1a886f74-d737-46b2-9647-232deec47e87\" (UID: \"1a886f74-d737-46b2-9647-232deec47e87\") " Nov 25 15:21:08 crc kubenswrapper[4890]: I1125 15:21:08.483872 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rk2gl\" (UniqueName: \"kubernetes.io/projected/1a886f74-d737-46b2-9647-232deec47e87-kube-api-access-rk2gl\") pod \"1a886f74-d737-46b2-9647-232deec47e87\" (UID: \"1a886f74-d737-46b2-9647-232deec47e87\") " Nov 25 15:21:08 crc kubenswrapper[4890]: I1125 15:21:08.483961 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a886f74-d737-46b2-9647-232deec47e87-run-httpd\") pod \"1a886f74-d737-46b2-9647-232deec47e87\" (UID: \"1a886f74-d737-46b2-9647-232deec47e87\") " Nov 25 15:21:08 crc kubenswrapper[4890]: I1125 15:21:08.484065 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a886f74-d737-46b2-9647-232deec47e87-config-data\") pod \"1a886f74-d737-46b2-9647-232deec47e87\" (UID: \"1a886f74-d737-46b2-9647-232deec47e87\") " Nov 25 15:21:08 crc kubenswrapper[4890]: I1125 15:21:08.484149 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a886f74-d737-46b2-9647-232deec47e87-combined-ca-bundle\") pod \"1a886f74-d737-46b2-9647-232deec47e87\" (UID: \"1a886f74-d737-46b2-9647-232deec47e87\") " Nov 25 15:21:08 crc kubenswrapper[4890]: I1125 15:21:08.484438 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a886f74-d737-46b2-9647-232deec47e87-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "1a886f74-d737-46b2-9647-232deec47e87" (UID: "1a886f74-d737-46b2-9647-232deec47e87"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:21:08 crc kubenswrapper[4890]: I1125 15:21:08.484526 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a886f74-d737-46b2-9647-232deec47e87-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "1a886f74-d737-46b2-9647-232deec47e87" (UID: "1a886f74-d737-46b2-9647-232deec47e87"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:21:08 crc kubenswrapper[4890]: I1125 15:21:08.484845 4890 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a886f74-d737-46b2-9647-232deec47e87-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:08 crc kubenswrapper[4890]: I1125 15:21:08.484866 4890 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a886f74-d737-46b2-9647-232deec47e87-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:08 crc kubenswrapper[4890]: I1125 15:21:08.493098 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a886f74-d737-46b2-9647-232deec47e87-scripts" (OuterVolumeSpecName: "scripts") pod "1a886f74-d737-46b2-9647-232deec47e87" (UID: "1a886f74-d737-46b2-9647-232deec47e87"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:21:08 crc kubenswrapper[4890]: I1125 15:21:08.493123 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a886f74-d737-46b2-9647-232deec47e87-kube-api-access-rk2gl" (OuterVolumeSpecName: "kube-api-access-rk2gl") pod "1a886f74-d737-46b2-9647-232deec47e87" (UID: "1a886f74-d737-46b2-9647-232deec47e87"). InnerVolumeSpecName "kube-api-access-rk2gl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:21:08 crc kubenswrapper[4890]: I1125 15:21:08.524318 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a886f74-d737-46b2-9647-232deec47e87-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "1a886f74-d737-46b2-9647-232deec47e87" (UID: "1a886f74-d737-46b2-9647-232deec47e87"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:21:08 crc kubenswrapper[4890]: I1125 15:21:08.538838 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a886f74-d737-46b2-9647-232deec47e87-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1a886f74-d737-46b2-9647-232deec47e87" (UID: "1a886f74-d737-46b2-9647-232deec47e87"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:21:08 crc kubenswrapper[4890]: I1125 15:21:08.561694 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a886f74-d737-46b2-9647-232deec47e87-config-data" (OuterVolumeSpecName: "config-data") pod "1a886f74-d737-46b2-9647-232deec47e87" (UID: "1a886f74-d737-46b2-9647-232deec47e87"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:21:08 crc kubenswrapper[4890]: I1125 15:21:08.586718 4890 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a886f74-d737-46b2-9647-232deec47e87-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:08 crc kubenswrapper[4890]: I1125 15:21:08.586771 4890 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a886f74-d737-46b2-9647-232deec47e87-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:08 crc kubenswrapper[4890]: I1125 15:21:08.586782 4890 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1a886f74-d737-46b2-9647-232deec47e87-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:08 crc kubenswrapper[4890]: I1125 15:21:08.586792 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rk2gl\" (UniqueName: \"kubernetes.io/projected/1a886f74-d737-46b2-9647-232deec47e87-kube-api-access-rk2gl\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:08 crc kubenswrapper[4890]: I1125 15:21:08.586805 4890 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a886f74-d737-46b2-9647-232deec47e87-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:08 crc kubenswrapper[4890]: I1125 15:21:08.869222 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a886f74-d737-46b2-9647-232deec47e87","Type":"ContainerDied","Data":"bda5bba1161b8bf25188069ccbc59156bbc3aaaeae183aafe126e72ca564a143"} Nov 25 15:21:08 crc kubenswrapper[4890]: I1125 15:21:08.869269 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 15:21:08 crc kubenswrapper[4890]: I1125 15:21:08.869351 4890 scope.go:117] "RemoveContainer" containerID="1e4367f1ecb3a5409dfa48315f7c61d1d4b3843638744d991f42f0429f63358f" Nov 25 15:21:08 crc kubenswrapper[4890]: I1125 15:21:08.944039 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:21:08 crc kubenswrapper[4890]: I1125 15:21:08.962100 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:21:08 crc kubenswrapper[4890]: I1125 15:21:08.963796 4890 scope.go:117] "RemoveContainer" containerID="86359cd1725531e6f39368d66c19f7fbf05633d7f16fcb20f6cab85de7b914cc" Nov 25 15:21:08 crc kubenswrapper[4890]: I1125 15:21:08.976618 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:21:08 crc kubenswrapper[4890]: E1125 15:21:08.977064 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a886f74-d737-46b2-9647-232deec47e87" containerName="proxy-httpd" Nov 25 15:21:08 crc kubenswrapper[4890]: I1125 15:21:08.977085 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a886f74-d737-46b2-9647-232deec47e87" containerName="proxy-httpd" Nov 25 15:21:08 crc kubenswrapper[4890]: E1125 15:21:08.977098 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a886f74-d737-46b2-9647-232deec47e87" containerName="sg-core" Nov 25 15:21:08 crc kubenswrapper[4890]: I1125 15:21:08.977107 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a886f74-d737-46b2-9647-232deec47e87" containerName="sg-core" Nov 25 15:21:08 crc kubenswrapper[4890]: E1125 15:21:08.977132 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d760f8c7-b47e-455f-81c3-0ce064da5c96" containerName="dnsmasq-dns" Nov 25 15:21:08 crc kubenswrapper[4890]: I1125 15:21:08.977140 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="d760f8c7-b47e-455f-81c3-0ce064da5c96" containerName="dnsmasq-dns" Nov 25 15:21:08 crc kubenswrapper[4890]: E1125 15:21:08.977174 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a886f74-d737-46b2-9647-232deec47e87" containerName="ceilometer-notification-agent" Nov 25 15:21:08 crc kubenswrapper[4890]: I1125 15:21:08.977183 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a886f74-d737-46b2-9647-232deec47e87" containerName="ceilometer-notification-agent" Nov 25 15:21:08 crc kubenswrapper[4890]: E1125 15:21:08.977199 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d760f8c7-b47e-455f-81c3-0ce064da5c96" containerName="init" Nov 25 15:21:08 crc kubenswrapper[4890]: I1125 15:21:08.977206 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="d760f8c7-b47e-455f-81c3-0ce064da5c96" containerName="init" Nov 25 15:21:08 crc kubenswrapper[4890]: I1125 15:21:08.977430 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a886f74-d737-46b2-9647-232deec47e87" containerName="sg-core" Nov 25 15:21:08 crc kubenswrapper[4890]: I1125 15:21:08.977455 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a886f74-d737-46b2-9647-232deec47e87" containerName="proxy-httpd" Nov 25 15:21:08 crc kubenswrapper[4890]: I1125 15:21:08.977466 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="d760f8c7-b47e-455f-81c3-0ce064da5c96" containerName="dnsmasq-dns" Nov 25 15:21:08 crc kubenswrapper[4890]: I1125 15:21:08.977484 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a886f74-d737-46b2-9647-232deec47e87" containerName="ceilometer-notification-agent" Nov 25 15:21:08 crc kubenswrapper[4890]: I1125 15:21:08.980413 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 15:21:08 crc kubenswrapper[4890]: I1125 15:21:08.986054 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 25 15:21:08 crc kubenswrapper[4890]: I1125 15:21:08.986686 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:21:08 crc kubenswrapper[4890]: I1125 15:21:08.987429 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.007583 4890 scope.go:117] "RemoveContainer" containerID="ebee09cd8dc4bb3bdcf5d00c763735682f030f46aedd33153df2043de970ac49" Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.094967 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/62cd6c55-698d-49c9-aa1b-de9d3df17830-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"62cd6c55-698d-49c9-aa1b-de9d3df17830\") " pod="openstack/ceilometer-0" Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.095419 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62cd6c55-698d-49c9-aa1b-de9d3df17830-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"62cd6c55-698d-49c9-aa1b-de9d3df17830\") " pod="openstack/ceilometer-0" Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.095470 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62cd6c55-698d-49c9-aa1b-de9d3df17830-run-httpd\") pod \"ceilometer-0\" (UID: \"62cd6c55-698d-49c9-aa1b-de9d3df17830\") " pod="openstack/ceilometer-0" Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.095505 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62cd6c55-698d-49c9-aa1b-de9d3df17830-config-data\") pod \"ceilometer-0\" (UID: \"62cd6c55-698d-49c9-aa1b-de9d3df17830\") " pod="openstack/ceilometer-0" Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.095553 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2frz\" (UniqueName: \"kubernetes.io/projected/62cd6c55-698d-49c9-aa1b-de9d3df17830-kube-api-access-x2frz\") pod \"ceilometer-0\" (UID: \"62cd6c55-698d-49c9-aa1b-de9d3df17830\") " pod="openstack/ceilometer-0" Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.095575 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62cd6c55-698d-49c9-aa1b-de9d3df17830-log-httpd\") pod \"ceilometer-0\" (UID: \"62cd6c55-698d-49c9-aa1b-de9d3df17830\") " pod="openstack/ceilometer-0" Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.095626 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62cd6c55-698d-49c9-aa1b-de9d3df17830-scripts\") pod \"ceilometer-0\" (UID: \"62cd6c55-698d-49c9-aa1b-de9d3df17830\") " pod="openstack/ceilometer-0" Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.196755 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62cd6c55-698d-49c9-aa1b-de9d3df17830-run-httpd\") pod \"ceilometer-0\" (UID: \"62cd6c55-698d-49c9-aa1b-de9d3df17830\") " pod="openstack/ceilometer-0" Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.196835 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62cd6c55-698d-49c9-aa1b-de9d3df17830-config-data\") pod \"ceilometer-0\" (UID: \"62cd6c55-698d-49c9-aa1b-de9d3df17830\") " pod="openstack/ceilometer-0" Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.196895 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2frz\" (UniqueName: \"kubernetes.io/projected/62cd6c55-698d-49c9-aa1b-de9d3df17830-kube-api-access-x2frz\") pod \"ceilometer-0\" (UID: \"62cd6c55-698d-49c9-aa1b-de9d3df17830\") " pod="openstack/ceilometer-0" Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.196927 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62cd6c55-698d-49c9-aa1b-de9d3df17830-log-httpd\") pod \"ceilometer-0\" (UID: \"62cd6c55-698d-49c9-aa1b-de9d3df17830\") " pod="openstack/ceilometer-0" Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.196986 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62cd6c55-698d-49c9-aa1b-de9d3df17830-scripts\") pod \"ceilometer-0\" (UID: \"62cd6c55-698d-49c9-aa1b-de9d3df17830\") " pod="openstack/ceilometer-0" Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.197046 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/62cd6c55-698d-49c9-aa1b-de9d3df17830-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"62cd6c55-698d-49c9-aa1b-de9d3df17830\") " pod="openstack/ceilometer-0" Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.197068 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62cd6c55-698d-49c9-aa1b-de9d3df17830-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"62cd6c55-698d-49c9-aa1b-de9d3df17830\") " pod="openstack/ceilometer-0" Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.199088 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62cd6c55-698d-49c9-aa1b-de9d3df17830-run-httpd\") pod \"ceilometer-0\" (UID: \"62cd6c55-698d-49c9-aa1b-de9d3df17830\") " pod="openstack/ceilometer-0" Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.203368 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62cd6c55-698d-49c9-aa1b-de9d3df17830-log-httpd\") pod \"ceilometer-0\" (UID: \"62cd6c55-698d-49c9-aa1b-de9d3df17830\") " pod="openstack/ceilometer-0" Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.203656 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/62cd6c55-698d-49c9-aa1b-de9d3df17830-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"62cd6c55-698d-49c9-aa1b-de9d3df17830\") " pod="openstack/ceilometer-0" Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.208905 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62cd6c55-698d-49c9-aa1b-de9d3df17830-scripts\") pod \"ceilometer-0\" (UID: \"62cd6c55-698d-49c9-aa1b-de9d3df17830\") " pod="openstack/ceilometer-0" Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.209605 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62cd6c55-698d-49c9-aa1b-de9d3df17830-config-data\") pod \"ceilometer-0\" (UID: \"62cd6c55-698d-49c9-aa1b-de9d3df17830\") " pod="openstack/ceilometer-0" Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.214884 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62cd6c55-698d-49c9-aa1b-de9d3df17830-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"62cd6c55-698d-49c9-aa1b-de9d3df17830\") " pod="openstack/ceilometer-0" Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.219937 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2frz\" (UniqueName: \"kubernetes.io/projected/62cd6c55-698d-49c9-aa1b-de9d3df17830-kube-api-access-x2frz\") pod \"ceilometer-0\" (UID: \"62cd6c55-698d-49c9-aa1b-de9d3df17830\") " pod="openstack/ceilometer-0" Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.278382 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-j7w9k" Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.309244 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.399331 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/997a078a-52bd-4eb2-8fde-f80d05d9b18c-logs\") pod \"997a078a-52bd-4eb2-8fde-f80d05d9b18c\" (UID: \"997a078a-52bd-4eb2-8fde-f80d05d9b18c\") " Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.399468 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/997a078a-52bd-4eb2-8fde-f80d05d9b18c-config-data\") pod \"997a078a-52bd-4eb2-8fde-f80d05d9b18c\" (UID: \"997a078a-52bd-4eb2-8fde-f80d05d9b18c\") " Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.399539 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pw2ct\" (UniqueName: \"kubernetes.io/projected/997a078a-52bd-4eb2-8fde-f80d05d9b18c-kube-api-access-pw2ct\") pod \"997a078a-52bd-4eb2-8fde-f80d05d9b18c\" (UID: \"997a078a-52bd-4eb2-8fde-f80d05d9b18c\") " Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.399559 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/997a078a-52bd-4eb2-8fde-f80d05d9b18c-scripts\") pod \"997a078a-52bd-4eb2-8fde-f80d05d9b18c\" (UID: \"997a078a-52bd-4eb2-8fde-f80d05d9b18c\") " Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.399637 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/997a078a-52bd-4eb2-8fde-f80d05d9b18c-combined-ca-bundle\") pod \"997a078a-52bd-4eb2-8fde-f80d05d9b18c\" (UID: \"997a078a-52bd-4eb2-8fde-f80d05d9b18c\") " Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.399994 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/997a078a-52bd-4eb2-8fde-f80d05d9b18c-logs" (OuterVolumeSpecName: "logs") pod "997a078a-52bd-4eb2-8fde-f80d05d9b18c" (UID: "997a078a-52bd-4eb2-8fde-f80d05d9b18c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.404009 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/997a078a-52bd-4eb2-8fde-f80d05d9b18c-kube-api-access-pw2ct" (OuterVolumeSpecName: "kube-api-access-pw2ct") pod "997a078a-52bd-4eb2-8fde-f80d05d9b18c" (UID: "997a078a-52bd-4eb2-8fde-f80d05d9b18c"). InnerVolumeSpecName "kube-api-access-pw2ct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.408972 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/997a078a-52bd-4eb2-8fde-f80d05d9b18c-scripts" (OuterVolumeSpecName: "scripts") pod "997a078a-52bd-4eb2-8fde-f80d05d9b18c" (UID: "997a078a-52bd-4eb2-8fde-f80d05d9b18c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.441180 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/997a078a-52bd-4eb2-8fde-f80d05d9b18c-config-data" (OuterVolumeSpecName: "config-data") pod "997a078a-52bd-4eb2-8fde-f80d05d9b18c" (UID: "997a078a-52bd-4eb2-8fde-f80d05d9b18c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.481928 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/997a078a-52bd-4eb2-8fde-f80d05d9b18c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "997a078a-52bd-4eb2-8fde-f80d05d9b18c" (UID: "997a078a-52bd-4eb2-8fde-f80d05d9b18c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.501097 4890 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/997a078a-52bd-4eb2-8fde-f80d05d9b18c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.501129 4890 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/997a078a-52bd-4eb2-8fde-f80d05d9b18c-logs\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.501138 4890 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/997a078a-52bd-4eb2-8fde-f80d05d9b18c-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.501147 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pw2ct\" (UniqueName: \"kubernetes.io/projected/997a078a-52bd-4eb2-8fde-f80d05d9b18c-kube-api-access-pw2ct\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.501188 4890 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/997a078a-52bd-4eb2-8fde-f80d05d9b18c-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.842627 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:21:09 crc kubenswrapper[4890]: W1125 15:21:09.848408 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod62cd6c55_698d_49c9_aa1b_de9d3df17830.slice/crio-f8922b08dda2a1fb0a8e5d472196fecaf8fb947469f950eadb0811f40df6ae5f WatchSource:0}: Error finding container f8922b08dda2a1fb0a8e5d472196fecaf8fb947469f950eadb0811f40df6ae5f: Status 404 returned error can't find the container with id f8922b08dda2a1fb0a8e5d472196fecaf8fb947469f950eadb0811f40df6ae5f Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.881013 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-j7w9k" event={"ID":"997a078a-52bd-4eb2-8fde-f80d05d9b18c","Type":"ContainerDied","Data":"cf4e901c7a76d094a83862a406a6fb9c94053a7cf4e304013cf95d3d582eae9f"} Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.881564 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cf4e901c7a76d094a83862a406a6fb9c94053a7cf4e304013cf95d3d582eae9f" Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.881251 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-j7w9k" Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.897430 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"62cd6c55-698d-49c9-aa1b-de9d3df17830","Type":"ContainerStarted","Data":"f8922b08dda2a1fb0a8e5d472196fecaf8fb947469f950eadb0811f40df6ae5f"} Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.975773 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-74c777b4cb-m7k57"] Nov 25 15:21:09 crc kubenswrapper[4890]: E1125 15:21:09.977442 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="997a078a-52bd-4eb2-8fde-f80d05d9b18c" containerName="placement-db-sync" Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.977553 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="997a078a-52bd-4eb2-8fde-f80d05d9b18c" containerName="placement-db-sync" Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.977804 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="997a078a-52bd-4eb2-8fde-f80d05d9b18c" containerName="placement-db-sync" Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.979014 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-74c777b4cb-m7k57" Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.981248 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.981507 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.981723 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.982328 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-bgppv" Nov 25 15:21:09 crc kubenswrapper[4890]: I1125 15:21:09.982823 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Nov 25 15:21:10 crc kubenswrapper[4890]: I1125 15:21:10.002825 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-74c777b4cb-m7k57"] Nov 25 15:21:10 crc kubenswrapper[4890]: I1125 15:21:10.114558 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46d00836-60a2-459f-b4c9-ed2d5e3b1fb1-logs\") pod \"placement-74c777b4cb-m7k57\" (UID: \"46d00836-60a2-459f-b4c9-ed2d5e3b1fb1\") " pod="openstack/placement-74c777b4cb-m7k57" Nov 25 15:21:10 crc kubenswrapper[4890]: I1125 15:21:10.114718 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46d00836-60a2-459f-b4c9-ed2d5e3b1fb1-scripts\") pod \"placement-74c777b4cb-m7k57\" (UID: \"46d00836-60a2-459f-b4c9-ed2d5e3b1fb1\") " pod="openstack/placement-74c777b4cb-m7k57" Nov 25 15:21:10 crc kubenswrapper[4890]: I1125 15:21:10.115084 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/46d00836-60a2-459f-b4c9-ed2d5e3b1fb1-internal-tls-certs\") pod \"placement-74c777b4cb-m7k57\" (UID: \"46d00836-60a2-459f-b4c9-ed2d5e3b1fb1\") " pod="openstack/placement-74c777b4cb-m7k57" Nov 25 15:21:10 crc kubenswrapper[4890]: I1125 15:21:10.115212 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kx68p\" (UniqueName: \"kubernetes.io/projected/46d00836-60a2-459f-b4c9-ed2d5e3b1fb1-kube-api-access-kx68p\") pod \"placement-74c777b4cb-m7k57\" (UID: \"46d00836-60a2-459f-b4c9-ed2d5e3b1fb1\") " pod="openstack/placement-74c777b4cb-m7k57" Nov 25 15:21:10 crc kubenswrapper[4890]: I1125 15:21:10.115337 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46d00836-60a2-459f-b4c9-ed2d5e3b1fb1-combined-ca-bundle\") pod \"placement-74c777b4cb-m7k57\" (UID: \"46d00836-60a2-459f-b4c9-ed2d5e3b1fb1\") " pod="openstack/placement-74c777b4cb-m7k57" Nov 25 15:21:10 crc kubenswrapper[4890]: I1125 15:21:10.115550 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46d00836-60a2-459f-b4c9-ed2d5e3b1fb1-config-data\") pod \"placement-74c777b4cb-m7k57\" (UID: \"46d00836-60a2-459f-b4c9-ed2d5e3b1fb1\") " pod="openstack/placement-74c777b4cb-m7k57" Nov 25 15:21:10 crc kubenswrapper[4890]: I1125 15:21:10.115584 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/46d00836-60a2-459f-b4c9-ed2d5e3b1fb1-public-tls-certs\") pod \"placement-74c777b4cb-m7k57\" (UID: \"46d00836-60a2-459f-b4c9-ed2d5e3b1fb1\") " pod="openstack/placement-74c777b4cb-m7k57" Nov 25 15:21:10 crc kubenswrapper[4890]: I1125 15:21:10.187252 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a886f74-d737-46b2-9647-232deec47e87" path="/var/lib/kubelet/pods/1a886f74-d737-46b2-9647-232deec47e87/volumes" Nov 25 15:21:10 crc kubenswrapper[4890]: I1125 15:21:10.217272 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/46d00836-60a2-459f-b4c9-ed2d5e3b1fb1-internal-tls-certs\") pod \"placement-74c777b4cb-m7k57\" (UID: \"46d00836-60a2-459f-b4c9-ed2d5e3b1fb1\") " pod="openstack/placement-74c777b4cb-m7k57" Nov 25 15:21:10 crc kubenswrapper[4890]: I1125 15:21:10.217345 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kx68p\" (UniqueName: \"kubernetes.io/projected/46d00836-60a2-459f-b4c9-ed2d5e3b1fb1-kube-api-access-kx68p\") pod \"placement-74c777b4cb-m7k57\" (UID: \"46d00836-60a2-459f-b4c9-ed2d5e3b1fb1\") " pod="openstack/placement-74c777b4cb-m7k57" Nov 25 15:21:10 crc kubenswrapper[4890]: I1125 15:21:10.217403 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46d00836-60a2-459f-b4c9-ed2d5e3b1fb1-combined-ca-bundle\") pod \"placement-74c777b4cb-m7k57\" (UID: \"46d00836-60a2-459f-b4c9-ed2d5e3b1fb1\") " pod="openstack/placement-74c777b4cb-m7k57" Nov 25 15:21:10 crc kubenswrapper[4890]: I1125 15:21:10.217461 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46d00836-60a2-459f-b4c9-ed2d5e3b1fb1-config-data\") pod \"placement-74c777b4cb-m7k57\" (UID: \"46d00836-60a2-459f-b4c9-ed2d5e3b1fb1\") " pod="openstack/placement-74c777b4cb-m7k57" Nov 25 15:21:10 crc kubenswrapper[4890]: I1125 15:21:10.217486 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/46d00836-60a2-459f-b4c9-ed2d5e3b1fb1-public-tls-certs\") pod \"placement-74c777b4cb-m7k57\" (UID: \"46d00836-60a2-459f-b4c9-ed2d5e3b1fb1\") " pod="openstack/placement-74c777b4cb-m7k57" Nov 25 15:21:10 crc kubenswrapper[4890]: I1125 15:21:10.217512 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46d00836-60a2-459f-b4c9-ed2d5e3b1fb1-logs\") pod \"placement-74c777b4cb-m7k57\" (UID: \"46d00836-60a2-459f-b4c9-ed2d5e3b1fb1\") " pod="openstack/placement-74c777b4cb-m7k57" Nov 25 15:21:10 crc kubenswrapper[4890]: I1125 15:21:10.217566 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46d00836-60a2-459f-b4c9-ed2d5e3b1fb1-scripts\") pod \"placement-74c777b4cb-m7k57\" (UID: \"46d00836-60a2-459f-b4c9-ed2d5e3b1fb1\") " pod="openstack/placement-74c777b4cb-m7k57" Nov 25 15:21:10 crc kubenswrapper[4890]: I1125 15:21:10.218721 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46d00836-60a2-459f-b4c9-ed2d5e3b1fb1-logs\") pod \"placement-74c777b4cb-m7k57\" (UID: \"46d00836-60a2-459f-b4c9-ed2d5e3b1fb1\") " pod="openstack/placement-74c777b4cb-m7k57" Nov 25 15:21:10 crc kubenswrapper[4890]: I1125 15:21:10.229249 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46d00836-60a2-459f-b4c9-ed2d5e3b1fb1-scripts\") pod \"placement-74c777b4cb-m7k57\" (UID: \"46d00836-60a2-459f-b4c9-ed2d5e3b1fb1\") " pod="openstack/placement-74c777b4cb-m7k57" Nov 25 15:21:10 crc kubenswrapper[4890]: I1125 15:21:10.229355 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/46d00836-60a2-459f-b4c9-ed2d5e3b1fb1-internal-tls-certs\") pod \"placement-74c777b4cb-m7k57\" (UID: \"46d00836-60a2-459f-b4c9-ed2d5e3b1fb1\") " pod="openstack/placement-74c777b4cb-m7k57" Nov 25 15:21:10 crc kubenswrapper[4890]: I1125 15:21:10.229583 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46d00836-60a2-459f-b4c9-ed2d5e3b1fb1-combined-ca-bundle\") pod \"placement-74c777b4cb-m7k57\" (UID: \"46d00836-60a2-459f-b4c9-ed2d5e3b1fb1\") " pod="openstack/placement-74c777b4cb-m7k57" Nov 25 15:21:10 crc kubenswrapper[4890]: I1125 15:21:10.231513 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/46d00836-60a2-459f-b4c9-ed2d5e3b1fb1-public-tls-certs\") pod \"placement-74c777b4cb-m7k57\" (UID: \"46d00836-60a2-459f-b4c9-ed2d5e3b1fb1\") " pod="openstack/placement-74c777b4cb-m7k57" Nov 25 15:21:10 crc kubenswrapper[4890]: I1125 15:21:10.232253 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46d00836-60a2-459f-b4c9-ed2d5e3b1fb1-config-data\") pod \"placement-74c777b4cb-m7k57\" (UID: \"46d00836-60a2-459f-b4c9-ed2d5e3b1fb1\") " pod="openstack/placement-74c777b4cb-m7k57" Nov 25 15:21:10 crc kubenswrapper[4890]: I1125 15:21:10.247223 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kx68p\" (UniqueName: \"kubernetes.io/projected/46d00836-60a2-459f-b4c9-ed2d5e3b1fb1-kube-api-access-kx68p\") pod \"placement-74c777b4cb-m7k57\" (UID: \"46d00836-60a2-459f-b4c9-ed2d5e3b1fb1\") " pod="openstack/placement-74c777b4cb-m7k57" Nov 25 15:21:10 crc kubenswrapper[4890]: I1125 15:21:10.295064 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-74c777b4cb-m7k57" Nov 25 15:21:10 crc kubenswrapper[4890]: I1125 15:21:10.589476 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-7d75bbdb6d-5dvcn" Nov 25 15:21:10 crc kubenswrapper[4890]: I1125 15:21:10.792779 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-74c777b4cb-m7k57"] Nov 25 15:21:10 crc kubenswrapper[4890]: I1125 15:21:10.908044 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-74c777b4cb-m7k57" event={"ID":"46d00836-60a2-459f-b4c9-ed2d5e3b1fb1","Type":"ContainerStarted","Data":"dc73289f9882d06601e0665806171a01a414bd1b79ca8afa21d777d70f2bca95"} Nov 25 15:21:12 crc kubenswrapper[4890]: I1125 15:21:12.295969 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-7d75bbdb6d-5dvcn" Nov 25 15:21:12 crc kubenswrapper[4890]: I1125 15:21:12.364832 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-556858788d-s2ppl"] Nov 25 15:21:12 crc kubenswrapper[4890]: I1125 15:21:12.367656 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-556858788d-s2ppl" podUID="1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f" containerName="horizon-log" containerID="cri-o://d1b1f601af90078f6742d375b3a9c3a0ce2fefdf39c2cde7a1057b3e2a3f8950" gracePeriod=30 Nov 25 15:21:12 crc kubenswrapper[4890]: I1125 15:21:12.367771 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-556858788d-s2ppl" podUID="1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f" containerName="horizon" containerID="cri-o://97df7bc09544ad16b4be8d113257f79ac856eb0352b81415283734caeb5030bb" gracePeriod=30 Nov 25 15:21:13 crc kubenswrapper[4890]: I1125 15:21:13.074315 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-64fdb55f6-22c8w" Nov 25 15:21:13 crc kubenswrapper[4890]: I1125 15:21:13.938570 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-74c777b4cb-m7k57" event={"ID":"46d00836-60a2-459f-b4c9-ed2d5e3b1fb1","Type":"ContainerStarted","Data":"a7b57bb63cd200d666894d259fe08987874764df96b3f31dd39e2cc7adec7db6"} Nov 25 15:21:14 crc kubenswrapper[4890]: I1125 15:21:14.955439 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-74c777b4cb-m7k57" event={"ID":"46d00836-60a2-459f-b4c9-ed2d5e3b1fb1","Type":"ContainerStarted","Data":"6072996e2e9cd814c362b918be3c446a0a1245b9317403e00fb31fc48c203f8a"} Nov 25 15:21:14 crc kubenswrapper[4890]: I1125 15:21:14.956118 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-74c777b4cb-m7k57" Nov 25 15:21:14 crc kubenswrapper[4890]: I1125 15:21:14.956137 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-74c777b4cb-m7k57" Nov 25 15:21:14 crc kubenswrapper[4890]: I1125 15:21:14.958536 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"62cd6c55-698d-49c9-aa1b-de9d3df17830","Type":"ContainerStarted","Data":"e1d56f3489c16630f7b8c17b3c3f7c8f73458c493793f12080e550c802afc376"} Nov 25 15:21:14 crc kubenswrapper[4890]: I1125 15:21:14.982833 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-74c777b4cb-m7k57" podStartSLOduration=5.982806481 podStartE2EDuration="5.982806481s" podCreationTimestamp="2025-11-25 15:21:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:21:14.979252201 +0000 UTC m=+1133.421714821" watchObservedRunningTime="2025-11-25 15:21:14.982806481 +0000 UTC m=+1133.425269091" Nov 25 15:21:15 crc kubenswrapper[4890]: I1125 15:21:15.386026 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 25 15:21:15 crc kubenswrapper[4890]: I1125 15:21:15.388298 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 25 15:21:15 crc kubenswrapper[4890]: I1125 15:21:15.393733 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-2j29m" Nov 25 15:21:15 crc kubenswrapper[4890]: I1125 15:21:15.394231 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Nov 25 15:21:15 crc kubenswrapper[4890]: I1125 15:21:15.394428 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Nov 25 15:21:15 crc kubenswrapper[4890]: I1125 15:21:15.419357 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 25 15:21:15 crc kubenswrapper[4890]: I1125 15:21:15.528607 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1d4df90d-315a-4483-a6e5-c6e5b322db54-openstack-config-secret\") pod \"openstackclient\" (UID: \"1d4df90d-315a-4483-a6e5-c6e5b322db54\") " pod="openstack/openstackclient" Nov 25 15:21:15 crc kubenswrapper[4890]: I1125 15:21:15.528725 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cl44x\" (UniqueName: \"kubernetes.io/projected/1d4df90d-315a-4483-a6e5-c6e5b322db54-kube-api-access-cl44x\") pod \"openstackclient\" (UID: \"1d4df90d-315a-4483-a6e5-c6e5b322db54\") " pod="openstack/openstackclient" Nov 25 15:21:15 crc kubenswrapper[4890]: I1125 15:21:15.528801 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1d4df90d-315a-4483-a6e5-c6e5b322db54-openstack-config\") pod \"openstackclient\" (UID: \"1d4df90d-315a-4483-a6e5-c6e5b322db54\") " pod="openstack/openstackclient" Nov 25 15:21:15 crc kubenswrapper[4890]: I1125 15:21:15.529067 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d4df90d-315a-4483-a6e5-c6e5b322db54-combined-ca-bundle\") pod \"openstackclient\" (UID: \"1d4df90d-315a-4483-a6e5-c6e5b322db54\") " pod="openstack/openstackclient" Nov 25 15:21:15 crc kubenswrapper[4890]: I1125 15:21:15.631182 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1d4df90d-315a-4483-a6e5-c6e5b322db54-openstack-config-secret\") pod \"openstackclient\" (UID: \"1d4df90d-315a-4483-a6e5-c6e5b322db54\") " pod="openstack/openstackclient" Nov 25 15:21:15 crc kubenswrapper[4890]: I1125 15:21:15.631565 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cl44x\" (UniqueName: \"kubernetes.io/projected/1d4df90d-315a-4483-a6e5-c6e5b322db54-kube-api-access-cl44x\") pod \"openstackclient\" (UID: \"1d4df90d-315a-4483-a6e5-c6e5b322db54\") " pod="openstack/openstackclient" Nov 25 15:21:15 crc kubenswrapper[4890]: I1125 15:21:15.631708 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1d4df90d-315a-4483-a6e5-c6e5b322db54-openstack-config\") pod \"openstackclient\" (UID: \"1d4df90d-315a-4483-a6e5-c6e5b322db54\") " pod="openstack/openstackclient" Nov 25 15:21:15 crc kubenswrapper[4890]: I1125 15:21:15.631768 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d4df90d-315a-4483-a6e5-c6e5b322db54-combined-ca-bundle\") pod \"openstackclient\" (UID: \"1d4df90d-315a-4483-a6e5-c6e5b322db54\") " pod="openstack/openstackclient" Nov 25 15:21:15 crc kubenswrapper[4890]: I1125 15:21:15.632649 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1d4df90d-315a-4483-a6e5-c6e5b322db54-openstack-config\") pod \"openstackclient\" (UID: \"1d4df90d-315a-4483-a6e5-c6e5b322db54\") " pod="openstack/openstackclient" Nov 25 15:21:15 crc kubenswrapper[4890]: I1125 15:21:15.638430 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d4df90d-315a-4483-a6e5-c6e5b322db54-combined-ca-bundle\") pod \"openstackclient\" (UID: \"1d4df90d-315a-4483-a6e5-c6e5b322db54\") " pod="openstack/openstackclient" Nov 25 15:21:15 crc kubenswrapper[4890]: I1125 15:21:15.641811 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1d4df90d-315a-4483-a6e5-c6e5b322db54-openstack-config-secret\") pod \"openstackclient\" (UID: \"1d4df90d-315a-4483-a6e5-c6e5b322db54\") " pod="openstack/openstackclient" Nov 25 15:21:15 crc kubenswrapper[4890]: I1125 15:21:15.651054 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cl44x\" (UniqueName: \"kubernetes.io/projected/1d4df90d-315a-4483-a6e5-c6e5b322db54-kube-api-access-cl44x\") pod \"openstackclient\" (UID: \"1d4df90d-315a-4483-a6e5-c6e5b322db54\") " pod="openstack/openstackclient" Nov 25 15:21:15 crc kubenswrapper[4890]: I1125 15:21:15.763298 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 25 15:21:15 crc kubenswrapper[4890]: I1125 15:21:15.972787 4890 generic.go:334] "Generic (PLEG): container finished" podID="1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f" containerID="97df7bc09544ad16b4be8d113257f79ac856eb0352b81415283734caeb5030bb" exitCode=0 Nov 25 15:21:15 crc kubenswrapper[4890]: I1125 15:21:15.972853 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-556858788d-s2ppl" event={"ID":"1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f","Type":"ContainerDied","Data":"97df7bc09544ad16b4be8d113257f79ac856eb0352b81415283734caeb5030bb"} Nov 25 15:21:15 crc kubenswrapper[4890]: I1125 15:21:15.975249 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"62cd6c55-698d-49c9-aa1b-de9d3df17830","Type":"ContainerStarted","Data":"4a9c9357e3114a61b1e169391ca67c5ecce86ad6415f1d0448587bf46f6d58fb"} Nov 25 15:21:15 crc kubenswrapper[4890]: I1125 15:21:15.976875 4890 generic.go:334] "Generic (PLEG): container finished" podID="fdcacd06-b857-46cf-8516-345e271cf882" containerID="040c842a972c23a9692699ce5354d20d5e885f515d047301cf76e6ca7da7c2c7" exitCode=0 Nov 25 15:21:15 crc kubenswrapper[4890]: I1125 15:21:15.977731 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-f8ctf" event={"ID":"fdcacd06-b857-46cf-8516-345e271cf882","Type":"ContainerDied","Data":"040c842a972c23a9692699ce5354d20d5e885f515d047301cf76e6ca7da7c2c7"} Nov 25 15:21:16 crc kubenswrapper[4890]: I1125 15:21:16.208591 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 25 15:21:16 crc kubenswrapper[4890]: I1125 15:21:16.987362 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"62cd6c55-698d-49c9-aa1b-de9d3df17830","Type":"ContainerStarted","Data":"73db5aac32d7d48e1de64725daec80bf80a30bc36525e783faecbbca1b88b262"} Nov 25 15:21:16 crc kubenswrapper[4890]: I1125 15:21:16.989325 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"1d4df90d-315a-4483-a6e5-c6e5b322db54","Type":"ContainerStarted","Data":"a09d41f08b3498415e7e41e88ac16eb40508b0bd82f28a4b72760521c4750d16"} Nov 25 15:21:17 crc kubenswrapper[4890]: I1125 15:21:17.346773 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-f8ctf" Nov 25 15:21:17 crc kubenswrapper[4890]: I1125 15:21:17.467082 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdcacd06-b857-46cf-8516-345e271cf882-combined-ca-bundle\") pod \"fdcacd06-b857-46cf-8516-345e271cf882\" (UID: \"fdcacd06-b857-46cf-8516-345e271cf882\") " Nov 25 15:21:17 crc kubenswrapper[4890]: I1125 15:21:17.467289 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fdcacd06-b857-46cf-8516-345e271cf882-db-sync-config-data\") pod \"fdcacd06-b857-46cf-8516-345e271cf882\" (UID: \"fdcacd06-b857-46cf-8516-345e271cf882\") " Nov 25 15:21:17 crc kubenswrapper[4890]: I1125 15:21:17.467390 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-szt2s\" (UniqueName: \"kubernetes.io/projected/fdcacd06-b857-46cf-8516-345e271cf882-kube-api-access-szt2s\") pod \"fdcacd06-b857-46cf-8516-345e271cf882\" (UID: \"fdcacd06-b857-46cf-8516-345e271cf882\") " Nov 25 15:21:17 crc kubenswrapper[4890]: I1125 15:21:17.472827 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdcacd06-b857-46cf-8516-345e271cf882-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "fdcacd06-b857-46cf-8516-345e271cf882" (UID: "fdcacd06-b857-46cf-8516-345e271cf882"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:21:17 crc kubenswrapper[4890]: I1125 15:21:17.486106 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fdcacd06-b857-46cf-8516-345e271cf882-kube-api-access-szt2s" (OuterVolumeSpecName: "kube-api-access-szt2s") pod "fdcacd06-b857-46cf-8516-345e271cf882" (UID: "fdcacd06-b857-46cf-8516-345e271cf882"). InnerVolumeSpecName "kube-api-access-szt2s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:21:17 crc kubenswrapper[4890]: I1125 15:21:17.498639 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdcacd06-b857-46cf-8516-345e271cf882-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fdcacd06-b857-46cf-8516-345e271cf882" (UID: "fdcacd06-b857-46cf-8516-345e271cf882"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:21:17 crc kubenswrapper[4890]: I1125 15:21:17.570091 4890 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fdcacd06-b857-46cf-8516-345e271cf882-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:17 crc kubenswrapper[4890]: I1125 15:21:17.570123 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-szt2s\" (UniqueName: \"kubernetes.io/projected/fdcacd06-b857-46cf-8516-345e271cf882-kube-api-access-szt2s\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:17 crc kubenswrapper[4890]: I1125 15:21:17.570139 4890 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdcacd06-b857-46cf-8516-345e271cf882-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.001881 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-f8ctf" event={"ID":"fdcacd06-b857-46cf-8516-345e271cf882","Type":"ContainerDied","Data":"507bac2bdb8d51b224e50337dc96073d3ec3a50ad9762887ab9180a4379d584d"} Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.001922 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="507bac2bdb8d51b224e50337dc96073d3ec3a50ad9762887ab9180a4379d584d" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.001940 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-f8ctf" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.306238 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-54dc864df6-phkw8"] Nov 25 15:21:18 crc kubenswrapper[4890]: E1125 15:21:18.313663 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdcacd06-b857-46cf-8516-345e271cf882" containerName="barbican-db-sync" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.313689 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdcacd06-b857-46cf-8516-345e271cf882" containerName="barbican-db-sync" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.313891 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdcacd06-b857-46cf-8516-345e271cf882" containerName="barbican-db-sync" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.314836 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-54dc864df6-phkw8" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.321233 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.321520 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-44nwj" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.321800 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.351306 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-77c9dbb88c-b285m"] Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.353211 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-77c9dbb88c-b285m" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.356840 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.366657 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-77c9dbb88c-b285m"] Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.382219 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-54dc864df6-phkw8"] Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.456381 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-58957f86ff-jbsf6"] Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.458137 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58957f86ff-jbsf6" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.463041 4890 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-556858788d-s2ppl" podUID="1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.145:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.145:8443: connect: connection refused" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.468701 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58957f86ff-jbsf6"] Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.491732 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b033caa-b23d-43c7-b7d4-a5465d13588b-logs\") pod \"barbican-keystone-listener-54dc864df6-phkw8\" (UID: \"0b033caa-b23d-43c7-b7d4-a5465d13588b\") " pod="openstack/barbican-keystone-listener-54dc864df6-phkw8" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.491814 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0b033caa-b23d-43c7-b7d4-a5465d13588b-config-data-custom\") pod \"barbican-keystone-listener-54dc864df6-phkw8\" (UID: \"0b033caa-b23d-43c7-b7d4-a5465d13588b\") " pod="openstack/barbican-keystone-listener-54dc864df6-phkw8" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.491892 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dc4lp\" (UniqueName: \"kubernetes.io/projected/0b033caa-b23d-43c7-b7d4-a5465d13588b-kube-api-access-dc4lp\") pod \"barbican-keystone-listener-54dc864df6-phkw8\" (UID: \"0b033caa-b23d-43c7-b7d4-a5465d13588b\") " pod="openstack/barbican-keystone-listener-54dc864df6-phkw8" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.491963 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b033caa-b23d-43c7-b7d4-a5465d13588b-config-data\") pod \"barbican-keystone-listener-54dc864df6-phkw8\" (UID: \"0b033caa-b23d-43c7-b7d4-a5465d13588b\") " pod="openstack/barbican-keystone-listener-54dc864df6-phkw8" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.491991 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hlxj\" (UniqueName: \"kubernetes.io/projected/56f2ab61-294c-4099-99d8-e5177128adec-kube-api-access-4hlxj\") pod \"barbican-worker-77c9dbb88c-b285m\" (UID: \"56f2ab61-294c-4099-99d8-e5177128adec\") " pod="openstack/barbican-worker-77c9dbb88c-b285m" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.492046 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b033caa-b23d-43c7-b7d4-a5465d13588b-combined-ca-bundle\") pod \"barbican-keystone-listener-54dc864df6-phkw8\" (UID: \"0b033caa-b23d-43c7-b7d4-a5465d13588b\") " pod="openstack/barbican-keystone-listener-54dc864df6-phkw8" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.492101 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/56f2ab61-294c-4099-99d8-e5177128adec-config-data-custom\") pod \"barbican-worker-77c9dbb88c-b285m\" (UID: \"56f2ab61-294c-4099-99d8-e5177128adec\") " pod="openstack/barbican-worker-77c9dbb88c-b285m" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.492147 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56f2ab61-294c-4099-99d8-e5177128adec-combined-ca-bundle\") pod \"barbican-worker-77c9dbb88c-b285m\" (UID: \"56f2ab61-294c-4099-99d8-e5177128adec\") " pod="openstack/barbican-worker-77c9dbb88c-b285m" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.492196 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56f2ab61-294c-4099-99d8-e5177128adec-config-data\") pod \"barbican-worker-77c9dbb88c-b285m\" (UID: \"56f2ab61-294c-4099-99d8-e5177128adec\") " pod="openstack/barbican-worker-77c9dbb88c-b285m" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.492231 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56f2ab61-294c-4099-99d8-e5177128adec-logs\") pod \"barbican-worker-77c9dbb88c-b285m\" (UID: \"56f2ab61-294c-4099-99d8-e5177128adec\") " pod="openstack/barbican-worker-77c9dbb88c-b285m" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.533185 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-8f5fb9544-r49m4"] Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.535080 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-8f5fb9544-r49m4" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.548265 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.566278 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-8f5fb9544-r49m4"] Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.593875 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hlxj\" (UniqueName: \"kubernetes.io/projected/56f2ab61-294c-4099-99d8-e5177128adec-kube-api-access-4hlxj\") pod \"barbican-worker-77c9dbb88c-b285m\" (UID: \"56f2ab61-294c-4099-99d8-e5177128adec\") " pod="openstack/barbican-worker-77c9dbb88c-b285m" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.595520 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b033caa-b23d-43c7-b7d4-a5465d13588b-combined-ca-bundle\") pod \"barbican-keystone-listener-54dc864df6-phkw8\" (UID: \"0b033caa-b23d-43c7-b7d4-a5465d13588b\") " pod="openstack/barbican-keystone-listener-54dc864df6-phkw8" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.595640 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/34bffacc-3ed2-4f5a-9338-34f1293b3776-dns-svc\") pod \"dnsmasq-dns-58957f86ff-jbsf6\" (UID: \"34bffacc-3ed2-4f5a-9338-34f1293b3776\") " pod="openstack/dnsmasq-dns-58957f86ff-jbsf6" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.595775 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/56f2ab61-294c-4099-99d8-e5177128adec-config-data-custom\") pod \"barbican-worker-77c9dbb88c-b285m\" (UID: \"56f2ab61-294c-4099-99d8-e5177128adec\") " pod="openstack/barbican-worker-77c9dbb88c-b285m" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.598836 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56f2ab61-294c-4099-99d8-e5177128adec-combined-ca-bundle\") pod \"barbican-worker-77c9dbb88c-b285m\" (UID: \"56f2ab61-294c-4099-99d8-e5177128adec\") " pod="openstack/barbican-worker-77c9dbb88c-b285m" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.598900 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56f2ab61-294c-4099-99d8-e5177128adec-config-data\") pod \"barbican-worker-77c9dbb88c-b285m\" (UID: \"56f2ab61-294c-4099-99d8-e5177128adec\") " pod="openstack/barbican-worker-77c9dbb88c-b285m" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.598930 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfr6z\" (UniqueName: \"kubernetes.io/projected/34bffacc-3ed2-4f5a-9338-34f1293b3776-kube-api-access-sfr6z\") pod \"dnsmasq-dns-58957f86ff-jbsf6\" (UID: \"34bffacc-3ed2-4f5a-9338-34f1293b3776\") " pod="openstack/dnsmasq-dns-58957f86ff-jbsf6" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.598976 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56f2ab61-294c-4099-99d8-e5177128adec-logs\") pod \"barbican-worker-77c9dbb88c-b285m\" (UID: \"56f2ab61-294c-4099-99d8-e5177128adec\") " pod="openstack/barbican-worker-77c9dbb88c-b285m" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.599033 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34bffacc-3ed2-4f5a-9338-34f1293b3776-config\") pod \"dnsmasq-dns-58957f86ff-jbsf6\" (UID: \"34bffacc-3ed2-4f5a-9338-34f1293b3776\") " pod="openstack/dnsmasq-dns-58957f86ff-jbsf6" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.599066 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/34bffacc-3ed2-4f5a-9338-34f1293b3776-ovsdbserver-nb\") pod \"dnsmasq-dns-58957f86ff-jbsf6\" (UID: \"34bffacc-3ed2-4f5a-9338-34f1293b3776\") " pod="openstack/dnsmasq-dns-58957f86ff-jbsf6" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.599142 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/34bffacc-3ed2-4f5a-9338-34f1293b3776-dns-swift-storage-0\") pod \"dnsmasq-dns-58957f86ff-jbsf6\" (UID: \"34bffacc-3ed2-4f5a-9338-34f1293b3776\") " pod="openstack/dnsmasq-dns-58957f86ff-jbsf6" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.599203 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/34bffacc-3ed2-4f5a-9338-34f1293b3776-ovsdbserver-sb\") pod \"dnsmasq-dns-58957f86ff-jbsf6\" (UID: \"34bffacc-3ed2-4f5a-9338-34f1293b3776\") " pod="openstack/dnsmasq-dns-58957f86ff-jbsf6" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.599853 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56f2ab61-294c-4099-99d8-e5177128adec-logs\") pod \"barbican-worker-77c9dbb88c-b285m\" (UID: \"56f2ab61-294c-4099-99d8-e5177128adec\") " pod="openstack/barbican-worker-77c9dbb88c-b285m" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.600044 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b033caa-b23d-43c7-b7d4-a5465d13588b-logs\") pod \"barbican-keystone-listener-54dc864df6-phkw8\" (UID: \"0b033caa-b23d-43c7-b7d4-a5465d13588b\") " pod="openstack/barbican-keystone-listener-54dc864df6-phkw8" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.600075 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0b033caa-b23d-43c7-b7d4-a5465d13588b-config-data-custom\") pod \"barbican-keystone-listener-54dc864df6-phkw8\" (UID: \"0b033caa-b23d-43c7-b7d4-a5465d13588b\") " pod="openstack/barbican-keystone-listener-54dc864df6-phkw8" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.600701 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dc4lp\" (UniqueName: \"kubernetes.io/projected/0b033caa-b23d-43c7-b7d4-a5465d13588b-kube-api-access-dc4lp\") pod \"barbican-keystone-listener-54dc864df6-phkw8\" (UID: \"0b033caa-b23d-43c7-b7d4-a5465d13588b\") " pod="openstack/barbican-keystone-listener-54dc864df6-phkw8" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.600791 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b033caa-b23d-43c7-b7d4-a5465d13588b-config-data\") pod \"barbican-keystone-listener-54dc864df6-phkw8\" (UID: \"0b033caa-b23d-43c7-b7d4-a5465d13588b\") " pod="openstack/barbican-keystone-listener-54dc864df6-phkw8" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.605037 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b033caa-b23d-43c7-b7d4-a5465d13588b-logs\") pod \"barbican-keystone-listener-54dc864df6-phkw8\" (UID: \"0b033caa-b23d-43c7-b7d4-a5465d13588b\") " pod="openstack/barbican-keystone-listener-54dc864df6-phkw8" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.608710 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/56f2ab61-294c-4099-99d8-e5177128adec-config-data-custom\") pod \"barbican-worker-77c9dbb88c-b285m\" (UID: \"56f2ab61-294c-4099-99d8-e5177128adec\") " pod="openstack/barbican-worker-77c9dbb88c-b285m" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.617274 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hlxj\" (UniqueName: \"kubernetes.io/projected/56f2ab61-294c-4099-99d8-e5177128adec-kube-api-access-4hlxj\") pod \"barbican-worker-77c9dbb88c-b285m\" (UID: \"56f2ab61-294c-4099-99d8-e5177128adec\") " pod="openstack/barbican-worker-77c9dbb88c-b285m" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.619560 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0b033caa-b23d-43c7-b7d4-a5465d13588b-config-data-custom\") pod \"barbican-keystone-listener-54dc864df6-phkw8\" (UID: \"0b033caa-b23d-43c7-b7d4-a5465d13588b\") " pod="openstack/barbican-keystone-listener-54dc864df6-phkw8" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.641595 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b033caa-b23d-43c7-b7d4-a5465d13588b-config-data\") pod \"barbican-keystone-listener-54dc864df6-phkw8\" (UID: \"0b033caa-b23d-43c7-b7d4-a5465d13588b\") " pod="openstack/barbican-keystone-listener-54dc864df6-phkw8" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.641646 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dc4lp\" (UniqueName: \"kubernetes.io/projected/0b033caa-b23d-43c7-b7d4-a5465d13588b-kube-api-access-dc4lp\") pod \"barbican-keystone-listener-54dc864df6-phkw8\" (UID: \"0b033caa-b23d-43c7-b7d4-a5465d13588b\") " pod="openstack/barbican-keystone-listener-54dc864df6-phkw8" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.642648 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56f2ab61-294c-4099-99d8-e5177128adec-combined-ca-bundle\") pod \"barbican-worker-77c9dbb88c-b285m\" (UID: \"56f2ab61-294c-4099-99d8-e5177128adec\") " pod="openstack/barbican-worker-77c9dbb88c-b285m" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.665590 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b033caa-b23d-43c7-b7d4-a5465d13588b-combined-ca-bundle\") pod \"barbican-keystone-listener-54dc864df6-phkw8\" (UID: \"0b033caa-b23d-43c7-b7d4-a5465d13588b\") " pod="openstack/barbican-keystone-listener-54dc864df6-phkw8" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.667899 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56f2ab61-294c-4099-99d8-e5177128adec-config-data\") pod \"barbican-worker-77c9dbb88c-b285m\" (UID: \"56f2ab61-294c-4099-99d8-e5177128adec\") " pod="openstack/barbican-worker-77c9dbb88c-b285m" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.704753 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-77c9dbb88c-b285m" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.715652 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f6fb686-a04a-40d5-a67a-caddba7e8d4d-config-data\") pod \"barbican-api-8f5fb9544-r49m4\" (UID: \"2f6fb686-a04a-40d5-a67a-caddba7e8d4d\") " pod="openstack/barbican-api-8f5fb9544-r49m4" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.715890 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2v8rh\" (UniqueName: \"kubernetes.io/projected/2f6fb686-a04a-40d5-a67a-caddba7e8d4d-kube-api-access-2v8rh\") pod \"barbican-api-8f5fb9544-r49m4\" (UID: \"2f6fb686-a04a-40d5-a67a-caddba7e8d4d\") " pod="openstack/barbican-api-8f5fb9544-r49m4" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.715942 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/34bffacc-3ed2-4f5a-9338-34f1293b3776-dns-svc\") pod \"dnsmasq-dns-58957f86ff-jbsf6\" (UID: \"34bffacc-3ed2-4f5a-9338-34f1293b3776\") " pod="openstack/dnsmasq-dns-58957f86ff-jbsf6" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.715979 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f6fb686-a04a-40d5-a67a-caddba7e8d4d-logs\") pod \"barbican-api-8f5fb9544-r49m4\" (UID: \"2f6fb686-a04a-40d5-a67a-caddba7e8d4d\") " pod="openstack/barbican-api-8f5fb9544-r49m4" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.716053 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfr6z\" (UniqueName: \"kubernetes.io/projected/34bffacc-3ed2-4f5a-9338-34f1293b3776-kube-api-access-sfr6z\") pod \"dnsmasq-dns-58957f86ff-jbsf6\" (UID: \"34bffacc-3ed2-4f5a-9338-34f1293b3776\") " pod="openstack/dnsmasq-dns-58957f86ff-jbsf6" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.716089 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2f6fb686-a04a-40d5-a67a-caddba7e8d4d-config-data-custom\") pod \"barbican-api-8f5fb9544-r49m4\" (UID: \"2f6fb686-a04a-40d5-a67a-caddba7e8d4d\") " pod="openstack/barbican-api-8f5fb9544-r49m4" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.716138 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f6fb686-a04a-40d5-a67a-caddba7e8d4d-combined-ca-bundle\") pod \"barbican-api-8f5fb9544-r49m4\" (UID: \"2f6fb686-a04a-40d5-a67a-caddba7e8d4d\") " pod="openstack/barbican-api-8f5fb9544-r49m4" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.716200 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34bffacc-3ed2-4f5a-9338-34f1293b3776-config\") pod \"dnsmasq-dns-58957f86ff-jbsf6\" (UID: \"34bffacc-3ed2-4f5a-9338-34f1293b3776\") " pod="openstack/dnsmasq-dns-58957f86ff-jbsf6" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.716228 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/34bffacc-3ed2-4f5a-9338-34f1293b3776-ovsdbserver-nb\") pod \"dnsmasq-dns-58957f86ff-jbsf6\" (UID: \"34bffacc-3ed2-4f5a-9338-34f1293b3776\") " pod="openstack/dnsmasq-dns-58957f86ff-jbsf6" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.716285 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/34bffacc-3ed2-4f5a-9338-34f1293b3776-dns-swift-storage-0\") pod \"dnsmasq-dns-58957f86ff-jbsf6\" (UID: \"34bffacc-3ed2-4f5a-9338-34f1293b3776\") " pod="openstack/dnsmasq-dns-58957f86ff-jbsf6" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.716323 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/34bffacc-3ed2-4f5a-9338-34f1293b3776-ovsdbserver-sb\") pod \"dnsmasq-dns-58957f86ff-jbsf6\" (UID: \"34bffacc-3ed2-4f5a-9338-34f1293b3776\") " pod="openstack/dnsmasq-dns-58957f86ff-jbsf6" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.721394 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/34bffacc-3ed2-4f5a-9338-34f1293b3776-dns-svc\") pod \"dnsmasq-dns-58957f86ff-jbsf6\" (UID: \"34bffacc-3ed2-4f5a-9338-34f1293b3776\") " pod="openstack/dnsmasq-dns-58957f86ff-jbsf6" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.722038 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34bffacc-3ed2-4f5a-9338-34f1293b3776-config\") pod \"dnsmasq-dns-58957f86ff-jbsf6\" (UID: \"34bffacc-3ed2-4f5a-9338-34f1293b3776\") " pod="openstack/dnsmasq-dns-58957f86ff-jbsf6" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.722684 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/34bffacc-3ed2-4f5a-9338-34f1293b3776-ovsdbserver-nb\") pod \"dnsmasq-dns-58957f86ff-jbsf6\" (UID: \"34bffacc-3ed2-4f5a-9338-34f1293b3776\") " pod="openstack/dnsmasq-dns-58957f86ff-jbsf6" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.723326 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/34bffacc-3ed2-4f5a-9338-34f1293b3776-dns-swift-storage-0\") pod \"dnsmasq-dns-58957f86ff-jbsf6\" (UID: \"34bffacc-3ed2-4f5a-9338-34f1293b3776\") " pod="openstack/dnsmasq-dns-58957f86ff-jbsf6" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.726263 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/34bffacc-3ed2-4f5a-9338-34f1293b3776-ovsdbserver-sb\") pod \"dnsmasq-dns-58957f86ff-jbsf6\" (UID: \"34bffacc-3ed2-4f5a-9338-34f1293b3776\") " pod="openstack/dnsmasq-dns-58957f86ff-jbsf6" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.739468 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfr6z\" (UniqueName: \"kubernetes.io/projected/34bffacc-3ed2-4f5a-9338-34f1293b3776-kube-api-access-sfr6z\") pod \"dnsmasq-dns-58957f86ff-jbsf6\" (UID: \"34bffacc-3ed2-4f5a-9338-34f1293b3776\") " pod="openstack/dnsmasq-dns-58957f86ff-jbsf6" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.794516 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58957f86ff-jbsf6" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.820106 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2v8rh\" (UniqueName: \"kubernetes.io/projected/2f6fb686-a04a-40d5-a67a-caddba7e8d4d-kube-api-access-2v8rh\") pod \"barbican-api-8f5fb9544-r49m4\" (UID: \"2f6fb686-a04a-40d5-a67a-caddba7e8d4d\") " pod="openstack/barbican-api-8f5fb9544-r49m4" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.820195 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f6fb686-a04a-40d5-a67a-caddba7e8d4d-logs\") pod \"barbican-api-8f5fb9544-r49m4\" (UID: \"2f6fb686-a04a-40d5-a67a-caddba7e8d4d\") " pod="openstack/barbican-api-8f5fb9544-r49m4" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.820307 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2f6fb686-a04a-40d5-a67a-caddba7e8d4d-config-data-custom\") pod \"barbican-api-8f5fb9544-r49m4\" (UID: \"2f6fb686-a04a-40d5-a67a-caddba7e8d4d\") " pod="openstack/barbican-api-8f5fb9544-r49m4" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.820732 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f6fb686-a04a-40d5-a67a-caddba7e8d4d-logs\") pod \"barbican-api-8f5fb9544-r49m4\" (UID: \"2f6fb686-a04a-40d5-a67a-caddba7e8d4d\") " pod="openstack/barbican-api-8f5fb9544-r49m4" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.820886 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f6fb686-a04a-40d5-a67a-caddba7e8d4d-combined-ca-bundle\") pod \"barbican-api-8f5fb9544-r49m4\" (UID: \"2f6fb686-a04a-40d5-a67a-caddba7e8d4d\") " pod="openstack/barbican-api-8f5fb9544-r49m4" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.821147 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f6fb686-a04a-40d5-a67a-caddba7e8d4d-config-data\") pod \"barbican-api-8f5fb9544-r49m4\" (UID: \"2f6fb686-a04a-40d5-a67a-caddba7e8d4d\") " pod="openstack/barbican-api-8f5fb9544-r49m4" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.824906 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2f6fb686-a04a-40d5-a67a-caddba7e8d4d-config-data-custom\") pod \"barbican-api-8f5fb9544-r49m4\" (UID: \"2f6fb686-a04a-40d5-a67a-caddba7e8d4d\") " pod="openstack/barbican-api-8f5fb9544-r49m4" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.826323 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f6fb686-a04a-40d5-a67a-caddba7e8d4d-config-data\") pod \"barbican-api-8f5fb9544-r49m4\" (UID: \"2f6fb686-a04a-40d5-a67a-caddba7e8d4d\") " pod="openstack/barbican-api-8f5fb9544-r49m4" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.828278 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f6fb686-a04a-40d5-a67a-caddba7e8d4d-combined-ca-bundle\") pod \"barbican-api-8f5fb9544-r49m4\" (UID: \"2f6fb686-a04a-40d5-a67a-caddba7e8d4d\") " pod="openstack/barbican-api-8f5fb9544-r49m4" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.854341 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2v8rh\" (UniqueName: \"kubernetes.io/projected/2f6fb686-a04a-40d5-a67a-caddba7e8d4d-kube-api-access-2v8rh\") pod \"barbican-api-8f5fb9544-r49m4\" (UID: \"2f6fb686-a04a-40d5-a67a-caddba7e8d4d\") " pod="openstack/barbican-api-8f5fb9544-r49m4" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.863288 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-8f5fb9544-r49m4" Nov 25 15:21:18 crc kubenswrapper[4890]: I1125 15:21:18.957290 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-54dc864df6-phkw8" Nov 25 15:21:19 crc kubenswrapper[4890]: I1125 15:21:19.063588 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"62cd6c55-698d-49c9-aa1b-de9d3df17830","Type":"ContainerStarted","Data":"295115d40276402bcfd8a7d506c376c1c7439de3950f5c0e6874e0a25ddfb5bf"} Nov 25 15:21:19 crc kubenswrapper[4890]: I1125 15:21:19.063867 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 25 15:21:19 crc kubenswrapper[4890]: I1125 15:21:19.113520 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.914954898 podStartE2EDuration="11.113502938s" podCreationTimestamp="2025-11-25 15:21:08 +0000 UTC" firstStartedPulling="2025-11-25 15:21:09.851417593 +0000 UTC m=+1128.293880233" lastFinishedPulling="2025-11-25 15:21:18.049965663 +0000 UTC m=+1136.492428273" observedRunningTime="2025-11-25 15:21:19.101067675 +0000 UTC m=+1137.543530315" watchObservedRunningTime="2025-11-25 15:21:19.113502938 +0000 UTC m=+1137.555965548" Nov 25 15:21:19 crc kubenswrapper[4890]: I1125 15:21:19.218839 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-77c9dbb88c-b285m"] Nov 25 15:21:19 crc kubenswrapper[4890]: W1125 15:21:19.247457 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod56f2ab61_294c_4099_99d8_e5177128adec.slice/crio-1ac421092cca804f56e419bc3024fe34da9af65e23292121757e9d19f0645583 WatchSource:0}: Error finding container 1ac421092cca804f56e419bc3024fe34da9af65e23292121757e9d19f0645583: Status 404 returned error can't find the container with id 1ac421092cca804f56e419bc3024fe34da9af65e23292121757e9d19f0645583 Nov 25 15:21:19 crc kubenswrapper[4890]: I1125 15:21:19.520755 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58957f86ff-jbsf6"] Nov 25 15:21:19 crc kubenswrapper[4890]: I1125 15:21:19.554113 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-8f5fb9544-r49m4"] Nov 25 15:21:19 crc kubenswrapper[4890]: I1125 15:21:19.685137 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-54dc864df6-phkw8"] Nov 25 15:21:19 crc kubenswrapper[4890]: W1125 15:21:19.763237 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0b033caa_b23d_43c7_b7d4_a5465d13588b.slice/crio-5de1b5383b325c2aadc86eb9a5cc4ba0304222634e9f132a77b08ee553ac72d8 WatchSource:0}: Error finding container 5de1b5383b325c2aadc86eb9a5cc4ba0304222634e9f132a77b08ee553ac72d8: Status 404 returned error can't find the container with id 5de1b5383b325c2aadc86eb9a5cc4ba0304222634e9f132a77b08ee553ac72d8 Nov 25 15:21:20 crc kubenswrapper[4890]: I1125 15:21:20.092704 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-54dc864df6-phkw8" event={"ID":"0b033caa-b23d-43c7-b7d4-a5465d13588b","Type":"ContainerStarted","Data":"5de1b5383b325c2aadc86eb9a5cc4ba0304222634e9f132a77b08ee553ac72d8"} Nov 25 15:21:20 crc kubenswrapper[4890]: I1125 15:21:20.096851 4890 generic.go:334] "Generic (PLEG): container finished" podID="dfc6fde5-9e4c-4224-8200-0f230e127f7e" containerID="30fb995c1452e7fb88486c70e38cdf87240b7d5488b0d1aaa70d5a725f84c6bf" exitCode=0 Nov 25 15:21:20 crc kubenswrapper[4890]: I1125 15:21:20.097099 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-c5b26" event={"ID":"dfc6fde5-9e4c-4224-8200-0f230e127f7e","Type":"ContainerDied","Data":"30fb995c1452e7fb88486c70e38cdf87240b7d5488b0d1aaa70d5a725f84c6bf"} Nov 25 15:21:20 crc kubenswrapper[4890]: I1125 15:21:20.105259 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-77c9dbb88c-b285m" event={"ID":"56f2ab61-294c-4099-99d8-e5177128adec","Type":"ContainerStarted","Data":"1ac421092cca804f56e419bc3024fe34da9af65e23292121757e9d19f0645583"} Nov 25 15:21:20 crc kubenswrapper[4890]: I1125 15:21:20.108595 4890 generic.go:334] "Generic (PLEG): container finished" podID="34bffacc-3ed2-4f5a-9338-34f1293b3776" containerID="c620d993985400412f84721f35e5c75fcad55b4c3d737d01500f6f2ad21a589b" exitCode=0 Nov 25 15:21:20 crc kubenswrapper[4890]: I1125 15:21:20.108688 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58957f86ff-jbsf6" event={"ID":"34bffacc-3ed2-4f5a-9338-34f1293b3776","Type":"ContainerDied","Data":"c620d993985400412f84721f35e5c75fcad55b4c3d737d01500f6f2ad21a589b"} Nov 25 15:21:20 crc kubenswrapper[4890]: I1125 15:21:20.108739 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58957f86ff-jbsf6" event={"ID":"34bffacc-3ed2-4f5a-9338-34f1293b3776","Type":"ContainerStarted","Data":"20be58ca723b1cb56f2be0b27b47cdd9f8c3c7b5f5d8ae1c4588ecc2eb955cf1"} Nov 25 15:21:20 crc kubenswrapper[4890]: I1125 15:21:20.118992 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8f5fb9544-r49m4" event={"ID":"2f6fb686-a04a-40d5-a67a-caddba7e8d4d","Type":"ContainerStarted","Data":"a1b729f7f24bb522a987cb0b04c7b9a5b52df51799327aa3e4d15458ec94d95c"} Nov 25 15:21:20 crc kubenswrapper[4890]: I1125 15:21:20.119038 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8f5fb9544-r49m4" event={"ID":"2f6fb686-a04a-40d5-a67a-caddba7e8d4d","Type":"ContainerStarted","Data":"02082ee4a3ea4ab77848fb789b7ba14148aa57252500150eaf8c8ae2605bccf1"} Nov 25 15:21:20 crc kubenswrapper[4890]: I1125 15:21:20.119055 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-8f5fb9544-r49m4" Nov 25 15:21:20 crc kubenswrapper[4890]: I1125 15:21:20.119080 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-8f5fb9544-r49m4" Nov 25 15:21:20 crc kubenswrapper[4890]: I1125 15:21:20.172212 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-8f5fb9544-r49m4" podStartSLOduration=2.172186531 podStartE2EDuration="2.172186531s" podCreationTimestamp="2025-11-25 15:21:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:21:20.167845231 +0000 UTC m=+1138.610307841" watchObservedRunningTime="2025-11-25 15:21:20.172186531 +0000 UTC m=+1138.614649141" Nov 25 15:21:21 crc kubenswrapper[4890]: I1125 15:21:21.137124 4890 generic.go:334] "Generic (PLEG): container finished" podID="449c7dd2-e3c8-48c2-8770-c1d6a77fce7b" containerID="92f0c8a918d324a2aabee016770d39884c7dbbe53d97c21e28654daf6eef0a8a" exitCode=0 Nov 25 15:21:21 crc kubenswrapper[4890]: I1125 15:21:21.137459 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-wlvxc" event={"ID":"449c7dd2-e3c8-48c2-8770-c1d6a77fce7b","Type":"ContainerDied","Data":"92f0c8a918d324a2aabee016770d39884c7dbbe53d97c21e28654daf6eef0a8a"} Nov 25 15:21:21 crc kubenswrapper[4890]: I1125 15:21:21.151407 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58957f86ff-jbsf6" event={"ID":"34bffacc-3ed2-4f5a-9338-34f1293b3776","Type":"ContainerStarted","Data":"be14c60cab33ab2ded2e1d36df0016a66348977ed17f64bc0f141bc63f78da6c"} Nov 25 15:21:21 crc kubenswrapper[4890]: I1125 15:21:21.152447 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-58957f86ff-jbsf6" Nov 25 15:21:21 crc kubenswrapper[4890]: I1125 15:21:21.163880 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8f5fb9544-r49m4" event={"ID":"2f6fb686-a04a-40d5-a67a-caddba7e8d4d","Type":"ContainerStarted","Data":"b1e404d377b43303ab6cf27c3866fde1baa533c9810625bcbf24c2695dfc0d89"} Nov 25 15:21:21 crc kubenswrapper[4890]: I1125 15:21:21.181903 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-58957f86ff-jbsf6" podStartSLOduration=3.181867428 podStartE2EDuration="3.181867428s" podCreationTimestamp="2025-11-25 15:21:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:21:21.171563789 +0000 UTC m=+1139.614026399" watchObservedRunningTime="2025-11-25 15:21:21.181867428 +0000 UTC m=+1139.624330038" Nov 25 15:21:21 crc kubenswrapper[4890]: I1125 15:21:21.730933 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7d4dcdcc94-g9drz"] Nov 25 15:21:21 crc kubenswrapper[4890]: I1125 15:21:21.732498 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7d4dcdcc94-g9drz" Nov 25 15:21:21 crc kubenswrapper[4890]: I1125 15:21:21.735228 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Nov 25 15:21:21 crc kubenswrapper[4890]: I1125 15:21:21.736523 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Nov 25 15:21:21 crc kubenswrapper[4890]: I1125 15:21:21.767694 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7d4dcdcc94-g9drz"] Nov 25 15:21:21 crc kubenswrapper[4890]: I1125 15:21:21.917310 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gt5hq\" (UniqueName: \"kubernetes.io/projected/117b4c5d-110a-472c-adc1-80846e0d37cb-kube-api-access-gt5hq\") pod \"barbican-api-7d4dcdcc94-g9drz\" (UID: \"117b4c5d-110a-472c-adc1-80846e0d37cb\") " pod="openstack/barbican-api-7d4dcdcc94-g9drz" Nov 25 15:21:21 crc kubenswrapper[4890]: I1125 15:21:21.917383 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/117b4c5d-110a-472c-adc1-80846e0d37cb-config-data\") pod \"barbican-api-7d4dcdcc94-g9drz\" (UID: \"117b4c5d-110a-472c-adc1-80846e0d37cb\") " pod="openstack/barbican-api-7d4dcdcc94-g9drz" Nov 25 15:21:21 crc kubenswrapper[4890]: I1125 15:21:21.917417 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/117b4c5d-110a-472c-adc1-80846e0d37cb-combined-ca-bundle\") pod \"barbican-api-7d4dcdcc94-g9drz\" (UID: \"117b4c5d-110a-472c-adc1-80846e0d37cb\") " pod="openstack/barbican-api-7d4dcdcc94-g9drz" Nov 25 15:21:21 crc kubenswrapper[4890]: I1125 15:21:21.917497 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/117b4c5d-110a-472c-adc1-80846e0d37cb-internal-tls-certs\") pod \"barbican-api-7d4dcdcc94-g9drz\" (UID: \"117b4c5d-110a-472c-adc1-80846e0d37cb\") " pod="openstack/barbican-api-7d4dcdcc94-g9drz" Nov 25 15:21:21 crc kubenswrapper[4890]: I1125 15:21:21.917527 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/117b4c5d-110a-472c-adc1-80846e0d37cb-public-tls-certs\") pod \"barbican-api-7d4dcdcc94-g9drz\" (UID: \"117b4c5d-110a-472c-adc1-80846e0d37cb\") " pod="openstack/barbican-api-7d4dcdcc94-g9drz" Nov 25 15:21:21 crc kubenswrapper[4890]: I1125 15:21:21.917545 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/117b4c5d-110a-472c-adc1-80846e0d37cb-logs\") pod \"barbican-api-7d4dcdcc94-g9drz\" (UID: \"117b4c5d-110a-472c-adc1-80846e0d37cb\") " pod="openstack/barbican-api-7d4dcdcc94-g9drz" Nov 25 15:21:21 crc kubenswrapper[4890]: I1125 15:21:21.917566 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/117b4c5d-110a-472c-adc1-80846e0d37cb-config-data-custom\") pod \"barbican-api-7d4dcdcc94-g9drz\" (UID: \"117b4c5d-110a-472c-adc1-80846e0d37cb\") " pod="openstack/barbican-api-7d4dcdcc94-g9drz" Nov 25 15:21:22 crc kubenswrapper[4890]: I1125 15:21:22.019713 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/117b4c5d-110a-472c-adc1-80846e0d37cb-config-data-custom\") pod \"barbican-api-7d4dcdcc94-g9drz\" (UID: \"117b4c5d-110a-472c-adc1-80846e0d37cb\") " pod="openstack/barbican-api-7d4dcdcc94-g9drz" Nov 25 15:21:22 crc kubenswrapper[4890]: I1125 15:21:22.020421 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gt5hq\" (UniqueName: \"kubernetes.io/projected/117b4c5d-110a-472c-adc1-80846e0d37cb-kube-api-access-gt5hq\") pod \"barbican-api-7d4dcdcc94-g9drz\" (UID: \"117b4c5d-110a-472c-adc1-80846e0d37cb\") " pod="openstack/barbican-api-7d4dcdcc94-g9drz" Nov 25 15:21:22 crc kubenswrapper[4890]: I1125 15:21:22.020580 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/117b4c5d-110a-472c-adc1-80846e0d37cb-config-data\") pod \"barbican-api-7d4dcdcc94-g9drz\" (UID: \"117b4c5d-110a-472c-adc1-80846e0d37cb\") " pod="openstack/barbican-api-7d4dcdcc94-g9drz" Nov 25 15:21:22 crc kubenswrapper[4890]: I1125 15:21:22.020714 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/117b4c5d-110a-472c-adc1-80846e0d37cb-combined-ca-bundle\") pod \"barbican-api-7d4dcdcc94-g9drz\" (UID: \"117b4c5d-110a-472c-adc1-80846e0d37cb\") " pod="openstack/barbican-api-7d4dcdcc94-g9drz" Nov 25 15:21:22 crc kubenswrapper[4890]: I1125 15:21:22.020959 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/117b4c5d-110a-472c-adc1-80846e0d37cb-internal-tls-certs\") pod \"barbican-api-7d4dcdcc94-g9drz\" (UID: \"117b4c5d-110a-472c-adc1-80846e0d37cb\") " pod="openstack/barbican-api-7d4dcdcc94-g9drz" Nov 25 15:21:22 crc kubenswrapper[4890]: I1125 15:21:22.021079 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/117b4c5d-110a-472c-adc1-80846e0d37cb-public-tls-certs\") pod \"barbican-api-7d4dcdcc94-g9drz\" (UID: \"117b4c5d-110a-472c-adc1-80846e0d37cb\") " pod="openstack/barbican-api-7d4dcdcc94-g9drz" Nov 25 15:21:22 crc kubenswrapper[4890]: I1125 15:21:22.021205 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/117b4c5d-110a-472c-adc1-80846e0d37cb-logs\") pod \"barbican-api-7d4dcdcc94-g9drz\" (UID: \"117b4c5d-110a-472c-adc1-80846e0d37cb\") " pod="openstack/barbican-api-7d4dcdcc94-g9drz" Nov 25 15:21:22 crc kubenswrapper[4890]: I1125 15:21:22.021625 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/117b4c5d-110a-472c-adc1-80846e0d37cb-logs\") pod \"barbican-api-7d4dcdcc94-g9drz\" (UID: \"117b4c5d-110a-472c-adc1-80846e0d37cb\") " pod="openstack/barbican-api-7d4dcdcc94-g9drz" Nov 25 15:21:22 crc kubenswrapper[4890]: I1125 15:21:22.026943 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/117b4c5d-110a-472c-adc1-80846e0d37cb-combined-ca-bundle\") pod \"barbican-api-7d4dcdcc94-g9drz\" (UID: \"117b4c5d-110a-472c-adc1-80846e0d37cb\") " pod="openstack/barbican-api-7d4dcdcc94-g9drz" Nov 25 15:21:22 crc kubenswrapper[4890]: I1125 15:21:22.026986 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/117b4c5d-110a-472c-adc1-80846e0d37cb-internal-tls-certs\") pod \"barbican-api-7d4dcdcc94-g9drz\" (UID: \"117b4c5d-110a-472c-adc1-80846e0d37cb\") " pod="openstack/barbican-api-7d4dcdcc94-g9drz" Nov 25 15:21:22 crc kubenswrapper[4890]: I1125 15:21:22.028126 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/117b4c5d-110a-472c-adc1-80846e0d37cb-config-data\") pod \"barbican-api-7d4dcdcc94-g9drz\" (UID: \"117b4c5d-110a-472c-adc1-80846e0d37cb\") " pod="openstack/barbican-api-7d4dcdcc94-g9drz" Nov 25 15:21:22 crc kubenswrapper[4890]: I1125 15:21:22.039703 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/117b4c5d-110a-472c-adc1-80846e0d37cb-public-tls-certs\") pod \"barbican-api-7d4dcdcc94-g9drz\" (UID: \"117b4c5d-110a-472c-adc1-80846e0d37cb\") " pod="openstack/barbican-api-7d4dcdcc94-g9drz" Nov 25 15:21:22 crc kubenswrapper[4890]: I1125 15:21:22.040132 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/117b4c5d-110a-472c-adc1-80846e0d37cb-config-data-custom\") pod \"barbican-api-7d4dcdcc94-g9drz\" (UID: \"117b4c5d-110a-472c-adc1-80846e0d37cb\") " pod="openstack/barbican-api-7d4dcdcc94-g9drz" Nov 25 15:21:22 crc kubenswrapper[4890]: I1125 15:21:22.042566 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gt5hq\" (UniqueName: \"kubernetes.io/projected/117b4c5d-110a-472c-adc1-80846e0d37cb-kube-api-access-gt5hq\") pod \"barbican-api-7d4dcdcc94-g9drz\" (UID: \"117b4c5d-110a-472c-adc1-80846e0d37cb\") " pod="openstack/barbican-api-7d4dcdcc94-g9drz" Nov 25 15:21:22 crc kubenswrapper[4890]: I1125 15:21:22.078945 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7d4dcdcc94-g9drz" Nov 25 15:21:22 crc kubenswrapper[4890]: I1125 15:21:22.198067 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-5f45cd8b8f-8wpds"] Nov 25 15:21:22 crc kubenswrapper[4890]: I1125 15:21:22.199791 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5f45cd8b8f-8wpds" Nov 25 15:21:22 crc kubenswrapper[4890]: I1125 15:21:22.204642 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Nov 25 15:21:22 crc kubenswrapper[4890]: I1125 15:21:22.204894 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 25 15:21:22 crc kubenswrapper[4890]: I1125 15:21:22.205030 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Nov 25 15:21:22 crc kubenswrapper[4890]: I1125 15:21:22.236423 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-5f45cd8b8f-8wpds"] Nov 25 15:21:22 crc kubenswrapper[4890]: I1125 15:21:22.346049 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lfmw\" (UniqueName: \"kubernetes.io/projected/66cfb07e-ac1a-4aee-8fbd-7871e2e30d88-kube-api-access-6lfmw\") pod \"swift-proxy-5f45cd8b8f-8wpds\" (UID: \"66cfb07e-ac1a-4aee-8fbd-7871e2e30d88\") " pod="openstack/swift-proxy-5f45cd8b8f-8wpds" Nov 25 15:21:22 crc kubenswrapper[4890]: I1125 15:21:22.346105 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/66cfb07e-ac1a-4aee-8fbd-7871e2e30d88-etc-swift\") pod \"swift-proxy-5f45cd8b8f-8wpds\" (UID: \"66cfb07e-ac1a-4aee-8fbd-7871e2e30d88\") " pod="openstack/swift-proxy-5f45cd8b8f-8wpds" Nov 25 15:21:22 crc kubenswrapper[4890]: I1125 15:21:22.346154 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/66cfb07e-ac1a-4aee-8fbd-7871e2e30d88-internal-tls-certs\") pod \"swift-proxy-5f45cd8b8f-8wpds\" (UID: \"66cfb07e-ac1a-4aee-8fbd-7871e2e30d88\") " pod="openstack/swift-proxy-5f45cd8b8f-8wpds" Nov 25 15:21:22 crc kubenswrapper[4890]: I1125 15:21:22.346185 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66cfb07e-ac1a-4aee-8fbd-7871e2e30d88-config-data\") pod \"swift-proxy-5f45cd8b8f-8wpds\" (UID: \"66cfb07e-ac1a-4aee-8fbd-7871e2e30d88\") " pod="openstack/swift-proxy-5f45cd8b8f-8wpds" Nov 25 15:21:22 crc kubenswrapper[4890]: I1125 15:21:22.346207 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66cfb07e-ac1a-4aee-8fbd-7871e2e30d88-combined-ca-bundle\") pod \"swift-proxy-5f45cd8b8f-8wpds\" (UID: \"66cfb07e-ac1a-4aee-8fbd-7871e2e30d88\") " pod="openstack/swift-proxy-5f45cd8b8f-8wpds" Nov 25 15:21:22 crc kubenswrapper[4890]: I1125 15:21:22.346238 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/66cfb07e-ac1a-4aee-8fbd-7871e2e30d88-log-httpd\") pod \"swift-proxy-5f45cd8b8f-8wpds\" (UID: \"66cfb07e-ac1a-4aee-8fbd-7871e2e30d88\") " pod="openstack/swift-proxy-5f45cd8b8f-8wpds" Nov 25 15:21:22 crc kubenswrapper[4890]: I1125 15:21:22.346274 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/66cfb07e-ac1a-4aee-8fbd-7871e2e30d88-public-tls-certs\") pod \"swift-proxy-5f45cd8b8f-8wpds\" (UID: \"66cfb07e-ac1a-4aee-8fbd-7871e2e30d88\") " pod="openstack/swift-proxy-5f45cd8b8f-8wpds" Nov 25 15:21:22 crc kubenswrapper[4890]: I1125 15:21:22.346433 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/66cfb07e-ac1a-4aee-8fbd-7871e2e30d88-run-httpd\") pod \"swift-proxy-5f45cd8b8f-8wpds\" (UID: \"66cfb07e-ac1a-4aee-8fbd-7871e2e30d88\") " pod="openstack/swift-proxy-5f45cd8b8f-8wpds" Nov 25 15:21:22 crc kubenswrapper[4890]: I1125 15:21:22.448658 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66cfb07e-ac1a-4aee-8fbd-7871e2e30d88-combined-ca-bundle\") pod \"swift-proxy-5f45cd8b8f-8wpds\" (UID: \"66cfb07e-ac1a-4aee-8fbd-7871e2e30d88\") " pod="openstack/swift-proxy-5f45cd8b8f-8wpds" Nov 25 15:21:22 crc kubenswrapper[4890]: I1125 15:21:22.449003 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/66cfb07e-ac1a-4aee-8fbd-7871e2e30d88-log-httpd\") pod \"swift-proxy-5f45cd8b8f-8wpds\" (UID: \"66cfb07e-ac1a-4aee-8fbd-7871e2e30d88\") " pod="openstack/swift-proxy-5f45cd8b8f-8wpds" Nov 25 15:21:22 crc kubenswrapper[4890]: I1125 15:21:22.449037 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/66cfb07e-ac1a-4aee-8fbd-7871e2e30d88-public-tls-certs\") pod \"swift-proxy-5f45cd8b8f-8wpds\" (UID: \"66cfb07e-ac1a-4aee-8fbd-7871e2e30d88\") " pod="openstack/swift-proxy-5f45cd8b8f-8wpds" Nov 25 15:21:22 crc kubenswrapper[4890]: I1125 15:21:22.449119 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/66cfb07e-ac1a-4aee-8fbd-7871e2e30d88-run-httpd\") pod \"swift-proxy-5f45cd8b8f-8wpds\" (UID: \"66cfb07e-ac1a-4aee-8fbd-7871e2e30d88\") " pod="openstack/swift-proxy-5f45cd8b8f-8wpds" Nov 25 15:21:22 crc kubenswrapper[4890]: I1125 15:21:22.449138 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lfmw\" (UniqueName: \"kubernetes.io/projected/66cfb07e-ac1a-4aee-8fbd-7871e2e30d88-kube-api-access-6lfmw\") pod \"swift-proxy-5f45cd8b8f-8wpds\" (UID: \"66cfb07e-ac1a-4aee-8fbd-7871e2e30d88\") " pod="openstack/swift-proxy-5f45cd8b8f-8wpds" Nov 25 15:21:22 crc kubenswrapper[4890]: I1125 15:21:22.449420 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/66cfb07e-ac1a-4aee-8fbd-7871e2e30d88-etc-swift\") pod \"swift-proxy-5f45cd8b8f-8wpds\" (UID: \"66cfb07e-ac1a-4aee-8fbd-7871e2e30d88\") " pod="openstack/swift-proxy-5f45cd8b8f-8wpds" Nov 25 15:21:22 crc kubenswrapper[4890]: I1125 15:21:22.449458 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/66cfb07e-ac1a-4aee-8fbd-7871e2e30d88-internal-tls-certs\") pod \"swift-proxy-5f45cd8b8f-8wpds\" (UID: \"66cfb07e-ac1a-4aee-8fbd-7871e2e30d88\") " pod="openstack/swift-proxy-5f45cd8b8f-8wpds" Nov 25 15:21:22 crc kubenswrapper[4890]: I1125 15:21:22.449479 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66cfb07e-ac1a-4aee-8fbd-7871e2e30d88-config-data\") pod \"swift-proxy-5f45cd8b8f-8wpds\" (UID: \"66cfb07e-ac1a-4aee-8fbd-7871e2e30d88\") " pod="openstack/swift-proxy-5f45cd8b8f-8wpds" Nov 25 15:21:22 crc kubenswrapper[4890]: I1125 15:21:22.450105 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/66cfb07e-ac1a-4aee-8fbd-7871e2e30d88-log-httpd\") pod \"swift-proxy-5f45cd8b8f-8wpds\" (UID: \"66cfb07e-ac1a-4aee-8fbd-7871e2e30d88\") " pod="openstack/swift-proxy-5f45cd8b8f-8wpds" Nov 25 15:21:22 crc kubenswrapper[4890]: I1125 15:21:22.450194 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/66cfb07e-ac1a-4aee-8fbd-7871e2e30d88-run-httpd\") pod \"swift-proxy-5f45cd8b8f-8wpds\" (UID: \"66cfb07e-ac1a-4aee-8fbd-7871e2e30d88\") " pod="openstack/swift-proxy-5f45cd8b8f-8wpds" Nov 25 15:21:22 crc kubenswrapper[4890]: I1125 15:21:22.455336 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/66cfb07e-ac1a-4aee-8fbd-7871e2e30d88-public-tls-certs\") pod \"swift-proxy-5f45cd8b8f-8wpds\" (UID: \"66cfb07e-ac1a-4aee-8fbd-7871e2e30d88\") " pod="openstack/swift-proxy-5f45cd8b8f-8wpds" Nov 25 15:21:22 crc kubenswrapper[4890]: I1125 15:21:22.456897 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/66cfb07e-ac1a-4aee-8fbd-7871e2e30d88-internal-tls-certs\") pod \"swift-proxy-5f45cd8b8f-8wpds\" (UID: \"66cfb07e-ac1a-4aee-8fbd-7871e2e30d88\") " pod="openstack/swift-proxy-5f45cd8b8f-8wpds" Nov 25 15:21:22 crc kubenswrapper[4890]: I1125 15:21:22.458813 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66cfb07e-ac1a-4aee-8fbd-7871e2e30d88-combined-ca-bundle\") pod \"swift-proxy-5f45cd8b8f-8wpds\" (UID: \"66cfb07e-ac1a-4aee-8fbd-7871e2e30d88\") " pod="openstack/swift-proxy-5f45cd8b8f-8wpds" Nov 25 15:21:22 crc kubenswrapper[4890]: I1125 15:21:22.462077 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66cfb07e-ac1a-4aee-8fbd-7871e2e30d88-config-data\") pod \"swift-proxy-5f45cd8b8f-8wpds\" (UID: \"66cfb07e-ac1a-4aee-8fbd-7871e2e30d88\") " pod="openstack/swift-proxy-5f45cd8b8f-8wpds" Nov 25 15:21:22 crc kubenswrapper[4890]: I1125 15:21:22.464023 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/66cfb07e-ac1a-4aee-8fbd-7871e2e30d88-etc-swift\") pod \"swift-proxy-5f45cd8b8f-8wpds\" (UID: \"66cfb07e-ac1a-4aee-8fbd-7871e2e30d88\") " pod="openstack/swift-proxy-5f45cd8b8f-8wpds" Nov 25 15:21:22 crc kubenswrapper[4890]: I1125 15:21:22.473732 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lfmw\" (UniqueName: \"kubernetes.io/projected/66cfb07e-ac1a-4aee-8fbd-7871e2e30d88-kube-api-access-6lfmw\") pod \"swift-proxy-5f45cd8b8f-8wpds\" (UID: \"66cfb07e-ac1a-4aee-8fbd-7871e2e30d88\") " pod="openstack/swift-proxy-5f45cd8b8f-8wpds" Nov 25 15:21:22 crc kubenswrapper[4890]: I1125 15:21:22.558867 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5f45cd8b8f-8wpds" Nov 25 15:21:23 crc kubenswrapper[4890]: I1125 15:21:23.323915 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:21:23 crc kubenswrapper[4890]: I1125 15:21:23.324524 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="62cd6c55-698d-49c9-aa1b-de9d3df17830" containerName="ceilometer-central-agent" containerID="cri-o://e1d56f3489c16630f7b8c17b3c3f7c8f73458c493793f12080e550c802afc376" gracePeriod=30 Nov 25 15:21:23 crc kubenswrapper[4890]: I1125 15:21:23.325007 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="62cd6c55-698d-49c9-aa1b-de9d3df17830" containerName="proxy-httpd" containerID="cri-o://295115d40276402bcfd8a7d506c376c1c7439de3950f5c0e6874e0a25ddfb5bf" gracePeriod=30 Nov 25 15:21:23 crc kubenswrapper[4890]: I1125 15:21:23.325066 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="62cd6c55-698d-49c9-aa1b-de9d3df17830" containerName="sg-core" containerID="cri-o://73db5aac32d7d48e1de64725daec80bf80a30bc36525e783faecbbca1b88b262" gracePeriod=30 Nov 25 15:21:23 crc kubenswrapper[4890]: I1125 15:21:23.325179 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="62cd6c55-698d-49c9-aa1b-de9d3df17830" containerName="ceilometer-notification-agent" containerID="cri-o://4a9c9357e3114a61b1e169391ca67c5ecce86ad6415f1d0448587bf46f6d58fb" gracePeriod=30 Nov 25 15:21:24 crc kubenswrapper[4890]: I1125 15:21:24.207429 4890 generic.go:334] "Generic (PLEG): container finished" podID="62cd6c55-698d-49c9-aa1b-de9d3df17830" containerID="295115d40276402bcfd8a7d506c376c1c7439de3950f5c0e6874e0a25ddfb5bf" exitCode=0 Nov 25 15:21:24 crc kubenswrapper[4890]: I1125 15:21:24.207827 4890 generic.go:334] "Generic (PLEG): container finished" podID="62cd6c55-698d-49c9-aa1b-de9d3df17830" containerID="73db5aac32d7d48e1de64725daec80bf80a30bc36525e783faecbbca1b88b262" exitCode=2 Nov 25 15:21:24 crc kubenswrapper[4890]: I1125 15:21:24.207842 4890 generic.go:334] "Generic (PLEG): container finished" podID="62cd6c55-698d-49c9-aa1b-de9d3df17830" containerID="4a9c9357e3114a61b1e169391ca67c5ecce86ad6415f1d0448587bf46f6d58fb" exitCode=0 Nov 25 15:21:24 crc kubenswrapper[4890]: I1125 15:21:24.207851 4890 generic.go:334] "Generic (PLEG): container finished" podID="62cd6c55-698d-49c9-aa1b-de9d3df17830" containerID="e1d56f3489c16630f7b8c17b3c3f7c8f73458c493793f12080e550c802afc376" exitCode=0 Nov 25 15:21:24 crc kubenswrapper[4890]: I1125 15:21:24.207493 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"62cd6c55-698d-49c9-aa1b-de9d3df17830","Type":"ContainerDied","Data":"295115d40276402bcfd8a7d506c376c1c7439de3950f5c0e6874e0a25ddfb5bf"} Nov 25 15:21:24 crc kubenswrapper[4890]: I1125 15:21:24.207889 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"62cd6c55-698d-49c9-aa1b-de9d3df17830","Type":"ContainerDied","Data":"73db5aac32d7d48e1de64725daec80bf80a30bc36525e783faecbbca1b88b262"} Nov 25 15:21:24 crc kubenswrapper[4890]: I1125 15:21:24.207907 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"62cd6c55-698d-49c9-aa1b-de9d3df17830","Type":"ContainerDied","Data":"4a9c9357e3114a61b1e169391ca67c5ecce86ad6415f1d0448587bf46f6d58fb"} Nov 25 15:21:24 crc kubenswrapper[4890]: I1125 15:21:24.207921 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"62cd6c55-698d-49c9-aa1b-de9d3df17830","Type":"ContainerDied","Data":"e1d56f3489c16630f7b8c17b3c3f7c8f73458c493793f12080e550c802afc376"} Nov 25 15:21:26 crc kubenswrapper[4890]: I1125 15:21:26.450434 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 15:21:26 crc kubenswrapper[4890]: I1125 15:21:26.451003 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 15:21:26 crc kubenswrapper[4890]: I1125 15:21:26.451135 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" Nov 25 15:21:26 crc kubenswrapper[4890]: I1125 15:21:26.452956 4890 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"53708c37db2dded329d6f7de3f6f52a10c460a5fc2cf14a1b264e8539a50f494"} pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 15:21:26 crc kubenswrapper[4890]: I1125 15:21:26.453134 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" containerID="cri-o://53708c37db2dded329d6f7de3f6f52a10c460a5fc2cf14a1b264e8539a50f494" gracePeriod=600 Nov 25 15:21:27 crc kubenswrapper[4890]: I1125 15:21:27.241602 4890 generic.go:334] "Generic (PLEG): container finished" podID="4e4f849d-f239-4727-a73e-18327856929a" containerID="53708c37db2dded329d6f7de3f6f52a10c460a5fc2cf14a1b264e8539a50f494" exitCode=0 Nov 25 15:21:27 crc kubenswrapper[4890]: I1125 15:21:27.241645 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" event={"ID":"4e4f849d-f239-4727-a73e-18327856929a","Type":"ContainerDied","Data":"53708c37db2dded329d6f7de3f6f52a10c460a5fc2cf14a1b264e8539a50f494"} Nov 25 15:21:27 crc kubenswrapper[4890]: I1125 15:21:27.241676 4890 scope.go:117] "RemoveContainer" containerID="b51c3e1b3270ec2e9b5e884e9f00a4274f242cf4d6c3240274a36d149cb2249e" Nov 25 15:21:28 crc kubenswrapper[4890]: I1125 15:21:28.461755 4890 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-556858788d-s2ppl" podUID="1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.145:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.145:8443: connect: connection refused" Nov 25 15:21:28 crc kubenswrapper[4890]: I1125 15:21:28.797125 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-58957f86ff-jbsf6" Nov 25 15:21:28 crc kubenswrapper[4890]: I1125 15:21:28.949080 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-2hjbr"] Nov 25 15:21:28 crc kubenswrapper[4890]: I1125 15:21:28.949474 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-58dd9ff6bc-2hjbr" podUID="e5750e4c-060e-4dfa-9592-9295e1e44359" containerName="dnsmasq-dns" containerID="cri-o://098c37f5179061c75b6eebf65637f3531baf46cd67545d729b0204eef4197600" gracePeriod=10 Nov 25 15:21:30 crc kubenswrapper[4890]: I1125 15:21:30.508948 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-8f5fb9544-r49m4" Nov 25 15:21:30 crc kubenswrapper[4890]: I1125 15:21:30.955988 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-8f5fb9544-r49m4" Nov 25 15:21:31 crc kubenswrapper[4890]: I1125 15:21:31.287826 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-2hjbr" event={"ID":"e5750e4c-060e-4dfa-9592-9295e1e44359","Type":"ContainerDied","Data":"098c37f5179061c75b6eebf65637f3531baf46cd67545d729b0204eef4197600"} Nov 25 15:21:31 crc kubenswrapper[4890]: I1125 15:21:31.287774 4890 generic.go:334] "Generic (PLEG): container finished" podID="e5750e4c-060e-4dfa-9592-9295e1e44359" containerID="098c37f5179061c75b6eebf65637f3531baf46cd67545d729b0204eef4197600" exitCode=0 Nov 25 15:21:32 crc kubenswrapper[4890]: I1125 15:21:32.410098 4890 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-58dd9ff6bc-2hjbr" podUID="e5750e4c-060e-4dfa-9592-9295e1e44359" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.149:5353: connect: connection refused" Nov 25 15:21:34 crc kubenswrapper[4890]: E1125 15:21:34.534293 4890 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified" Nov 25 15:21:34 crc kubenswrapper[4890]: E1125 15:21:34.534777 4890 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:openstackclient,Image:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,Command:[/bin/sleep],Args:[infinity],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n58dh6fh54fhcch5cch5bfh66ch58ch98hfhcbh655h56fh646h74h696h5c4h68dh598hc9h565h4h7hd9hc8h59dh68ch557h9ch5f9h577h59cq,ValueFrom:nil,},EnvVar{Name:OS_CLOUD,Value:default,ValueFrom:nil,},EnvVar{Name:PROMETHEUS_HOST,Value:metric-storage-prometheus.openstack.svc,ValueFrom:nil,},EnvVar{Name:PROMETHEUS_PORT,Value:9090,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:openstack-config,ReadOnly:false,MountPath:/home/cloud-admin/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/home/cloud-admin/.config/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/home/cloud-admin/cloudrc,SubPath:cloudrc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cl44x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42401,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42401,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstackclient_openstack(1d4df90d-315a-4483-a6e5-c6e5b322db54): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 15:21:34 crc kubenswrapper[4890]: E1125 15:21:34.536500 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstackclient\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstackclient" podUID="1d4df90d-315a-4483-a6e5-c6e5b322db54" Nov 25 15:21:34 crc kubenswrapper[4890]: I1125 15:21:34.846901 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-c5b26" Nov 25 15:21:34 crc kubenswrapper[4890]: I1125 15:21:34.868625 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-wlvxc" Nov 25 15:21:34 crc kubenswrapper[4890]: I1125 15:21:34.935095 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/449c7dd2-e3c8-48c2-8770-c1d6a77fce7b-db-sync-config-data\") pod \"449c7dd2-e3c8-48c2-8770-c1d6a77fce7b\" (UID: \"449c7dd2-e3c8-48c2-8770-c1d6a77fce7b\") " Nov 25 15:21:34 crc kubenswrapper[4890]: I1125 15:21:34.935488 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfc6fde5-9e4c-4224-8200-0f230e127f7e-config-data\") pod \"dfc6fde5-9e4c-4224-8200-0f230e127f7e\" (UID: \"dfc6fde5-9e4c-4224-8200-0f230e127f7e\") " Nov 25 15:21:34 crc kubenswrapper[4890]: I1125 15:21:34.935541 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/449c7dd2-e3c8-48c2-8770-c1d6a77fce7b-combined-ca-bundle\") pod \"449c7dd2-e3c8-48c2-8770-c1d6a77fce7b\" (UID: \"449c7dd2-e3c8-48c2-8770-c1d6a77fce7b\") " Nov 25 15:21:34 crc kubenswrapper[4890]: I1125 15:21:34.935602 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfc6fde5-9e4c-4224-8200-0f230e127f7e-combined-ca-bundle\") pod \"dfc6fde5-9e4c-4224-8200-0f230e127f7e\" (UID: \"dfc6fde5-9e4c-4224-8200-0f230e127f7e\") " Nov 25 15:21:34 crc kubenswrapper[4890]: I1125 15:21:34.935664 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/dfc6fde5-9e4c-4224-8200-0f230e127f7e-db-sync-config-data\") pod \"dfc6fde5-9e4c-4224-8200-0f230e127f7e\" (UID: \"dfc6fde5-9e4c-4224-8200-0f230e127f7e\") " Nov 25 15:21:34 crc kubenswrapper[4890]: I1125 15:21:34.935691 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/449c7dd2-e3c8-48c2-8770-c1d6a77fce7b-config-data\") pod \"449c7dd2-e3c8-48c2-8770-c1d6a77fce7b\" (UID: \"449c7dd2-e3c8-48c2-8770-c1d6a77fce7b\") " Nov 25 15:21:34 crc kubenswrapper[4890]: I1125 15:21:34.935752 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxv7c\" (UniqueName: \"kubernetes.io/projected/dfc6fde5-9e4c-4224-8200-0f230e127f7e-kube-api-access-fxv7c\") pod \"dfc6fde5-9e4c-4224-8200-0f230e127f7e\" (UID: \"dfc6fde5-9e4c-4224-8200-0f230e127f7e\") " Nov 25 15:21:34 crc kubenswrapper[4890]: I1125 15:21:34.935813 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/449c7dd2-e3c8-48c2-8770-c1d6a77fce7b-etc-machine-id\") pod \"449c7dd2-e3c8-48c2-8770-c1d6a77fce7b\" (UID: \"449c7dd2-e3c8-48c2-8770-c1d6a77fce7b\") " Nov 25 15:21:34 crc kubenswrapper[4890]: I1125 15:21:34.935836 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/449c7dd2-e3c8-48c2-8770-c1d6a77fce7b-scripts\") pod \"449c7dd2-e3c8-48c2-8770-c1d6a77fce7b\" (UID: \"449c7dd2-e3c8-48c2-8770-c1d6a77fce7b\") " Nov 25 15:21:34 crc kubenswrapper[4890]: I1125 15:21:34.935894 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q98k8\" (UniqueName: \"kubernetes.io/projected/449c7dd2-e3c8-48c2-8770-c1d6a77fce7b-kube-api-access-q98k8\") pod \"449c7dd2-e3c8-48c2-8770-c1d6a77fce7b\" (UID: \"449c7dd2-e3c8-48c2-8770-c1d6a77fce7b\") " Nov 25 15:21:34 crc kubenswrapper[4890]: I1125 15:21:34.942277 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/449c7dd2-e3c8-48c2-8770-c1d6a77fce7b-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "449c7dd2-e3c8-48c2-8770-c1d6a77fce7b" (UID: "449c7dd2-e3c8-48c2-8770-c1d6a77fce7b"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 15:21:34 crc kubenswrapper[4890]: I1125 15:21:34.947714 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/449c7dd2-e3c8-48c2-8770-c1d6a77fce7b-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "449c7dd2-e3c8-48c2-8770-c1d6a77fce7b" (UID: "449c7dd2-e3c8-48c2-8770-c1d6a77fce7b"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:21:34 crc kubenswrapper[4890]: I1125 15:21:34.962525 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/449c7dd2-e3c8-48c2-8770-c1d6a77fce7b-kube-api-access-q98k8" (OuterVolumeSpecName: "kube-api-access-q98k8") pod "449c7dd2-e3c8-48c2-8770-c1d6a77fce7b" (UID: "449c7dd2-e3c8-48c2-8770-c1d6a77fce7b"). InnerVolumeSpecName "kube-api-access-q98k8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:21:34 crc kubenswrapper[4890]: I1125 15:21:34.964408 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfc6fde5-9e4c-4224-8200-0f230e127f7e-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "dfc6fde5-9e4c-4224-8200-0f230e127f7e" (UID: "dfc6fde5-9e4c-4224-8200-0f230e127f7e"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:21:34 crc kubenswrapper[4890]: I1125 15:21:34.964647 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfc6fde5-9e4c-4224-8200-0f230e127f7e-kube-api-access-fxv7c" (OuterVolumeSpecName: "kube-api-access-fxv7c") pod "dfc6fde5-9e4c-4224-8200-0f230e127f7e" (UID: "dfc6fde5-9e4c-4224-8200-0f230e127f7e"). InnerVolumeSpecName "kube-api-access-fxv7c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:21:34 crc kubenswrapper[4890]: I1125 15:21:34.970930 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/449c7dd2-e3c8-48c2-8770-c1d6a77fce7b-scripts" (OuterVolumeSpecName: "scripts") pod "449c7dd2-e3c8-48c2-8770-c1d6a77fce7b" (UID: "449c7dd2-e3c8-48c2-8770-c1d6a77fce7b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.038659 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxv7c\" (UniqueName: \"kubernetes.io/projected/dfc6fde5-9e4c-4224-8200-0f230e127f7e-kube-api-access-fxv7c\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.038698 4890 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/449c7dd2-e3c8-48c2-8770-c1d6a77fce7b-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.038711 4890 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/449c7dd2-e3c8-48c2-8770-c1d6a77fce7b-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.038741 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q98k8\" (UniqueName: \"kubernetes.io/projected/449c7dd2-e3c8-48c2-8770-c1d6a77fce7b-kube-api-access-q98k8\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.038755 4890 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/449c7dd2-e3c8-48c2-8770-c1d6a77fce7b-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.038767 4890 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/dfc6fde5-9e4c-4224-8200-0f230e127f7e-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.120276 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dd9ff6bc-2hjbr" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.124573 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.139484 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62cd6c55-698d-49c9-aa1b-de9d3df17830-run-httpd\") pod \"62cd6c55-698d-49c9-aa1b-de9d3df17830\" (UID: \"62cd6c55-698d-49c9-aa1b-de9d3df17830\") " Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.139525 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e5750e4c-060e-4dfa-9592-9295e1e44359-ovsdbserver-nb\") pod \"e5750e4c-060e-4dfa-9592-9295e1e44359\" (UID: \"e5750e4c-060e-4dfa-9592-9295e1e44359\") " Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.139551 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/62cd6c55-698d-49c9-aa1b-de9d3df17830-sg-core-conf-yaml\") pod \"62cd6c55-698d-49c9-aa1b-de9d3df17830\" (UID: \"62cd6c55-698d-49c9-aa1b-de9d3df17830\") " Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.139568 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62cd6c55-698d-49c9-aa1b-de9d3df17830-config-data\") pod \"62cd6c55-698d-49c9-aa1b-de9d3df17830\" (UID: \"62cd6c55-698d-49c9-aa1b-de9d3df17830\") " Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.139612 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e5750e4c-060e-4dfa-9592-9295e1e44359-dns-svc\") pod \"e5750e4c-060e-4dfa-9592-9295e1e44359\" (UID: \"e5750e4c-060e-4dfa-9592-9295e1e44359\") " Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.139644 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62cd6c55-698d-49c9-aa1b-de9d3df17830-combined-ca-bundle\") pod \"62cd6c55-698d-49c9-aa1b-de9d3df17830\" (UID: \"62cd6c55-698d-49c9-aa1b-de9d3df17830\") " Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.140020 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5750e4c-060e-4dfa-9592-9295e1e44359-config\") pod \"e5750e4c-060e-4dfa-9592-9295e1e44359\" (UID: \"e5750e4c-060e-4dfa-9592-9295e1e44359\") " Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.140056 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mctq\" (UniqueName: \"kubernetes.io/projected/e5750e4c-060e-4dfa-9592-9295e1e44359-kube-api-access-6mctq\") pod \"e5750e4c-060e-4dfa-9592-9295e1e44359\" (UID: \"e5750e4c-060e-4dfa-9592-9295e1e44359\") " Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.140035 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62cd6c55-698d-49c9-aa1b-de9d3df17830-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "62cd6c55-698d-49c9-aa1b-de9d3df17830" (UID: "62cd6c55-698d-49c9-aa1b-de9d3df17830"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.140111 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e5750e4c-060e-4dfa-9592-9295e1e44359-dns-swift-storage-0\") pod \"e5750e4c-060e-4dfa-9592-9295e1e44359\" (UID: \"e5750e4c-060e-4dfa-9592-9295e1e44359\") " Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.140182 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62cd6c55-698d-49c9-aa1b-de9d3df17830-scripts\") pod \"62cd6c55-698d-49c9-aa1b-de9d3df17830\" (UID: \"62cd6c55-698d-49c9-aa1b-de9d3df17830\") " Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.140206 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62cd6c55-698d-49c9-aa1b-de9d3df17830-log-httpd\") pod \"62cd6c55-698d-49c9-aa1b-de9d3df17830\" (UID: \"62cd6c55-698d-49c9-aa1b-de9d3df17830\") " Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.140279 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e5750e4c-060e-4dfa-9592-9295e1e44359-ovsdbserver-sb\") pod \"e5750e4c-060e-4dfa-9592-9295e1e44359\" (UID: \"e5750e4c-060e-4dfa-9592-9295e1e44359\") " Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.140308 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2frz\" (UniqueName: \"kubernetes.io/projected/62cd6c55-698d-49c9-aa1b-de9d3df17830-kube-api-access-x2frz\") pod \"62cd6c55-698d-49c9-aa1b-de9d3df17830\" (UID: \"62cd6c55-698d-49c9-aa1b-de9d3df17830\") " Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.140673 4890 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62cd6c55-698d-49c9-aa1b-de9d3df17830-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.141927 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62cd6c55-698d-49c9-aa1b-de9d3df17830-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "62cd6c55-698d-49c9-aa1b-de9d3df17830" (UID: "62cd6c55-698d-49c9-aa1b-de9d3df17830"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.174440 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62cd6c55-698d-49c9-aa1b-de9d3df17830-kube-api-access-x2frz" (OuterVolumeSpecName: "kube-api-access-x2frz") pod "62cd6c55-698d-49c9-aa1b-de9d3df17830" (UID: "62cd6c55-698d-49c9-aa1b-de9d3df17830"). InnerVolumeSpecName "kube-api-access-x2frz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.177501 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62cd6c55-698d-49c9-aa1b-de9d3df17830-scripts" (OuterVolumeSpecName: "scripts") pod "62cd6c55-698d-49c9-aa1b-de9d3df17830" (UID: "62cd6c55-698d-49c9-aa1b-de9d3df17830"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.210922 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5750e4c-060e-4dfa-9592-9295e1e44359-kube-api-access-6mctq" (OuterVolumeSpecName: "kube-api-access-6mctq") pod "e5750e4c-060e-4dfa-9592-9295e1e44359" (UID: "e5750e4c-060e-4dfa-9592-9295e1e44359"). InnerVolumeSpecName "kube-api-access-6mctq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.212251 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/449c7dd2-e3c8-48c2-8770-c1d6a77fce7b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "449c7dd2-e3c8-48c2-8770-c1d6a77fce7b" (UID: "449c7dd2-e3c8-48c2-8770-c1d6a77fce7b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.222697 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7d4dcdcc94-g9drz"] Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.242011 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2frz\" (UniqueName: \"kubernetes.io/projected/62cd6c55-698d-49c9-aa1b-de9d3df17830-kube-api-access-x2frz\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.242334 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mctq\" (UniqueName: \"kubernetes.io/projected/e5750e4c-060e-4dfa-9592-9295e1e44359-kube-api-access-6mctq\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.242360 4890 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62cd6c55-698d-49c9-aa1b-de9d3df17830-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.242373 4890 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62cd6c55-698d-49c9-aa1b-de9d3df17830-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.242384 4890 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/449c7dd2-e3c8-48c2-8770-c1d6a77fce7b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.250565 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfc6fde5-9e4c-4224-8200-0f230e127f7e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dfc6fde5-9e4c-4224-8200-0f230e127f7e" (UID: "dfc6fde5-9e4c-4224-8200-0f230e127f7e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.328599 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-wlvxc" event={"ID":"449c7dd2-e3c8-48c2-8770-c1d6a77fce7b","Type":"ContainerDied","Data":"42047d72e03e641622835078e3ab9955e61599a2e2d9caf1245c6703ca3426d5"} Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.328638 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42047d72e03e641622835078e3ab9955e61599a2e2d9caf1245c6703ca3426d5" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.328697 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-wlvxc" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.330371 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-c5b26" event={"ID":"dfc6fde5-9e4c-4224-8200-0f230e127f7e","Type":"ContainerDied","Data":"6aa17f501d336d60c4f2eccff69fa11460eb1feac76b5ad7dc2b5e14dee3ae4a"} Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.330394 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6aa17f501d336d60c4f2eccff69fa11460eb1feac76b5ad7dc2b5e14dee3ae4a" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.330428 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-c5b26" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.332082 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7d4dcdcc94-g9drz" event={"ID":"117b4c5d-110a-472c-adc1-80846e0d37cb","Type":"ContainerStarted","Data":"b6dbd0bf2d8aaf7e3516e7aa1be33334794e6ec559818d71ec08822aaf4e82b9"} Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.333477 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-2hjbr" event={"ID":"e5750e4c-060e-4dfa-9592-9295e1e44359","Type":"ContainerDied","Data":"630c580be58b970646e55355182892687cd7347528ab0ef1a920711906b3cdce"} Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.333508 4890 scope.go:117] "RemoveContainer" containerID="098c37f5179061c75b6eebf65637f3531baf46cd67545d729b0204eef4197600" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.333586 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dd9ff6bc-2hjbr" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.336325 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" event={"ID":"4e4f849d-f239-4727-a73e-18327856929a","Type":"ContainerStarted","Data":"325b4a907d6b6c6f59e5653baa1b3273925ff29fed8f729e95af24a885f5e532"} Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.340577 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.340752 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"62cd6c55-698d-49c9-aa1b-de9d3df17830","Type":"ContainerDied","Data":"f8922b08dda2a1fb0a8e5d472196fecaf8fb947469f950eadb0811f40df6ae5f"} Nov 25 15:21:35 crc kubenswrapper[4890]: E1125 15:21:35.342264 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstackclient\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified\\\"\"" pod="openstack/openstackclient" podUID="1d4df90d-315a-4483-a6e5-c6e5b322db54" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.350539 4890 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfc6fde5-9e4c-4224-8200-0f230e127f7e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.379634 4890 scope.go:117] "RemoveContainer" containerID="7d04dec3867ee45102cae2108850a714091b4e31593cb0da520c5d67c49c75d0" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.407148 4890 scope.go:117] "RemoveContainer" containerID="295115d40276402bcfd8a7d506c376c1c7439de3950f5c0e6874e0a25ddfb5bf" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.453905 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62cd6c55-698d-49c9-aa1b-de9d3df17830-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "62cd6c55-698d-49c9-aa1b-de9d3df17830" (UID: "62cd6c55-698d-49c9-aa1b-de9d3df17830"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.454616 4890 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/62cd6c55-698d-49c9-aa1b-de9d3df17830-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.479545 4890 scope.go:117] "RemoveContainer" containerID="73db5aac32d7d48e1de64725daec80bf80a30bc36525e783faecbbca1b88b262" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.484143 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-5f45cd8b8f-8wpds"] Nov 25 15:21:35 crc kubenswrapper[4890]: W1125 15:21:35.491047 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod66cfb07e_ac1a_4aee_8fbd_7871e2e30d88.slice/crio-b685c1c74b6128a96424dcf5d471780256c03c8ee64e5c988f6312afc00d7905 WatchSource:0}: Error finding container b685c1c74b6128a96424dcf5d471780256c03c8ee64e5c988f6312afc00d7905: Status 404 returned error can't find the container with id b685c1c74b6128a96424dcf5d471780256c03c8ee64e5c988f6312afc00d7905 Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.503009 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfc6fde5-9e4c-4224-8200-0f230e127f7e-config-data" (OuterVolumeSpecName: "config-data") pod "dfc6fde5-9e4c-4224-8200-0f230e127f7e" (UID: "dfc6fde5-9e4c-4224-8200-0f230e127f7e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.518931 4890 scope.go:117] "RemoveContainer" containerID="4a9c9357e3114a61b1e169391ca67c5ecce86ad6415f1d0448587bf46f6d58fb" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.549461 4890 scope.go:117] "RemoveContainer" containerID="e1d56f3489c16630f7b8c17b3c3f7c8f73458c493793f12080e550c802afc376" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.556561 4890 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfc6fde5-9e4c-4224-8200-0f230e127f7e-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.597448 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/449c7dd2-e3c8-48c2-8770-c1d6a77fce7b-config-data" (OuterVolumeSpecName: "config-data") pod "449c7dd2-e3c8-48c2-8770-c1d6a77fce7b" (UID: "449c7dd2-e3c8-48c2-8770-c1d6a77fce7b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.620641 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5750e4c-060e-4dfa-9592-9295e1e44359-config" (OuterVolumeSpecName: "config") pod "e5750e4c-060e-4dfa-9592-9295e1e44359" (UID: "e5750e4c-060e-4dfa-9592-9295e1e44359"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.631442 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62cd6c55-698d-49c9-aa1b-de9d3df17830-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "62cd6c55-698d-49c9-aa1b-de9d3df17830" (UID: "62cd6c55-698d-49c9-aa1b-de9d3df17830"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.634064 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5750e4c-060e-4dfa-9592-9295e1e44359-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e5750e4c-060e-4dfa-9592-9295e1e44359" (UID: "e5750e4c-060e-4dfa-9592-9295e1e44359"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.649994 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5750e4c-060e-4dfa-9592-9295e1e44359-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e5750e4c-060e-4dfa-9592-9295e1e44359" (UID: "e5750e4c-060e-4dfa-9592-9295e1e44359"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.651529 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5750e4c-060e-4dfa-9592-9295e1e44359-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e5750e4c-060e-4dfa-9592-9295e1e44359" (UID: "e5750e4c-060e-4dfa-9592-9295e1e44359"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.653829 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5750e4c-060e-4dfa-9592-9295e1e44359-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e5750e4c-060e-4dfa-9592-9295e1e44359" (UID: "e5750e4c-060e-4dfa-9592-9295e1e44359"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.657518 4890 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e5750e4c-060e-4dfa-9592-9295e1e44359-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.657578 4890 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e5750e4c-060e-4dfa-9592-9295e1e44359-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.657588 4890 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62cd6c55-698d-49c9-aa1b-de9d3df17830-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.657597 4890 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5750e4c-060e-4dfa-9592-9295e1e44359-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.657605 4890 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e5750e4c-060e-4dfa-9592-9295e1e44359-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.657613 4890 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e5750e4c-060e-4dfa-9592-9295e1e44359-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.657622 4890 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/449c7dd2-e3c8-48c2-8770-c1d6a77fce7b-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.695397 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62cd6c55-698d-49c9-aa1b-de9d3df17830-config-data" (OuterVolumeSpecName: "config-data") pod "62cd6c55-698d-49c9-aa1b-de9d3df17830" (UID: "62cd6c55-698d-49c9-aa1b-de9d3df17830"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.767913 4890 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62cd6c55-698d-49c9-aa1b-de9d3df17830-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.975335 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-2hjbr"] Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.981624 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-2hjbr"] Nov 25 15:21:35 crc kubenswrapper[4890]: I1125 15:21:35.995150 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.004892 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.014674 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:21:36 crc kubenswrapper[4890]: E1125 15:21:36.015740 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62cd6c55-698d-49c9-aa1b-de9d3df17830" containerName="sg-core" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.015760 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="62cd6c55-698d-49c9-aa1b-de9d3df17830" containerName="sg-core" Nov 25 15:21:36 crc kubenswrapper[4890]: E1125 15:21:36.015771 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62cd6c55-698d-49c9-aa1b-de9d3df17830" containerName="ceilometer-central-agent" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.015778 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="62cd6c55-698d-49c9-aa1b-de9d3df17830" containerName="ceilometer-central-agent" Nov 25 15:21:36 crc kubenswrapper[4890]: E1125 15:21:36.015794 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="449c7dd2-e3c8-48c2-8770-c1d6a77fce7b" containerName="cinder-db-sync" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.015801 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="449c7dd2-e3c8-48c2-8770-c1d6a77fce7b" containerName="cinder-db-sync" Nov 25 15:21:36 crc kubenswrapper[4890]: E1125 15:21:36.015812 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62cd6c55-698d-49c9-aa1b-de9d3df17830" containerName="proxy-httpd" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.015818 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="62cd6c55-698d-49c9-aa1b-de9d3df17830" containerName="proxy-httpd" Nov 25 15:21:36 crc kubenswrapper[4890]: E1125 15:21:36.015826 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5750e4c-060e-4dfa-9592-9295e1e44359" containerName="init" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.015831 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5750e4c-060e-4dfa-9592-9295e1e44359" containerName="init" Nov 25 15:21:36 crc kubenswrapper[4890]: E1125 15:21:36.015845 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfc6fde5-9e4c-4224-8200-0f230e127f7e" containerName="glance-db-sync" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.015851 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfc6fde5-9e4c-4224-8200-0f230e127f7e" containerName="glance-db-sync" Nov 25 15:21:36 crc kubenswrapper[4890]: E1125 15:21:36.015865 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62cd6c55-698d-49c9-aa1b-de9d3df17830" containerName="ceilometer-notification-agent" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.015870 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="62cd6c55-698d-49c9-aa1b-de9d3df17830" containerName="ceilometer-notification-agent" Nov 25 15:21:36 crc kubenswrapper[4890]: E1125 15:21:36.015878 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5750e4c-060e-4dfa-9592-9295e1e44359" containerName="dnsmasq-dns" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.015883 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5750e4c-060e-4dfa-9592-9295e1e44359" containerName="dnsmasq-dns" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.016634 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="62cd6c55-698d-49c9-aa1b-de9d3df17830" containerName="ceilometer-notification-agent" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.016657 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="62cd6c55-698d-49c9-aa1b-de9d3df17830" containerName="sg-core" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.016666 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="62cd6c55-698d-49c9-aa1b-de9d3df17830" containerName="ceilometer-central-agent" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.016675 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5750e4c-060e-4dfa-9592-9295e1e44359" containerName="dnsmasq-dns" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.016682 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="449c7dd2-e3c8-48c2-8770-c1d6a77fce7b" containerName="cinder-db-sync" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.016693 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfc6fde5-9e4c-4224-8200-0f230e127f7e" containerName="glance-db-sync" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.016700 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="62cd6c55-698d-49c9-aa1b-de9d3df17830" containerName="proxy-httpd" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.024584 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.027899 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.028097 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.028728 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.072369 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1ae99b8-5004-434a-812d-0ab8043bd031-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f1ae99b8-5004-434a-812d-0ab8043bd031\") " pod="openstack/ceilometer-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.072486 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f1ae99b8-5004-434a-812d-0ab8043bd031-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f1ae99b8-5004-434a-812d-0ab8043bd031\") " pod="openstack/ceilometer-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.072520 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1ae99b8-5004-434a-812d-0ab8043bd031-config-data\") pod \"ceilometer-0\" (UID: \"f1ae99b8-5004-434a-812d-0ab8043bd031\") " pod="openstack/ceilometer-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.072564 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f1ae99b8-5004-434a-812d-0ab8043bd031-log-httpd\") pod \"ceilometer-0\" (UID: \"f1ae99b8-5004-434a-812d-0ab8043bd031\") " pod="openstack/ceilometer-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.072579 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f1ae99b8-5004-434a-812d-0ab8043bd031-scripts\") pod \"ceilometer-0\" (UID: \"f1ae99b8-5004-434a-812d-0ab8043bd031\") " pod="openstack/ceilometer-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.072617 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvg6v\" (UniqueName: \"kubernetes.io/projected/f1ae99b8-5004-434a-812d-0ab8043bd031-kube-api-access-fvg6v\") pod \"ceilometer-0\" (UID: \"f1ae99b8-5004-434a-812d-0ab8043bd031\") " pod="openstack/ceilometer-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.072642 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f1ae99b8-5004-434a-812d-0ab8043bd031-run-httpd\") pod \"ceilometer-0\" (UID: \"f1ae99b8-5004-434a-812d-0ab8043bd031\") " pod="openstack/ceilometer-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.173915 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f1ae99b8-5004-434a-812d-0ab8043bd031-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f1ae99b8-5004-434a-812d-0ab8043bd031\") " pod="openstack/ceilometer-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.174006 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1ae99b8-5004-434a-812d-0ab8043bd031-config-data\") pod \"ceilometer-0\" (UID: \"f1ae99b8-5004-434a-812d-0ab8043bd031\") " pod="openstack/ceilometer-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.174119 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f1ae99b8-5004-434a-812d-0ab8043bd031-log-httpd\") pod \"ceilometer-0\" (UID: \"f1ae99b8-5004-434a-812d-0ab8043bd031\") " pod="openstack/ceilometer-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.174140 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f1ae99b8-5004-434a-812d-0ab8043bd031-scripts\") pod \"ceilometer-0\" (UID: \"f1ae99b8-5004-434a-812d-0ab8043bd031\") " pod="openstack/ceilometer-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.174196 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvg6v\" (UniqueName: \"kubernetes.io/projected/f1ae99b8-5004-434a-812d-0ab8043bd031-kube-api-access-fvg6v\") pod \"ceilometer-0\" (UID: \"f1ae99b8-5004-434a-812d-0ab8043bd031\") " pod="openstack/ceilometer-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.174224 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f1ae99b8-5004-434a-812d-0ab8043bd031-run-httpd\") pod \"ceilometer-0\" (UID: \"f1ae99b8-5004-434a-812d-0ab8043bd031\") " pod="openstack/ceilometer-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.174269 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1ae99b8-5004-434a-812d-0ab8043bd031-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f1ae99b8-5004-434a-812d-0ab8043bd031\") " pod="openstack/ceilometer-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.180305 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f1ae99b8-5004-434a-812d-0ab8043bd031-run-httpd\") pod \"ceilometer-0\" (UID: \"f1ae99b8-5004-434a-812d-0ab8043bd031\") " pod="openstack/ceilometer-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.184045 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f1ae99b8-5004-434a-812d-0ab8043bd031-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f1ae99b8-5004-434a-812d-0ab8043bd031\") " pod="openstack/ceilometer-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.184408 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f1ae99b8-5004-434a-812d-0ab8043bd031-log-httpd\") pod \"ceilometer-0\" (UID: \"f1ae99b8-5004-434a-812d-0ab8043bd031\") " pod="openstack/ceilometer-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.192975 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1ae99b8-5004-434a-812d-0ab8043bd031-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f1ae99b8-5004-434a-812d-0ab8043bd031\") " pod="openstack/ceilometer-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.197874 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f1ae99b8-5004-434a-812d-0ab8043bd031-scripts\") pod \"ceilometer-0\" (UID: \"f1ae99b8-5004-434a-812d-0ab8043bd031\") " pod="openstack/ceilometer-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.222713 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62cd6c55-698d-49c9-aa1b-de9d3df17830" path="/var/lib/kubelet/pods/62cd6c55-698d-49c9-aa1b-de9d3df17830/volumes" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.223416 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5750e4c-060e-4dfa-9592-9295e1e44359" path="/var/lib/kubelet/pods/e5750e4c-060e-4dfa-9592-9295e1e44359/volumes" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.225493 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1ae99b8-5004-434a-812d-0ab8043bd031-config-data\") pod \"ceilometer-0\" (UID: \"f1ae99b8-5004-434a-812d-0ab8043bd031\") " pod="openstack/ceilometer-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.231450 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.245340 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.253179 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvg6v\" (UniqueName: \"kubernetes.io/projected/f1ae99b8-5004-434a-812d-0ab8043bd031-kube-api-access-fvg6v\") pod \"ceilometer-0\" (UID: \"f1ae99b8-5004-434a-812d-0ab8043bd031\") " pod="openstack/ceilometer-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.275356 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d65af5aa-095d-465f-8fd7-ba76b4ec08af-config-data\") pod \"cinder-scheduler-0\" (UID: \"d65af5aa-095d-465f-8fd7-ba76b4ec08af\") " pod="openstack/cinder-scheduler-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.275401 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d65af5aa-095d-465f-8fd7-ba76b4ec08af-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"d65af5aa-095d-465f-8fd7-ba76b4ec08af\") " pod="openstack/cinder-scheduler-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.275427 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8fkv\" (UniqueName: \"kubernetes.io/projected/d65af5aa-095d-465f-8fd7-ba76b4ec08af-kube-api-access-q8fkv\") pod \"cinder-scheduler-0\" (UID: \"d65af5aa-095d-465f-8fd7-ba76b4ec08af\") " pod="openstack/cinder-scheduler-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.275503 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d65af5aa-095d-465f-8fd7-ba76b4ec08af-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"d65af5aa-095d-465f-8fd7-ba76b4ec08af\") " pod="openstack/cinder-scheduler-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.275575 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d65af5aa-095d-465f-8fd7-ba76b4ec08af-scripts\") pod \"cinder-scheduler-0\" (UID: \"d65af5aa-095d-465f-8fd7-ba76b4ec08af\") " pod="openstack/cinder-scheduler-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.275597 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d65af5aa-095d-465f-8fd7-ba76b4ec08af-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"d65af5aa-095d-465f-8fd7-ba76b4ec08af\") " pod="openstack/cinder-scheduler-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.275975 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.276216 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.276475 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.276533 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-bgrtc" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.288235 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.311779 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7bf4c8dd6c-n8rxv"] Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.313323 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bf4c8dd6c-n8rxv" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.342587 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bf4c8dd6c-n8rxv"] Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.358732 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.396847 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d65af5aa-095d-465f-8fd7-ba76b4ec08af-scripts\") pod \"cinder-scheduler-0\" (UID: \"d65af5aa-095d-465f-8fd7-ba76b4ec08af\") " pod="openstack/cinder-scheduler-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.396908 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3c7f9325-f498-4cd8-96fe-ec9ec9f57a83-ovsdbserver-nb\") pod \"dnsmasq-dns-7bf4c8dd6c-n8rxv\" (UID: \"3c7f9325-f498-4cd8-96fe-ec9ec9f57a83\") " pod="openstack/dnsmasq-dns-7bf4c8dd6c-n8rxv" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.396943 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d65af5aa-095d-465f-8fd7-ba76b4ec08af-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"d65af5aa-095d-465f-8fd7-ba76b4ec08af\") " pod="openstack/cinder-scheduler-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.397082 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3c7f9325-f498-4cd8-96fe-ec9ec9f57a83-dns-swift-storage-0\") pod \"dnsmasq-dns-7bf4c8dd6c-n8rxv\" (UID: \"3c7f9325-f498-4cd8-96fe-ec9ec9f57a83\") " pod="openstack/dnsmasq-dns-7bf4c8dd6c-n8rxv" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.397904 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3c7f9325-f498-4cd8-96fe-ec9ec9f57a83-ovsdbserver-sb\") pod \"dnsmasq-dns-7bf4c8dd6c-n8rxv\" (UID: \"3c7f9325-f498-4cd8-96fe-ec9ec9f57a83\") " pod="openstack/dnsmasq-dns-7bf4c8dd6c-n8rxv" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.397992 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d65af5aa-095d-465f-8fd7-ba76b4ec08af-config-data\") pod \"cinder-scheduler-0\" (UID: \"d65af5aa-095d-465f-8fd7-ba76b4ec08af\") " pod="openstack/cinder-scheduler-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.398139 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d65af5aa-095d-465f-8fd7-ba76b4ec08af-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"d65af5aa-095d-465f-8fd7-ba76b4ec08af\") " pod="openstack/cinder-scheduler-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.398423 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8fkv\" (UniqueName: \"kubernetes.io/projected/d65af5aa-095d-465f-8fd7-ba76b4ec08af-kube-api-access-q8fkv\") pod \"cinder-scheduler-0\" (UID: \"d65af5aa-095d-465f-8fd7-ba76b4ec08af\") " pod="openstack/cinder-scheduler-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.398603 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4dhq\" (UniqueName: \"kubernetes.io/projected/3c7f9325-f498-4cd8-96fe-ec9ec9f57a83-kube-api-access-c4dhq\") pod \"dnsmasq-dns-7bf4c8dd6c-n8rxv\" (UID: \"3c7f9325-f498-4cd8-96fe-ec9ec9f57a83\") " pod="openstack/dnsmasq-dns-7bf4c8dd6c-n8rxv" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.398771 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c7f9325-f498-4cd8-96fe-ec9ec9f57a83-config\") pod \"dnsmasq-dns-7bf4c8dd6c-n8rxv\" (UID: \"3c7f9325-f498-4cd8-96fe-ec9ec9f57a83\") " pod="openstack/dnsmasq-dns-7bf4c8dd6c-n8rxv" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.398923 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3c7f9325-f498-4cd8-96fe-ec9ec9f57a83-dns-svc\") pod \"dnsmasq-dns-7bf4c8dd6c-n8rxv\" (UID: \"3c7f9325-f498-4cd8-96fe-ec9ec9f57a83\") " pod="openstack/dnsmasq-dns-7bf4c8dd6c-n8rxv" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.399082 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d65af5aa-095d-465f-8fd7-ba76b4ec08af-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"d65af5aa-095d-465f-8fd7-ba76b4ec08af\") " pod="openstack/cinder-scheduler-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.399671 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d65af5aa-095d-465f-8fd7-ba76b4ec08af-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"d65af5aa-095d-465f-8fd7-ba76b4ec08af\") " pod="openstack/cinder-scheduler-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.403487 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d65af5aa-095d-465f-8fd7-ba76b4ec08af-scripts\") pod \"cinder-scheduler-0\" (UID: \"d65af5aa-095d-465f-8fd7-ba76b4ec08af\") " pod="openstack/cinder-scheduler-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.404341 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d65af5aa-095d-465f-8fd7-ba76b4ec08af-config-data\") pod \"cinder-scheduler-0\" (UID: \"d65af5aa-095d-465f-8fd7-ba76b4ec08af\") " pod="openstack/cinder-scheduler-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.408868 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d65af5aa-095d-465f-8fd7-ba76b4ec08af-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"d65af5aa-095d-465f-8fd7-ba76b4ec08af\") " pod="openstack/cinder-scheduler-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.418365 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d65af5aa-095d-465f-8fd7-ba76b4ec08af-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"d65af5aa-095d-465f-8fd7-ba76b4ec08af\") " pod="openstack/cinder-scheduler-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.423466 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-54dc864df6-phkw8" event={"ID":"0b033caa-b23d-43c7-b7d4-a5465d13588b","Type":"ContainerStarted","Data":"ea3cdf9491c1e43a21aaeae8a011c438c2879495c2444d6438098e2b412215e3"} Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.423519 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-54dc864df6-phkw8" event={"ID":"0b033caa-b23d-43c7-b7d4-a5465d13588b","Type":"ContainerStarted","Data":"2b5f0a3df11bc04fde93341154f00217b4a88bf2ee1b4eba0b65b7c9bbd185ca"} Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.451997 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8fkv\" (UniqueName: \"kubernetes.io/projected/d65af5aa-095d-465f-8fd7-ba76b4ec08af-kube-api-access-q8fkv\") pod \"cinder-scheduler-0\" (UID: \"d65af5aa-095d-465f-8fd7-ba76b4ec08af\") " pod="openstack/cinder-scheduler-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.464704 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-77c9dbb88c-b285m" event={"ID":"56f2ab61-294c-4099-99d8-e5177128adec","Type":"ContainerStarted","Data":"6ba0a7a3925f51d5e5003186a4c599aac7cd65f2bab248c9bcfd66e59010e976"} Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.464752 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-77c9dbb88c-b285m" event={"ID":"56f2ab61-294c-4099-99d8-e5177128adec","Type":"ContainerStarted","Data":"d2db77e0f056a1be0b2d45b5fea7212782488df438c6a3c3ed0da41ddcb44d09"} Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.485999 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7d4dcdcc94-g9drz" event={"ID":"117b4c5d-110a-472c-adc1-80846e0d37cb","Type":"ContainerStarted","Data":"039347263eebffdfe7a20843d02d8682ad3d0b0e5baf7741cc3e52bb1355b8b7"} Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.486042 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7d4dcdcc94-g9drz" event={"ID":"117b4c5d-110a-472c-adc1-80846e0d37cb","Type":"ContainerStarted","Data":"f9e1e8ba5f0ed49eb8d6c3a2f5b9bde14ebcab9a6edad3f504dcfe1ed7be26c2"} Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.486752 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7d4dcdcc94-g9drz" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.486774 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7d4dcdcc94-g9drz" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.490916 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5f45cd8b8f-8wpds" event={"ID":"66cfb07e-ac1a-4aee-8fbd-7871e2e30d88","Type":"ContainerStarted","Data":"f0a30f7f1b52f8f2914518522937c707c7cafddbf2c1a0da13113af3a65c4336"} Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.490956 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.492438 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5f45cd8b8f-8wpds" event={"ID":"66cfb07e-ac1a-4aee-8fbd-7871e2e30d88","Type":"ContainerStarted","Data":"b685c1c74b6128a96424dcf5d471780256c03c8ee64e5c988f6312afc00d7905"} Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.492517 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.502479 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.503345 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3c7f9325-f498-4cd8-96fe-ec9ec9f57a83-ovsdbserver-sb\") pod \"dnsmasq-dns-7bf4c8dd6c-n8rxv\" (UID: \"3c7f9325-f498-4cd8-96fe-ec9ec9f57a83\") " pod="openstack/dnsmasq-dns-7bf4c8dd6c-n8rxv" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.503505 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4dhq\" (UniqueName: \"kubernetes.io/projected/3c7f9325-f498-4cd8-96fe-ec9ec9f57a83-kube-api-access-c4dhq\") pod \"dnsmasq-dns-7bf4c8dd6c-n8rxv\" (UID: \"3c7f9325-f498-4cd8-96fe-ec9ec9f57a83\") " pod="openstack/dnsmasq-dns-7bf4c8dd6c-n8rxv" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.503558 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c7f9325-f498-4cd8-96fe-ec9ec9f57a83-config\") pod \"dnsmasq-dns-7bf4c8dd6c-n8rxv\" (UID: \"3c7f9325-f498-4cd8-96fe-ec9ec9f57a83\") " pod="openstack/dnsmasq-dns-7bf4c8dd6c-n8rxv" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.503584 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3c7f9325-f498-4cd8-96fe-ec9ec9f57a83-dns-svc\") pod \"dnsmasq-dns-7bf4c8dd6c-n8rxv\" (UID: \"3c7f9325-f498-4cd8-96fe-ec9ec9f57a83\") " pod="openstack/dnsmasq-dns-7bf4c8dd6c-n8rxv" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.503690 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3c7f9325-f498-4cd8-96fe-ec9ec9f57a83-ovsdbserver-nb\") pod \"dnsmasq-dns-7bf4c8dd6c-n8rxv\" (UID: \"3c7f9325-f498-4cd8-96fe-ec9ec9f57a83\") " pod="openstack/dnsmasq-dns-7bf4c8dd6c-n8rxv" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.503854 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3c7f9325-f498-4cd8-96fe-ec9ec9f57a83-dns-swift-storage-0\") pod \"dnsmasq-dns-7bf4c8dd6c-n8rxv\" (UID: \"3c7f9325-f498-4cd8-96fe-ec9ec9f57a83\") " pod="openstack/dnsmasq-dns-7bf4c8dd6c-n8rxv" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.505642 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c7f9325-f498-4cd8-96fe-ec9ec9f57a83-config\") pod \"dnsmasq-dns-7bf4c8dd6c-n8rxv\" (UID: \"3c7f9325-f498-4cd8-96fe-ec9ec9f57a83\") " pod="openstack/dnsmasq-dns-7bf4c8dd6c-n8rxv" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.509458 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3c7f9325-f498-4cd8-96fe-ec9ec9f57a83-ovsdbserver-nb\") pod \"dnsmasq-dns-7bf4c8dd6c-n8rxv\" (UID: \"3c7f9325-f498-4cd8-96fe-ec9ec9f57a83\") " pod="openstack/dnsmasq-dns-7bf4c8dd6c-n8rxv" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.510635 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3c7f9325-f498-4cd8-96fe-ec9ec9f57a83-ovsdbserver-sb\") pod \"dnsmasq-dns-7bf4c8dd6c-n8rxv\" (UID: \"3c7f9325-f498-4cd8-96fe-ec9ec9f57a83\") " pod="openstack/dnsmasq-dns-7bf4c8dd6c-n8rxv" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.512297 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3c7f9325-f498-4cd8-96fe-ec9ec9f57a83-dns-swift-storage-0\") pod \"dnsmasq-dns-7bf4c8dd6c-n8rxv\" (UID: \"3c7f9325-f498-4cd8-96fe-ec9ec9f57a83\") " pod="openstack/dnsmasq-dns-7bf4c8dd6c-n8rxv" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.512775 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3c7f9325-f498-4cd8-96fe-ec9ec9f57a83-dns-svc\") pod \"dnsmasq-dns-7bf4c8dd6c-n8rxv\" (UID: \"3c7f9325-f498-4cd8-96fe-ec9ec9f57a83\") " pod="openstack/dnsmasq-dns-7bf4c8dd6c-n8rxv" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.527328 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.542846 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-54dc864df6-phkw8" podStartSLOduration=3.599858309 podStartE2EDuration="18.542823027s" podCreationTimestamp="2025-11-25 15:21:18 +0000 UTC" firstStartedPulling="2025-11-25 15:21:19.766609783 +0000 UTC m=+1138.209072393" lastFinishedPulling="2025-11-25 15:21:34.709574501 +0000 UTC m=+1153.152037111" observedRunningTime="2025-11-25 15:21:36.476326511 +0000 UTC m=+1154.918789121" watchObservedRunningTime="2025-11-25 15:21:36.542823027 +0000 UTC m=+1154.985285637" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.590815 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4dhq\" (UniqueName: \"kubernetes.io/projected/3c7f9325-f498-4cd8-96fe-ec9ec9f57a83-kube-api-access-c4dhq\") pod \"dnsmasq-dns-7bf4c8dd6c-n8rxv\" (UID: \"3c7f9325-f498-4cd8-96fe-ec9ec9f57a83\") " pod="openstack/dnsmasq-dns-7bf4c8dd6c-n8rxv" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.602057 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bf4c8dd6c-n8rxv"] Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.602803 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bf4c8dd6c-n8rxv" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.605456 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd0c2b2d-c230-487c-ae4f-f231e973e1c3-scripts\") pod \"cinder-api-0\" (UID: \"bd0c2b2d-c230-487c-ae4f-f231e973e1c3\") " pod="openstack/cinder-api-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.605508 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd0c2b2d-c230-487c-ae4f-f231e973e1c3-config-data\") pod \"cinder-api-0\" (UID: \"bd0c2b2d-c230-487c-ae4f-f231e973e1c3\") " pod="openstack/cinder-api-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.605622 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd0c2b2d-c230-487c-ae4f-f231e973e1c3-logs\") pod \"cinder-api-0\" (UID: \"bd0c2b2d-c230-487c-ae4f-f231e973e1c3\") " pod="openstack/cinder-api-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.605662 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd0c2b2d-c230-487c-ae4f-f231e973e1c3-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"bd0c2b2d-c230-487c-ae4f-f231e973e1c3\") " pod="openstack/cinder-api-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.605689 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6w2g7\" (UniqueName: \"kubernetes.io/projected/bd0c2b2d-c230-487c-ae4f-f231e973e1c3-kube-api-access-6w2g7\") pod \"cinder-api-0\" (UID: \"bd0c2b2d-c230-487c-ae4f-f231e973e1c3\") " pod="openstack/cinder-api-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.605705 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bd0c2b2d-c230-487c-ae4f-f231e973e1c3-config-data-custom\") pod \"cinder-api-0\" (UID: \"bd0c2b2d-c230-487c-ae4f-f231e973e1c3\") " pod="openstack/cinder-api-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.605745 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bd0c2b2d-c230-487c-ae4f-f231e973e1c3-etc-machine-id\") pod \"cinder-api-0\" (UID: \"bd0c2b2d-c230-487c-ae4f-f231e973e1c3\") " pod="openstack/cinder-api-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.613902 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-77c9dbb88c-b285m" podStartSLOduration=3.232268008 podStartE2EDuration="18.613879737s" podCreationTimestamp="2025-11-25 15:21:18 +0000 UTC" firstStartedPulling="2025-11-25 15:21:19.255939367 +0000 UTC m=+1137.698401977" lastFinishedPulling="2025-11-25 15:21:34.637551096 +0000 UTC m=+1153.080013706" observedRunningTime="2025-11-25 15:21:36.507096137 +0000 UTC m=+1154.949558747" watchObservedRunningTime="2025-11-25 15:21:36.613879737 +0000 UTC m=+1155.056342347" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.654611 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.659210 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-795f4db4bc-xwdbs"] Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.660822 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-795f4db4bc-xwdbs" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.707190 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd0c2b2d-c230-487c-ae4f-f231e973e1c3-scripts\") pod \"cinder-api-0\" (UID: \"bd0c2b2d-c230-487c-ae4f-f231e973e1c3\") " pod="openstack/cinder-api-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.707239 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd0c2b2d-c230-487c-ae4f-f231e973e1c3-config-data\") pod \"cinder-api-0\" (UID: \"bd0c2b2d-c230-487c-ae4f-f231e973e1c3\") " pod="openstack/cinder-api-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.707264 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3dab95a0-f8d1-43c0-bd81-d3cafe943a16-ovsdbserver-nb\") pod \"dnsmasq-dns-795f4db4bc-xwdbs\" (UID: \"3dab95a0-f8d1-43c0-bd81-d3cafe943a16\") " pod="openstack/dnsmasq-dns-795f4db4bc-xwdbs" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.707289 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3dab95a0-f8d1-43c0-bd81-d3cafe943a16-config\") pod \"dnsmasq-dns-795f4db4bc-xwdbs\" (UID: \"3dab95a0-f8d1-43c0-bd81-d3cafe943a16\") " pod="openstack/dnsmasq-dns-795f4db4bc-xwdbs" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.707327 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3dab95a0-f8d1-43c0-bd81-d3cafe943a16-ovsdbserver-sb\") pod \"dnsmasq-dns-795f4db4bc-xwdbs\" (UID: \"3dab95a0-f8d1-43c0-bd81-d3cafe943a16\") " pod="openstack/dnsmasq-dns-795f4db4bc-xwdbs" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.707347 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s59nt\" (UniqueName: \"kubernetes.io/projected/3dab95a0-f8d1-43c0-bd81-d3cafe943a16-kube-api-access-s59nt\") pod \"dnsmasq-dns-795f4db4bc-xwdbs\" (UID: \"3dab95a0-f8d1-43c0-bd81-d3cafe943a16\") " pod="openstack/dnsmasq-dns-795f4db4bc-xwdbs" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.707395 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd0c2b2d-c230-487c-ae4f-f231e973e1c3-logs\") pod \"cinder-api-0\" (UID: \"bd0c2b2d-c230-487c-ae4f-f231e973e1c3\") " pod="openstack/cinder-api-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.707415 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3dab95a0-f8d1-43c0-bd81-d3cafe943a16-dns-swift-storage-0\") pod \"dnsmasq-dns-795f4db4bc-xwdbs\" (UID: \"3dab95a0-f8d1-43c0-bd81-d3cafe943a16\") " pod="openstack/dnsmasq-dns-795f4db4bc-xwdbs" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.707439 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3dab95a0-f8d1-43c0-bd81-d3cafe943a16-dns-svc\") pod \"dnsmasq-dns-795f4db4bc-xwdbs\" (UID: \"3dab95a0-f8d1-43c0-bd81-d3cafe943a16\") " pod="openstack/dnsmasq-dns-795f4db4bc-xwdbs" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.707469 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd0c2b2d-c230-487c-ae4f-f231e973e1c3-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"bd0c2b2d-c230-487c-ae4f-f231e973e1c3\") " pod="openstack/cinder-api-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.707496 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6w2g7\" (UniqueName: \"kubernetes.io/projected/bd0c2b2d-c230-487c-ae4f-f231e973e1c3-kube-api-access-6w2g7\") pod \"cinder-api-0\" (UID: \"bd0c2b2d-c230-487c-ae4f-f231e973e1c3\") " pod="openstack/cinder-api-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.707515 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bd0c2b2d-c230-487c-ae4f-f231e973e1c3-config-data-custom\") pod \"cinder-api-0\" (UID: \"bd0c2b2d-c230-487c-ae4f-f231e973e1c3\") " pod="openstack/cinder-api-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.707546 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bd0c2b2d-c230-487c-ae4f-f231e973e1c3-etc-machine-id\") pod \"cinder-api-0\" (UID: \"bd0c2b2d-c230-487c-ae4f-f231e973e1c3\") " pod="openstack/cinder-api-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.707638 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bd0c2b2d-c230-487c-ae4f-f231e973e1c3-etc-machine-id\") pod \"cinder-api-0\" (UID: \"bd0c2b2d-c230-487c-ae4f-f231e973e1c3\") " pod="openstack/cinder-api-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.710282 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd0c2b2d-c230-487c-ae4f-f231e973e1c3-logs\") pod \"cinder-api-0\" (UID: \"bd0c2b2d-c230-487c-ae4f-f231e973e1c3\") " pod="openstack/cinder-api-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.714903 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-795f4db4bc-xwdbs"] Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.727700 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd0c2b2d-c230-487c-ae4f-f231e973e1c3-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"bd0c2b2d-c230-487c-ae4f-f231e973e1c3\") " pod="openstack/cinder-api-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.737655 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd0c2b2d-c230-487c-ae4f-f231e973e1c3-config-data\") pod \"cinder-api-0\" (UID: \"bd0c2b2d-c230-487c-ae4f-f231e973e1c3\") " pod="openstack/cinder-api-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.737738 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd0c2b2d-c230-487c-ae4f-f231e973e1c3-scripts\") pod \"cinder-api-0\" (UID: \"bd0c2b2d-c230-487c-ae4f-f231e973e1c3\") " pod="openstack/cinder-api-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.740708 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bd0c2b2d-c230-487c-ae4f-f231e973e1c3-config-data-custom\") pod \"cinder-api-0\" (UID: \"bd0c2b2d-c230-487c-ae4f-f231e973e1c3\") " pod="openstack/cinder-api-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.744818 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6w2g7\" (UniqueName: \"kubernetes.io/projected/bd0c2b2d-c230-487c-ae4f-f231e973e1c3-kube-api-access-6w2g7\") pod \"cinder-api-0\" (UID: \"bd0c2b2d-c230-487c-ae4f-f231e973e1c3\") " pod="openstack/cinder-api-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.752324 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7d4dcdcc94-g9drz" podStartSLOduration=15.752298024 podStartE2EDuration="15.752298024s" podCreationTimestamp="2025-11-25 15:21:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:21:36.583636515 +0000 UTC m=+1155.026099145" watchObservedRunningTime="2025-11-25 15:21:36.752298024 +0000 UTC m=+1155.194760644" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.822179 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3dab95a0-f8d1-43c0-bd81-d3cafe943a16-dns-swift-storage-0\") pod \"dnsmasq-dns-795f4db4bc-xwdbs\" (UID: \"3dab95a0-f8d1-43c0-bd81-d3cafe943a16\") " pod="openstack/dnsmasq-dns-795f4db4bc-xwdbs" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.822223 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3dab95a0-f8d1-43c0-bd81-d3cafe943a16-dns-svc\") pod \"dnsmasq-dns-795f4db4bc-xwdbs\" (UID: \"3dab95a0-f8d1-43c0-bd81-d3cafe943a16\") " pod="openstack/dnsmasq-dns-795f4db4bc-xwdbs" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.822298 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3dab95a0-f8d1-43c0-bd81-d3cafe943a16-ovsdbserver-nb\") pod \"dnsmasq-dns-795f4db4bc-xwdbs\" (UID: \"3dab95a0-f8d1-43c0-bd81-d3cafe943a16\") " pod="openstack/dnsmasq-dns-795f4db4bc-xwdbs" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.822317 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3dab95a0-f8d1-43c0-bd81-d3cafe943a16-config\") pod \"dnsmasq-dns-795f4db4bc-xwdbs\" (UID: \"3dab95a0-f8d1-43c0-bd81-d3cafe943a16\") " pod="openstack/dnsmasq-dns-795f4db4bc-xwdbs" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.822347 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3dab95a0-f8d1-43c0-bd81-d3cafe943a16-ovsdbserver-sb\") pod \"dnsmasq-dns-795f4db4bc-xwdbs\" (UID: \"3dab95a0-f8d1-43c0-bd81-d3cafe943a16\") " pod="openstack/dnsmasq-dns-795f4db4bc-xwdbs" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.822367 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s59nt\" (UniqueName: \"kubernetes.io/projected/3dab95a0-f8d1-43c0-bd81-d3cafe943a16-kube-api-access-s59nt\") pod \"dnsmasq-dns-795f4db4bc-xwdbs\" (UID: \"3dab95a0-f8d1-43c0-bd81-d3cafe943a16\") " pod="openstack/dnsmasq-dns-795f4db4bc-xwdbs" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.823982 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3dab95a0-f8d1-43c0-bd81-d3cafe943a16-dns-swift-storage-0\") pod \"dnsmasq-dns-795f4db4bc-xwdbs\" (UID: \"3dab95a0-f8d1-43c0-bd81-d3cafe943a16\") " pod="openstack/dnsmasq-dns-795f4db4bc-xwdbs" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.824511 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3dab95a0-f8d1-43c0-bd81-d3cafe943a16-config\") pod \"dnsmasq-dns-795f4db4bc-xwdbs\" (UID: \"3dab95a0-f8d1-43c0-bd81-d3cafe943a16\") " pod="openstack/dnsmasq-dns-795f4db4bc-xwdbs" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.825009 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3dab95a0-f8d1-43c0-bd81-d3cafe943a16-ovsdbserver-sb\") pod \"dnsmasq-dns-795f4db4bc-xwdbs\" (UID: \"3dab95a0-f8d1-43c0-bd81-d3cafe943a16\") " pod="openstack/dnsmasq-dns-795f4db4bc-xwdbs" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.833608 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.838971 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3dab95a0-f8d1-43c0-bd81-d3cafe943a16-dns-svc\") pod \"dnsmasq-dns-795f4db4bc-xwdbs\" (UID: \"3dab95a0-f8d1-43c0-bd81-d3cafe943a16\") " pod="openstack/dnsmasq-dns-795f4db4bc-xwdbs" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.840549 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3dab95a0-f8d1-43c0-bd81-d3cafe943a16-ovsdbserver-nb\") pod \"dnsmasq-dns-795f4db4bc-xwdbs\" (UID: \"3dab95a0-f8d1-43c0-bd81-d3cafe943a16\") " pod="openstack/dnsmasq-dns-795f4db4bc-xwdbs" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.893323 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s59nt\" (UniqueName: \"kubernetes.io/projected/3dab95a0-f8d1-43c0-bd81-d3cafe943a16-kube-api-access-s59nt\") pod \"dnsmasq-dns-795f4db4bc-xwdbs\" (UID: \"3dab95a0-f8d1-43c0-bd81-d3cafe943a16\") " pod="openstack/dnsmasq-dns-795f4db4bc-xwdbs" Nov 25 15:21:36 crc kubenswrapper[4890]: I1125 15:21:36.914108 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-795f4db4bc-xwdbs" Nov 25 15:21:37 crc kubenswrapper[4890]: I1125 15:21:37.333849 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:21:37 crc kubenswrapper[4890]: I1125 15:21:37.390279 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 15:21:37 crc kubenswrapper[4890]: I1125 15:21:37.395765 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 15:21:37 crc kubenswrapper[4890]: I1125 15:21:37.411928 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 15:21:37 crc kubenswrapper[4890]: I1125 15:21:37.416787 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 25 15:21:37 crc kubenswrapper[4890]: I1125 15:21:37.417282 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 25 15:21:37 crc kubenswrapper[4890]: I1125 15:21:37.417617 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-qjthd" Nov 25 15:21:37 crc kubenswrapper[4890]: I1125 15:21:37.484555 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"47f833ef-ad6e-4a03-a1f0-9507d2d1f11e\") " pod="openstack/glance-default-external-api-0" Nov 25 15:21:37 crc kubenswrapper[4890]: I1125 15:21:37.484611 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhvw9\" (UniqueName: \"kubernetes.io/projected/47f833ef-ad6e-4a03-a1f0-9507d2d1f11e-kube-api-access-rhvw9\") pod \"glance-default-external-api-0\" (UID: \"47f833ef-ad6e-4a03-a1f0-9507d2d1f11e\") " pod="openstack/glance-default-external-api-0" Nov 25 15:21:37 crc kubenswrapper[4890]: I1125 15:21:37.484629 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47f833ef-ad6e-4a03-a1f0-9507d2d1f11e-scripts\") pod \"glance-default-external-api-0\" (UID: \"47f833ef-ad6e-4a03-a1f0-9507d2d1f11e\") " pod="openstack/glance-default-external-api-0" Nov 25 15:21:37 crc kubenswrapper[4890]: I1125 15:21:37.484658 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47f833ef-ad6e-4a03-a1f0-9507d2d1f11e-config-data\") pod \"glance-default-external-api-0\" (UID: \"47f833ef-ad6e-4a03-a1f0-9507d2d1f11e\") " pod="openstack/glance-default-external-api-0" Nov 25 15:21:37 crc kubenswrapper[4890]: I1125 15:21:37.484687 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47f833ef-ad6e-4a03-a1f0-9507d2d1f11e-logs\") pod \"glance-default-external-api-0\" (UID: \"47f833ef-ad6e-4a03-a1f0-9507d2d1f11e\") " pod="openstack/glance-default-external-api-0" Nov 25 15:21:37 crc kubenswrapper[4890]: I1125 15:21:37.484718 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/47f833ef-ad6e-4a03-a1f0-9507d2d1f11e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"47f833ef-ad6e-4a03-a1f0-9507d2d1f11e\") " pod="openstack/glance-default-external-api-0" Nov 25 15:21:37 crc kubenswrapper[4890]: I1125 15:21:37.484734 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47f833ef-ad6e-4a03-a1f0-9507d2d1f11e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"47f833ef-ad6e-4a03-a1f0-9507d2d1f11e\") " pod="openstack/glance-default-external-api-0" Nov 25 15:21:37 crc kubenswrapper[4890]: I1125 15:21:37.491215 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bf4c8dd6c-n8rxv"] Nov 25 15:21:37 crc kubenswrapper[4890]: I1125 15:21:37.544240 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5f45cd8b8f-8wpds" event={"ID":"66cfb07e-ac1a-4aee-8fbd-7871e2e30d88","Type":"ContainerStarted","Data":"848687404e769c718e8d947ebdb155e77bea9bb351cbdc9c10e57d0a137a435d"} Nov 25 15:21:37 crc kubenswrapper[4890]: I1125 15:21:37.547310 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-5f45cd8b8f-8wpds" Nov 25 15:21:37 crc kubenswrapper[4890]: I1125 15:21:37.547412 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-5f45cd8b8f-8wpds" Nov 25 15:21:37 crc kubenswrapper[4890]: I1125 15:21:37.560544 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bf4c8dd6c-n8rxv" event={"ID":"3c7f9325-f498-4cd8-96fe-ec9ec9f57a83","Type":"ContainerStarted","Data":"d01457c8aa348839219c9e294dd50342f2687b4f2171612921e5dad7f1bee06c"} Nov 25 15:21:37 crc kubenswrapper[4890]: I1125 15:21:37.575325 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f1ae99b8-5004-434a-812d-0ab8043bd031","Type":"ContainerStarted","Data":"ccc2122b33c3cad963740d47b2d74385bd34b4e270f6fae8f66ebc5de53f3a3a"} Nov 25 15:21:37 crc kubenswrapper[4890]: I1125 15:21:37.579648 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-5f45cd8b8f-8wpds" podStartSLOduration=15.579622628 podStartE2EDuration="15.579622628s" podCreationTimestamp="2025-11-25 15:21:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:21:37.574306594 +0000 UTC m=+1156.016769214" watchObservedRunningTime="2025-11-25 15:21:37.579622628 +0000 UTC m=+1156.022085238" Nov 25 15:21:37 crc kubenswrapper[4890]: I1125 15:21:37.589078 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"47f833ef-ad6e-4a03-a1f0-9507d2d1f11e\") " pod="openstack/glance-default-external-api-0" Nov 25 15:21:37 crc kubenswrapper[4890]: I1125 15:21:37.589249 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhvw9\" (UniqueName: \"kubernetes.io/projected/47f833ef-ad6e-4a03-a1f0-9507d2d1f11e-kube-api-access-rhvw9\") pod \"glance-default-external-api-0\" (UID: \"47f833ef-ad6e-4a03-a1f0-9507d2d1f11e\") " pod="openstack/glance-default-external-api-0" Nov 25 15:21:37 crc kubenswrapper[4890]: I1125 15:21:37.589277 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47f833ef-ad6e-4a03-a1f0-9507d2d1f11e-scripts\") pod \"glance-default-external-api-0\" (UID: \"47f833ef-ad6e-4a03-a1f0-9507d2d1f11e\") " pod="openstack/glance-default-external-api-0" Nov 25 15:21:37 crc kubenswrapper[4890]: I1125 15:21:37.589332 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47f833ef-ad6e-4a03-a1f0-9507d2d1f11e-config-data\") pod \"glance-default-external-api-0\" (UID: \"47f833ef-ad6e-4a03-a1f0-9507d2d1f11e\") " pod="openstack/glance-default-external-api-0" Nov 25 15:21:37 crc kubenswrapper[4890]: I1125 15:21:37.589368 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47f833ef-ad6e-4a03-a1f0-9507d2d1f11e-logs\") pod \"glance-default-external-api-0\" (UID: \"47f833ef-ad6e-4a03-a1f0-9507d2d1f11e\") " pod="openstack/glance-default-external-api-0" Nov 25 15:21:37 crc kubenswrapper[4890]: I1125 15:21:37.589415 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/47f833ef-ad6e-4a03-a1f0-9507d2d1f11e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"47f833ef-ad6e-4a03-a1f0-9507d2d1f11e\") " pod="openstack/glance-default-external-api-0" Nov 25 15:21:37 crc kubenswrapper[4890]: I1125 15:21:37.589439 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47f833ef-ad6e-4a03-a1f0-9507d2d1f11e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"47f833ef-ad6e-4a03-a1f0-9507d2d1f11e\") " pod="openstack/glance-default-external-api-0" Nov 25 15:21:37 crc kubenswrapper[4890]: I1125 15:21:37.590754 4890 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"47f833ef-ad6e-4a03-a1f0-9507d2d1f11e\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-external-api-0" Nov 25 15:21:37 crc kubenswrapper[4890]: I1125 15:21:37.594125 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47f833ef-ad6e-4a03-a1f0-9507d2d1f11e-logs\") pod \"glance-default-external-api-0\" (UID: \"47f833ef-ad6e-4a03-a1f0-9507d2d1f11e\") " pod="openstack/glance-default-external-api-0" Nov 25 15:21:37 crc kubenswrapper[4890]: I1125 15:21:37.594374 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/47f833ef-ad6e-4a03-a1f0-9507d2d1f11e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"47f833ef-ad6e-4a03-a1f0-9507d2d1f11e\") " pod="openstack/glance-default-external-api-0" Nov 25 15:21:37 crc kubenswrapper[4890]: I1125 15:21:37.614805 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47f833ef-ad6e-4a03-a1f0-9507d2d1f11e-config-data\") pod \"glance-default-external-api-0\" (UID: \"47f833ef-ad6e-4a03-a1f0-9507d2d1f11e\") " pod="openstack/glance-default-external-api-0" Nov 25 15:21:37 crc kubenswrapper[4890]: I1125 15:21:37.648388 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47f833ef-ad6e-4a03-a1f0-9507d2d1f11e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"47f833ef-ad6e-4a03-a1f0-9507d2d1f11e\") " pod="openstack/glance-default-external-api-0" Nov 25 15:21:37 crc kubenswrapper[4890]: I1125 15:21:37.671364 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhvw9\" (UniqueName: \"kubernetes.io/projected/47f833ef-ad6e-4a03-a1f0-9507d2d1f11e-kube-api-access-rhvw9\") pod \"glance-default-external-api-0\" (UID: \"47f833ef-ad6e-4a03-a1f0-9507d2d1f11e\") " pod="openstack/glance-default-external-api-0" Nov 25 15:21:37 crc kubenswrapper[4890]: I1125 15:21:37.704853 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 15:21:37 crc kubenswrapper[4890]: I1125 15:21:37.713846 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 15:21:37 crc kubenswrapper[4890]: I1125 15:21:37.716230 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 25 15:21:37 crc kubenswrapper[4890]: I1125 15:21:37.717938 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47f833ef-ad6e-4a03-a1f0-9507d2d1f11e-scripts\") pod \"glance-default-external-api-0\" (UID: \"47f833ef-ad6e-4a03-a1f0-9507d2d1f11e\") " pod="openstack/glance-default-external-api-0" Nov 25 15:21:37 crc kubenswrapper[4890]: I1125 15:21:37.746344 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 15:21:37 crc kubenswrapper[4890]: I1125 15:21:37.755467 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"47f833ef-ad6e-4a03-a1f0-9507d2d1f11e\") " pod="openstack/glance-default-external-api-0" Nov 25 15:21:37 crc kubenswrapper[4890]: I1125 15:21:37.779504 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 15:21:37 crc kubenswrapper[4890]: I1125 15:21:37.797077 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 15:21:37 crc kubenswrapper[4890]: I1125 15:21:37.902572 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4d53cba-d83e-4b1f-afe5-b6f21b939fb0-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b4d53cba-d83e-4b1f-afe5-b6f21b939fb0\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:37 crc kubenswrapper[4890]: I1125 15:21:37.903076 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"b4d53cba-d83e-4b1f-afe5-b6f21b939fb0\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:37 crc kubenswrapper[4890]: I1125 15:21:37.903099 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cwm2\" (UniqueName: \"kubernetes.io/projected/b4d53cba-d83e-4b1f-afe5-b6f21b939fb0-kube-api-access-8cwm2\") pod \"glance-default-internal-api-0\" (UID: \"b4d53cba-d83e-4b1f-afe5-b6f21b939fb0\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:37 crc kubenswrapper[4890]: I1125 15:21:37.903145 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b4d53cba-d83e-4b1f-afe5-b6f21b939fb0-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b4d53cba-d83e-4b1f-afe5-b6f21b939fb0\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:37 crc kubenswrapper[4890]: I1125 15:21:37.903196 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4d53cba-d83e-4b1f-afe5-b6f21b939fb0-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b4d53cba-d83e-4b1f-afe5-b6f21b939fb0\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:37 crc kubenswrapper[4890]: I1125 15:21:37.903256 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b4d53cba-d83e-4b1f-afe5-b6f21b939fb0-logs\") pod \"glance-default-internal-api-0\" (UID: \"b4d53cba-d83e-4b1f-afe5-b6f21b939fb0\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:37 crc kubenswrapper[4890]: I1125 15:21:37.903308 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4d53cba-d83e-4b1f-afe5-b6f21b939fb0-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b4d53cba-d83e-4b1f-afe5-b6f21b939fb0\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:37 crc kubenswrapper[4890]: I1125 15:21:37.984035 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-795f4db4bc-xwdbs"] Nov 25 15:21:38 crc kubenswrapper[4890]: I1125 15:21:38.004346 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b4d53cba-d83e-4b1f-afe5-b6f21b939fb0-logs\") pod \"glance-default-internal-api-0\" (UID: \"b4d53cba-d83e-4b1f-afe5-b6f21b939fb0\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:38 crc kubenswrapper[4890]: I1125 15:21:38.004392 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4d53cba-d83e-4b1f-afe5-b6f21b939fb0-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b4d53cba-d83e-4b1f-afe5-b6f21b939fb0\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:38 crc kubenswrapper[4890]: I1125 15:21:38.004440 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4d53cba-d83e-4b1f-afe5-b6f21b939fb0-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b4d53cba-d83e-4b1f-afe5-b6f21b939fb0\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:38 crc kubenswrapper[4890]: I1125 15:21:38.004498 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"b4d53cba-d83e-4b1f-afe5-b6f21b939fb0\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:38 crc kubenswrapper[4890]: I1125 15:21:38.004517 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cwm2\" (UniqueName: \"kubernetes.io/projected/b4d53cba-d83e-4b1f-afe5-b6f21b939fb0-kube-api-access-8cwm2\") pod \"glance-default-internal-api-0\" (UID: \"b4d53cba-d83e-4b1f-afe5-b6f21b939fb0\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:38 crc kubenswrapper[4890]: I1125 15:21:38.004548 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b4d53cba-d83e-4b1f-afe5-b6f21b939fb0-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b4d53cba-d83e-4b1f-afe5-b6f21b939fb0\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:38 crc kubenswrapper[4890]: I1125 15:21:38.004571 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4d53cba-d83e-4b1f-afe5-b6f21b939fb0-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b4d53cba-d83e-4b1f-afe5-b6f21b939fb0\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:38 crc kubenswrapper[4890]: I1125 15:21:38.006558 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b4d53cba-d83e-4b1f-afe5-b6f21b939fb0-logs\") pod \"glance-default-internal-api-0\" (UID: \"b4d53cba-d83e-4b1f-afe5-b6f21b939fb0\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:38 crc kubenswrapper[4890]: I1125 15:21:38.007028 4890 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"b4d53cba-d83e-4b1f-afe5-b6f21b939fb0\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-internal-api-0" Nov 25 15:21:38 crc kubenswrapper[4890]: I1125 15:21:38.011138 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b4d53cba-d83e-4b1f-afe5-b6f21b939fb0-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b4d53cba-d83e-4b1f-afe5-b6f21b939fb0\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:38 crc kubenswrapper[4890]: I1125 15:21:38.013769 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4d53cba-d83e-4b1f-afe5-b6f21b939fb0-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b4d53cba-d83e-4b1f-afe5-b6f21b939fb0\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:38 crc kubenswrapper[4890]: I1125 15:21:38.015484 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4d53cba-d83e-4b1f-afe5-b6f21b939fb0-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b4d53cba-d83e-4b1f-afe5-b6f21b939fb0\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:38 crc kubenswrapper[4890]: I1125 15:21:38.023052 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cwm2\" (UniqueName: \"kubernetes.io/projected/b4d53cba-d83e-4b1f-afe5-b6f21b939fb0-kube-api-access-8cwm2\") pod \"glance-default-internal-api-0\" (UID: \"b4d53cba-d83e-4b1f-afe5-b6f21b939fb0\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:38 crc kubenswrapper[4890]: I1125 15:21:38.026044 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4d53cba-d83e-4b1f-afe5-b6f21b939fb0-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b4d53cba-d83e-4b1f-afe5-b6f21b939fb0\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:38 crc kubenswrapper[4890]: I1125 15:21:38.042941 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 25 15:21:38 crc kubenswrapper[4890]: I1125 15:21:38.074147 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"b4d53cba-d83e-4b1f-afe5-b6f21b939fb0\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:38 crc kubenswrapper[4890]: I1125 15:21:38.198760 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 15:21:38 crc kubenswrapper[4890]: I1125 15:21:38.464727 4890 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-556858788d-s2ppl" podUID="1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.145:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.145:8443: connect: connection refused" Nov 25 15:21:38 crc kubenswrapper[4890]: I1125 15:21:38.465184 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-556858788d-s2ppl" Nov 25 15:21:38 crc kubenswrapper[4890]: I1125 15:21:38.512274 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 15:21:38 crc kubenswrapper[4890]: I1125 15:21:38.627939 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d65af5aa-095d-465f-8fd7-ba76b4ec08af","Type":"ContainerStarted","Data":"87a8b8c8de7fa3af02a2134b3a7e4fbc1de8ac4f257df8e52aad1d9f76d366a4"} Nov 25 15:21:38 crc kubenswrapper[4890]: I1125 15:21:38.642568 4890 generic.go:334] "Generic (PLEG): container finished" podID="3c7f9325-f498-4cd8-96fe-ec9ec9f57a83" containerID="bef223d3d7137d787686f9d444621cd523767d6b09121fd40231404064e3556c" exitCode=0 Nov 25 15:21:38 crc kubenswrapper[4890]: I1125 15:21:38.642650 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bf4c8dd6c-n8rxv" event={"ID":"3c7f9325-f498-4cd8-96fe-ec9ec9f57a83","Type":"ContainerDied","Data":"bef223d3d7137d787686f9d444621cd523767d6b09121fd40231404064e3556c"} Nov 25 15:21:38 crc kubenswrapper[4890]: I1125 15:21:38.683018 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"bd0c2b2d-c230-487c-ae4f-f231e973e1c3","Type":"ContainerStarted","Data":"77baf01d5d4a5ee3d68e83d21894d2fa1f336e9768fcd79e2b239493021bf205"} Nov 25 15:21:38 crc kubenswrapper[4890]: I1125 15:21:38.689307 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"47f833ef-ad6e-4a03-a1f0-9507d2d1f11e","Type":"ContainerStarted","Data":"c16d0c26b1e388fe63a5e78a81af8a7c8d2c6df38dbaa7a1788c51fe961f6235"} Nov 25 15:21:38 crc kubenswrapper[4890]: I1125 15:21:38.701455 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-795f4db4bc-xwdbs" event={"ID":"3dab95a0-f8d1-43c0-bd81-d3cafe943a16","Type":"ContainerStarted","Data":"f5cced960e733f360ae35de6797c9523edb2727f632862fc3204eb6efd8bb28d"} Nov 25 15:21:39 crc kubenswrapper[4890]: I1125 15:21:39.056781 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 15:21:39 crc kubenswrapper[4890]: W1125 15:21:39.354777 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb4d53cba_d83e_4b1f_afe5_b6f21b939fb0.slice/crio-65d5a66735e95719ff9ff16e32207333a2724eb8b6d296fedb32492b0d2cd16c WatchSource:0}: Error finding container 65d5a66735e95719ff9ff16e32207333a2724eb8b6d296fedb32492b0d2cd16c: Status 404 returned error can't find the container with id 65d5a66735e95719ff9ff16e32207333a2724eb8b6d296fedb32492b0d2cd16c Nov 25 15:21:39 crc kubenswrapper[4890]: I1125 15:21:39.455432 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bf4c8dd6c-n8rxv" Nov 25 15:21:39 crc kubenswrapper[4890]: I1125 15:21:39.580999 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3c7f9325-f498-4cd8-96fe-ec9ec9f57a83-dns-svc\") pod \"3c7f9325-f498-4cd8-96fe-ec9ec9f57a83\" (UID: \"3c7f9325-f498-4cd8-96fe-ec9ec9f57a83\") " Nov 25 15:21:39 crc kubenswrapper[4890]: I1125 15:21:39.581305 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3c7f9325-f498-4cd8-96fe-ec9ec9f57a83-ovsdbserver-sb\") pod \"3c7f9325-f498-4cd8-96fe-ec9ec9f57a83\" (UID: \"3c7f9325-f498-4cd8-96fe-ec9ec9f57a83\") " Nov 25 15:21:39 crc kubenswrapper[4890]: I1125 15:21:39.581413 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3c7f9325-f498-4cd8-96fe-ec9ec9f57a83-ovsdbserver-nb\") pod \"3c7f9325-f498-4cd8-96fe-ec9ec9f57a83\" (UID: \"3c7f9325-f498-4cd8-96fe-ec9ec9f57a83\") " Nov 25 15:21:39 crc kubenswrapper[4890]: I1125 15:21:39.581677 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c7f9325-f498-4cd8-96fe-ec9ec9f57a83-config\") pod \"3c7f9325-f498-4cd8-96fe-ec9ec9f57a83\" (UID: \"3c7f9325-f498-4cd8-96fe-ec9ec9f57a83\") " Nov 25 15:21:39 crc kubenswrapper[4890]: I1125 15:21:39.581784 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3c7f9325-f498-4cd8-96fe-ec9ec9f57a83-dns-swift-storage-0\") pod \"3c7f9325-f498-4cd8-96fe-ec9ec9f57a83\" (UID: \"3c7f9325-f498-4cd8-96fe-ec9ec9f57a83\") " Nov 25 15:21:39 crc kubenswrapper[4890]: I1125 15:21:39.581993 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4dhq\" (UniqueName: \"kubernetes.io/projected/3c7f9325-f498-4cd8-96fe-ec9ec9f57a83-kube-api-access-c4dhq\") pod \"3c7f9325-f498-4cd8-96fe-ec9ec9f57a83\" (UID: \"3c7f9325-f498-4cd8-96fe-ec9ec9f57a83\") " Nov 25 15:21:39 crc kubenswrapper[4890]: I1125 15:21:39.615464 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c7f9325-f498-4cd8-96fe-ec9ec9f57a83-kube-api-access-c4dhq" (OuterVolumeSpecName: "kube-api-access-c4dhq") pod "3c7f9325-f498-4cd8-96fe-ec9ec9f57a83" (UID: "3c7f9325-f498-4cd8-96fe-ec9ec9f57a83"). InnerVolumeSpecName "kube-api-access-c4dhq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:21:39 crc kubenswrapper[4890]: I1125 15:21:39.622120 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c7f9325-f498-4cd8-96fe-ec9ec9f57a83-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "3c7f9325-f498-4cd8-96fe-ec9ec9f57a83" (UID: "3c7f9325-f498-4cd8-96fe-ec9ec9f57a83"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:21:39 crc kubenswrapper[4890]: I1125 15:21:39.637081 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c7f9325-f498-4cd8-96fe-ec9ec9f57a83-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3c7f9325-f498-4cd8-96fe-ec9ec9f57a83" (UID: "3c7f9325-f498-4cd8-96fe-ec9ec9f57a83"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:21:39 crc kubenswrapper[4890]: I1125 15:21:39.664084 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c7f9325-f498-4cd8-96fe-ec9ec9f57a83-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3c7f9325-f498-4cd8-96fe-ec9ec9f57a83" (UID: "3c7f9325-f498-4cd8-96fe-ec9ec9f57a83"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:21:39 crc kubenswrapper[4890]: I1125 15:21:39.664807 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c7f9325-f498-4cd8-96fe-ec9ec9f57a83-config" (OuterVolumeSpecName: "config") pod "3c7f9325-f498-4cd8-96fe-ec9ec9f57a83" (UID: "3c7f9325-f498-4cd8-96fe-ec9ec9f57a83"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:21:39 crc kubenswrapper[4890]: I1125 15:21:39.665431 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c7f9325-f498-4cd8-96fe-ec9ec9f57a83-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3c7f9325-f498-4cd8-96fe-ec9ec9f57a83" (UID: "3c7f9325-f498-4cd8-96fe-ec9ec9f57a83"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:21:39 crc kubenswrapper[4890]: I1125 15:21:39.684640 4890 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c7f9325-f498-4cd8-96fe-ec9ec9f57a83-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:39 crc kubenswrapper[4890]: I1125 15:21:39.684681 4890 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3c7f9325-f498-4cd8-96fe-ec9ec9f57a83-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:39 crc kubenswrapper[4890]: I1125 15:21:39.684696 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4dhq\" (UniqueName: \"kubernetes.io/projected/3c7f9325-f498-4cd8-96fe-ec9ec9f57a83-kube-api-access-c4dhq\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:39 crc kubenswrapper[4890]: I1125 15:21:39.684707 4890 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3c7f9325-f498-4cd8-96fe-ec9ec9f57a83-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:39 crc kubenswrapper[4890]: I1125 15:21:39.684718 4890 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3c7f9325-f498-4cd8-96fe-ec9ec9f57a83-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:39 crc kubenswrapper[4890]: I1125 15:21:39.684729 4890 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3c7f9325-f498-4cd8-96fe-ec9ec9f57a83-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:39 crc kubenswrapper[4890]: I1125 15:21:39.759271 4890 generic.go:334] "Generic (PLEG): container finished" podID="3dab95a0-f8d1-43c0-bd81-d3cafe943a16" containerID="25b66f459674dfda658c6f48abf96a0faaa14fa26d4b81871a7d4eba9d78f692" exitCode=0 Nov 25 15:21:39 crc kubenswrapper[4890]: I1125 15:21:39.759348 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-795f4db4bc-xwdbs" event={"ID":"3dab95a0-f8d1-43c0-bd81-d3cafe943a16","Type":"ContainerDied","Data":"25b66f459674dfda658c6f48abf96a0faaa14fa26d4b81871a7d4eba9d78f692"} Nov 25 15:21:39 crc kubenswrapper[4890]: I1125 15:21:39.770758 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bf4c8dd6c-n8rxv" event={"ID":"3c7f9325-f498-4cd8-96fe-ec9ec9f57a83","Type":"ContainerDied","Data":"d01457c8aa348839219c9e294dd50342f2687b4f2171612921e5dad7f1bee06c"} Nov 25 15:21:39 crc kubenswrapper[4890]: I1125 15:21:39.770811 4890 scope.go:117] "RemoveContainer" containerID="bef223d3d7137d787686f9d444621cd523767d6b09121fd40231404064e3556c" Nov 25 15:21:39 crc kubenswrapper[4890]: I1125 15:21:39.770958 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bf4c8dd6c-n8rxv" Nov 25 15:21:39 crc kubenswrapper[4890]: I1125 15:21:39.779319 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b4d53cba-d83e-4b1f-afe5-b6f21b939fb0","Type":"ContainerStarted","Data":"65d5a66735e95719ff9ff16e32207333a2724eb8b6d296fedb32492b0d2cd16c"} Nov 25 15:21:39 crc kubenswrapper[4890]: I1125 15:21:39.792125 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"bd0c2b2d-c230-487c-ae4f-f231e973e1c3","Type":"ContainerStarted","Data":"fc645d4225419eb44aac652080ca3a27fbcc00627a803e750a4c82f85299871a"} Nov 25 15:21:39 crc kubenswrapper[4890]: I1125 15:21:39.805984 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f1ae99b8-5004-434a-812d-0ab8043bd031","Type":"ContainerStarted","Data":"05260f40f216ef0403254bbafc9d90e1f11f05da2c4e948c656caf0e28f2129c"} Nov 25 15:21:39 crc kubenswrapper[4890]: I1125 15:21:39.933900 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bf4c8dd6c-n8rxv"] Nov 25 15:21:39 crc kubenswrapper[4890]: I1125 15:21:39.963607 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7bf4c8dd6c-n8rxv"] Nov 25 15:21:40 crc kubenswrapper[4890]: I1125 15:21:40.214822 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c7f9325-f498-4cd8-96fe-ec9ec9f57a83" path="/var/lib/kubelet/pods/3c7f9325-f498-4cd8-96fe-ec9ec9f57a83/volumes" Nov 25 15:21:40 crc kubenswrapper[4890]: I1125 15:21:40.838688 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b4d53cba-d83e-4b1f-afe5-b6f21b939fb0","Type":"ContainerStarted","Data":"1c7d653f81fa1e71828fdb1eaff2feeb554e88bf38c9e22c4f90fbfcde3f2659"} Nov 25 15:21:40 crc kubenswrapper[4890]: I1125 15:21:40.840536 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f1ae99b8-5004-434a-812d-0ab8043bd031","Type":"ContainerStarted","Data":"89212db8a499c692049b06aa414e256a32d28f2a42604f084d5c0376ffb54df2"} Nov 25 15:21:40 crc kubenswrapper[4890]: I1125 15:21:40.847660 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"47f833ef-ad6e-4a03-a1f0-9507d2d1f11e","Type":"ContainerStarted","Data":"25f9a26993bce3c5084a666d32531c3c295f8fe62fd7b42825348d6e24d2c230"} Nov 25 15:21:40 crc kubenswrapper[4890]: I1125 15:21:40.849811 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-795f4db4bc-xwdbs" event={"ID":"3dab95a0-f8d1-43c0-bd81-d3cafe943a16","Type":"ContainerStarted","Data":"89e68f8d06e257d883b87fe64dd99aadb2b80168680ee45faf22df84dc8e8bb3"} Nov 25 15:21:40 crc kubenswrapper[4890]: I1125 15:21:40.850328 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-795f4db4bc-xwdbs" Nov 25 15:21:40 crc kubenswrapper[4890]: I1125 15:21:40.889266 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-795f4db4bc-xwdbs" podStartSLOduration=4.8892469290000005 podStartE2EDuration="4.889246929s" podCreationTimestamp="2025-11-25 15:21:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:21:40.870250461 +0000 UTC m=+1159.312713071" watchObservedRunningTime="2025-11-25 15:21:40.889246929 +0000 UTC m=+1159.331709539" Nov 25 15:21:41 crc kubenswrapper[4890]: I1125 15:21:41.482035 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 15:21:41 crc kubenswrapper[4890]: I1125 15:21:41.764889 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 15:21:41 crc kubenswrapper[4890]: I1125 15:21:41.939506 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f1ae99b8-5004-434a-812d-0ab8043bd031","Type":"ContainerStarted","Data":"ed409c3fa7896d5038584fa75afca3e6decc87f9db9f82aa2e9515bd2784a68f"} Nov 25 15:21:41 crc kubenswrapper[4890]: I1125 15:21:41.973555 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"47f833ef-ad6e-4a03-a1f0-9507d2d1f11e","Type":"ContainerStarted","Data":"5acb64d6ccbff760c46190787766719f1c75efd8e9c67af19facdafca534d0e7"} Nov 25 15:21:42 crc kubenswrapper[4890]: I1125 15:21:42.001316 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d65af5aa-095d-465f-8fd7-ba76b4ec08af","Type":"ContainerStarted","Data":"bd60a64ae0e4a9adde0fbb49c6acc7c6ce9c4b28a194e92a9791eacf02e81b52"} Nov 25 15:21:42 crc kubenswrapper[4890]: I1125 15:21:42.004110 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b4d53cba-d83e-4b1f-afe5-b6f21b939fb0","Type":"ContainerStarted","Data":"f387b29da54c75ad0c6513c82dbb7b3fa2bceb03bcf5e8cf39deb409c2f611a9"} Nov 25 15:21:42 crc kubenswrapper[4890]: I1125 15:21:42.004297 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="b4d53cba-d83e-4b1f-afe5-b6f21b939fb0" containerName="glance-log" containerID="cri-o://1c7d653f81fa1e71828fdb1eaff2feeb554e88bf38c9e22c4f90fbfcde3f2659" gracePeriod=30 Nov 25 15:21:42 crc kubenswrapper[4890]: I1125 15:21:42.004747 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="b4d53cba-d83e-4b1f-afe5-b6f21b939fb0" containerName="glance-httpd" containerID="cri-o://f387b29da54c75ad0c6513c82dbb7b3fa2bceb03bcf5e8cf39deb409c2f611a9" gracePeriod=30 Nov 25 15:21:42 crc kubenswrapper[4890]: I1125 15:21:42.007723 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.007705068 podStartE2EDuration="6.007705068s" podCreationTimestamp="2025-11-25 15:21:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:21:42.002061286 +0000 UTC m=+1160.444523906" watchObservedRunningTime="2025-11-25 15:21:42.007705068 +0000 UTC m=+1160.450167678" Nov 25 15:21:42 crc kubenswrapper[4890]: I1125 15:21:42.016899 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"bd0c2b2d-c230-487c-ae4f-f231e973e1c3","Type":"ContainerStarted","Data":"13f2e42ea2f12408dbaed3897ff7cee43f4e235b1ab146c7f90fa9f8b7efefd0"} Nov 25 15:21:42 crc kubenswrapper[4890]: I1125 15:21:42.016933 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 25 15:21:42 crc kubenswrapper[4890]: I1125 15:21:42.035026 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.035009016 podStartE2EDuration="6.035009016s" podCreationTimestamp="2025-11-25 15:21:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:21:42.033661722 +0000 UTC m=+1160.476124332" watchObservedRunningTime="2025-11-25 15:21:42.035009016 +0000 UTC m=+1160.477471626" Nov 25 15:21:42 crc kubenswrapper[4890]: I1125 15:21:42.068367 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=6.068335525 podStartE2EDuration="6.068335525s" podCreationTimestamp="2025-11-25 15:21:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:21:42.058431856 +0000 UTC m=+1160.500894466" watchObservedRunningTime="2025-11-25 15:21:42.068335525 +0000 UTC m=+1160.510798135" Nov 25 15:21:42 crc kubenswrapper[4890]: I1125 15:21:42.485036 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-74c777b4cb-m7k57" Nov 25 15:21:42 crc kubenswrapper[4890]: I1125 15:21:42.540665 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-74c777b4cb-m7k57" Nov 25 15:21:42 crc kubenswrapper[4890]: I1125 15:21:42.600205 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-5f45cd8b8f-8wpds" Nov 25 15:21:42 crc kubenswrapper[4890]: I1125 15:21:42.614522 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-5f45cd8b8f-8wpds" Nov 25 15:21:42 crc kubenswrapper[4890]: I1125 15:21:42.875935 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 15:21:42 crc kubenswrapper[4890]: I1125 15:21:42.895833 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"b4d53cba-d83e-4b1f-afe5-b6f21b939fb0\" (UID: \"b4d53cba-d83e-4b1f-afe5-b6f21b939fb0\") " Nov 25 15:21:42 crc kubenswrapper[4890]: I1125 15:21:42.895932 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b4d53cba-d83e-4b1f-afe5-b6f21b939fb0-httpd-run\") pod \"b4d53cba-d83e-4b1f-afe5-b6f21b939fb0\" (UID: \"b4d53cba-d83e-4b1f-afe5-b6f21b939fb0\") " Nov 25 15:21:42 crc kubenswrapper[4890]: I1125 15:21:42.895976 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4d53cba-d83e-4b1f-afe5-b6f21b939fb0-combined-ca-bundle\") pod \"b4d53cba-d83e-4b1f-afe5-b6f21b939fb0\" (UID: \"b4d53cba-d83e-4b1f-afe5-b6f21b939fb0\") " Nov 25 15:21:42 crc kubenswrapper[4890]: I1125 15:21:42.896038 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b4d53cba-d83e-4b1f-afe5-b6f21b939fb0-logs\") pod \"b4d53cba-d83e-4b1f-afe5-b6f21b939fb0\" (UID: \"b4d53cba-d83e-4b1f-afe5-b6f21b939fb0\") " Nov 25 15:21:42 crc kubenswrapper[4890]: I1125 15:21:42.896199 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4d53cba-d83e-4b1f-afe5-b6f21b939fb0-config-data\") pod \"b4d53cba-d83e-4b1f-afe5-b6f21b939fb0\" (UID: \"b4d53cba-d83e-4b1f-afe5-b6f21b939fb0\") " Nov 25 15:21:42 crc kubenswrapper[4890]: I1125 15:21:42.896235 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8cwm2\" (UniqueName: \"kubernetes.io/projected/b4d53cba-d83e-4b1f-afe5-b6f21b939fb0-kube-api-access-8cwm2\") pod \"b4d53cba-d83e-4b1f-afe5-b6f21b939fb0\" (UID: \"b4d53cba-d83e-4b1f-afe5-b6f21b939fb0\") " Nov 25 15:21:42 crc kubenswrapper[4890]: I1125 15:21:42.896266 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4d53cba-d83e-4b1f-afe5-b6f21b939fb0-scripts\") pod \"b4d53cba-d83e-4b1f-afe5-b6f21b939fb0\" (UID: \"b4d53cba-d83e-4b1f-afe5-b6f21b939fb0\") " Nov 25 15:21:42 crc kubenswrapper[4890]: I1125 15:21:42.897152 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4d53cba-d83e-4b1f-afe5-b6f21b939fb0-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "b4d53cba-d83e-4b1f-afe5-b6f21b939fb0" (UID: "b4d53cba-d83e-4b1f-afe5-b6f21b939fb0"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:21:42 crc kubenswrapper[4890]: I1125 15:21:42.897378 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4d53cba-d83e-4b1f-afe5-b6f21b939fb0-logs" (OuterVolumeSpecName: "logs") pod "b4d53cba-d83e-4b1f-afe5-b6f21b939fb0" (UID: "b4d53cba-d83e-4b1f-afe5-b6f21b939fb0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:21:42 crc kubenswrapper[4890]: I1125 15:21:42.906593 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4d53cba-d83e-4b1f-afe5-b6f21b939fb0-kube-api-access-8cwm2" (OuterVolumeSpecName: "kube-api-access-8cwm2") pod "b4d53cba-d83e-4b1f-afe5-b6f21b939fb0" (UID: "b4d53cba-d83e-4b1f-afe5-b6f21b939fb0"). InnerVolumeSpecName "kube-api-access-8cwm2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:21:42 crc kubenswrapper[4890]: I1125 15:21:42.907793 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4d53cba-d83e-4b1f-afe5-b6f21b939fb0-scripts" (OuterVolumeSpecName: "scripts") pod "b4d53cba-d83e-4b1f-afe5-b6f21b939fb0" (UID: "b4d53cba-d83e-4b1f-afe5-b6f21b939fb0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:21:42 crc kubenswrapper[4890]: I1125 15:21:42.912338 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "b4d53cba-d83e-4b1f-afe5-b6f21b939fb0" (UID: "b4d53cba-d83e-4b1f-afe5-b6f21b939fb0"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 15:21:42 crc kubenswrapper[4890]: I1125 15:21:42.973383 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4d53cba-d83e-4b1f-afe5-b6f21b939fb0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b4d53cba-d83e-4b1f-afe5-b6f21b939fb0" (UID: "b4d53cba-d83e-4b1f-afe5-b6f21b939fb0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:21:42 crc kubenswrapper[4890]: I1125 15:21:42.999362 4890 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b4d53cba-d83e-4b1f-afe5-b6f21b939fb0-logs\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:42 crc kubenswrapper[4890]: I1125 15:21:42.999405 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8cwm2\" (UniqueName: \"kubernetes.io/projected/b4d53cba-d83e-4b1f-afe5-b6f21b939fb0-kube-api-access-8cwm2\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:42 crc kubenswrapper[4890]: I1125 15:21:42.999419 4890 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4d53cba-d83e-4b1f-afe5-b6f21b939fb0-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:42 crc kubenswrapper[4890]: I1125 15:21:42.999446 4890 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Nov 25 15:21:42 crc kubenswrapper[4890]: I1125 15:21:42.999460 4890 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b4d53cba-d83e-4b1f-afe5-b6f21b939fb0-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:42 crc kubenswrapper[4890]: I1125 15:21:42.999473 4890 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4d53cba-d83e-4b1f-afe5-b6f21b939fb0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.051899 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4d53cba-d83e-4b1f-afe5-b6f21b939fb0-config-data" (OuterVolumeSpecName: "config-data") pod "b4d53cba-d83e-4b1f-afe5-b6f21b939fb0" (UID: "b4d53cba-d83e-4b1f-afe5-b6f21b939fb0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.061853 4890 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.062272 4890 generic.go:334] "Generic (PLEG): container finished" podID="b4d53cba-d83e-4b1f-afe5-b6f21b939fb0" containerID="f387b29da54c75ad0c6513c82dbb7b3fa2bceb03bcf5e8cf39deb409c2f611a9" exitCode=143 Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.062298 4890 generic.go:334] "Generic (PLEG): container finished" podID="b4d53cba-d83e-4b1f-afe5-b6f21b939fb0" containerID="1c7d653f81fa1e71828fdb1eaff2feeb554e88bf38c9e22c4f90fbfcde3f2659" exitCode=143 Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.062349 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b4d53cba-d83e-4b1f-afe5-b6f21b939fb0","Type":"ContainerDied","Data":"f387b29da54c75ad0c6513c82dbb7b3fa2bceb03bcf5e8cf39deb409c2f611a9"} Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.062380 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b4d53cba-d83e-4b1f-afe5-b6f21b939fb0","Type":"ContainerDied","Data":"1c7d653f81fa1e71828fdb1eaff2feeb554e88bf38c9e22c4f90fbfcde3f2659"} Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.062394 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b4d53cba-d83e-4b1f-afe5-b6f21b939fb0","Type":"ContainerDied","Data":"65d5a66735e95719ff9ff16e32207333a2724eb8b6d296fedb32492b0d2cd16c"} Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.062412 4890 scope.go:117] "RemoveContainer" containerID="f387b29da54c75ad0c6513c82dbb7b3fa2bceb03bcf5e8cf39deb409c2f611a9" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.062561 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.101786 4890 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4d53cba-d83e-4b1f-afe5-b6f21b939fb0-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.101831 4890 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.128434 4890 generic.go:334] "Generic (PLEG): container finished" podID="1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f" containerID="d1b1f601af90078f6742d375b3a9c3a0ce2fefdf39c2cde7a1057b3e2a3f8950" exitCode=137 Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.128511 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-556858788d-s2ppl" event={"ID":"1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f","Type":"ContainerDied","Data":"d1b1f601af90078f6742d375b3a9c3a0ce2fefdf39c2cde7a1057b3e2a3f8950"} Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.128547 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-556858788d-s2ppl" event={"ID":"1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f","Type":"ContainerDied","Data":"fa469e382145d10056c39ef4297c66072335b5c0cd9e1036a451051c60370bcb"} Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.128561 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa469e382145d10056c39ef4297c66072335b5c0cd9e1036a451051c60370bcb" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.134855 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d65af5aa-095d-465f-8fd7-ba76b4ec08af","Type":"ContainerStarted","Data":"0d6ac61429a65c1c7f8aaf11b9dbd4aa64709aa5aa05a214a384ec49ea22dd13"} Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.135028 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="47f833ef-ad6e-4a03-a1f0-9507d2d1f11e" containerName="glance-log" containerID="cri-o://25f9a26993bce3c5084a666d32531c3c295f8fe62fd7b42825348d6e24d2c230" gracePeriod=30 Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.136192 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="47f833ef-ad6e-4a03-a1f0-9507d2d1f11e" containerName="glance-httpd" containerID="cri-o://5acb64d6ccbff760c46190787766719f1c75efd8e9c67af19facdafca534d0e7" gracePeriod=30 Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.210213 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=5.420234707 podStartE2EDuration="7.210192512s" podCreationTimestamp="2025-11-25 15:21:36 +0000 UTC" firstStartedPulling="2025-11-25 15:21:37.80236649 +0000 UTC m=+1156.244829100" lastFinishedPulling="2025-11-25 15:21:39.592324295 +0000 UTC m=+1158.034786905" observedRunningTime="2025-11-25 15:21:43.186971527 +0000 UTC m=+1161.629434147" watchObservedRunningTime="2025-11-25 15:21:43.210192512 +0000 UTC m=+1161.652655122" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.216655 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-556858788d-s2ppl" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.279226 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.283742 4890 scope.go:117] "RemoveContainer" containerID="1c7d653f81fa1e71828fdb1eaff2feeb554e88bf38c9e22c4f90fbfcde3f2659" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.296906 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.307236 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f-scripts\") pod \"1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f\" (UID: \"1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f\") " Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.307299 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f-horizon-secret-key\") pod \"1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f\" (UID: \"1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f\") " Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.307379 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zcc2n\" (UniqueName: \"kubernetes.io/projected/1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f-kube-api-access-zcc2n\") pod \"1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f\" (UID: \"1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f\") " Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.307421 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f-logs\") pod \"1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f\" (UID: \"1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f\") " Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.307445 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f-combined-ca-bundle\") pod \"1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f\" (UID: \"1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f\") " Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.307459 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f-config-data\") pod \"1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f\" (UID: \"1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f\") " Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.307479 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f-horizon-tls-certs\") pod \"1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f\" (UID: \"1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f\") " Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.325871 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f-logs" (OuterVolumeSpecName: "logs") pod "1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f" (UID: "1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.395363 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f-kube-api-access-zcc2n" (OuterVolumeSpecName: "kube-api-access-zcc2n") pod "1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f" (UID: "1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f"). InnerVolumeSpecName "kube-api-access-zcc2n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.395721 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 15:21:43 crc kubenswrapper[4890]: E1125 15:21:43.396141 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c7f9325-f498-4cd8-96fe-ec9ec9f57a83" containerName="init" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.396154 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c7f9325-f498-4cd8-96fe-ec9ec9f57a83" containerName="init" Nov 25 15:21:43 crc kubenswrapper[4890]: E1125 15:21:43.396188 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4d53cba-d83e-4b1f-afe5-b6f21b939fb0" containerName="glance-log" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.396194 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4d53cba-d83e-4b1f-afe5-b6f21b939fb0" containerName="glance-log" Nov 25 15:21:43 crc kubenswrapper[4890]: E1125 15:21:43.396205 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f" containerName="horizon" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.396211 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f" containerName="horizon" Nov 25 15:21:43 crc kubenswrapper[4890]: E1125 15:21:43.396233 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f" containerName="horizon-log" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.396239 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f" containerName="horizon-log" Nov 25 15:21:43 crc kubenswrapper[4890]: E1125 15:21:43.396254 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4d53cba-d83e-4b1f-afe5-b6f21b939fb0" containerName="glance-httpd" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.396260 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4d53cba-d83e-4b1f-afe5-b6f21b939fb0" containerName="glance-httpd" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.396411 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c7f9325-f498-4cd8-96fe-ec9ec9f57a83" containerName="init" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.396430 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f" containerName="horizon" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.396448 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4d53cba-d83e-4b1f-afe5-b6f21b939fb0" containerName="glance-log" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.396456 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4d53cba-d83e-4b1f-afe5-b6f21b939fb0" containerName="glance-httpd" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.396465 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f" containerName="horizon-log" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.397361 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.400359 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.400570 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.421390 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f" (UID: "1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.435232 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.440431 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f-horizon-secret-key\") pod \"1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f\" (UID: \"1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f\") " Nov 25 15:21:43 crc kubenswrapper[4890]: W1125 15:21:43.440644 4890 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f/volumes/kubernetes.io~secret/horizon-secret-key Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.440667 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f" (UID: "1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.440667 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/59f62350-c15f-4893-ba59-79cca5353cb0-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"59f62350-c15f-4893-ba59-79cca5353cb0\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.440736 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59f62350-c15f-4893-ba59-79cca5353cb0-config-data\") pod \"glance-default-internal-api-0\" (UID: \"59f62350-c15f-4893-ba59-79cca5353cb0\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.440871 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59f62350-c15f-4893-ba59-79cca5353cb0-scripts\") pod \"glance-default-internal-api-0\" (UID: \"59f62350-c15f-4893-ba59-79cca5353cb0\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.440913 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbl7m\" (UniqueName: \"kubernetes.io/projected/59f62350-c15f-4893-ba59-79cca5353cb0-kube-api-access-hbl7m\") pod \"glance-default-internal-api-0\" (UID: \"59f62350-c15f-4893-ba59-79cca5353cb0\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.441038 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59f62350-c15f-4893-ba59-79cca5353cb0-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"59f62350-c15f-4893-ba59-79cca5353cb0\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.441106 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59f62350-c15f-4893-ba59-79cca5353cb0-logs\") pod \"glance-default-internal-api-0\" (UID: \"59f62350-c15f-4893-ba59-79cca5353cb0\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.441142 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"59f62350-c15f-4893-ba59-79cca5353cb0\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.441205 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/59f62350-c15f-4893-ba59-79cca5353cb0-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"59f62350-c15f-4893-ba59-79cca5353cb0\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.441269 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zcc2n\" (UniqueName: \"kubernetes.io/projected/1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f-kube-api-access-zcc2n\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.441281 4890 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f-logs\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.441292 4890 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.455623 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f-scripts" (OuterVolumeSpecName: "scripts") pod "1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f" (UID: "1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.487322 4890 scope.go:117] "RemoveContainer" containerID="f387b29da54c75ad0c6513c82dbb7b3fa2bceb03bcf5e8cf39deb409c2f611a9" Nov 25 15:21:43 crc kubenswrapper[4890]: E1125 15:21:43.489116 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f387b29da54c75ad0c6513c82dbb7b3fa2bceb03bcf5e8cf39deb409c2f611a9\": container with ID starting with f387b29da54c75ad0c6513c82dbb7b3fa2bceb03bcf5e8cf39deb409c2f611a9 not found: ID does not exist" containerID="f387b29da54c75ad0c6513c82dbb7b3fa2bceb03bcf5e8cf39deb409c2f611a9" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.489169 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f387b29da54c75ad0c6513c82dbb7b3fa2bceb03bcf5e8cf39deb409c2f611a9"} err="failed to get container status \"f387b29da54c75ad0c6513c82dbb7b3fa2bceb03bcf5e8cf39deb409c2f611a9\": rpc error: code = NotFound desc = could not find container \"f387b29da54c75ad0c6513c82dbb7b3fa2bceb03bcf5e8cf39deb409c2f611a9\": container with ID starting with f387b29da54c75ad0c6513c82dbb7b3fa2bceb03bcf5e8cf39deb409c2f611a9 not found: ID does not exist" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.489200 4890 scope.go:117] "RemoveContainer" containerID="1c7d653f81fa1e71828fdb1eaff2feeb554e88bf38c9e22c4f90fbfcde3f2659" Nov 25 15:21:43 crc kubenswrapper[4890]: E1125 15:21:43.493650 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c7d653f81fa1e71828fdb1eaff2feeb554e88bf38c9e22c4f90fbfcde3f2659\": container with ID starting with 1c7d653f81fa1e71828fdb1eaff2feeb554e88bf38c9e22c4f90fbfcde3f2659 not found: ID does not exist" containerID="1c7d653f81fa1e71828fdb1eaff2feeb554e88bf38c9e22c4f90fbfcde3f2659" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.493695 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c7d653f81fa1e71828fdb1eaff2feeb554e88bf38c9e22c4f90fbfcde3f2659"} err="failed to get container status \"1c7d653f81fa1e71828fdb1eaff2feeb554e88bf38c9e22c4f90fbfcde3f2659\": rpc error: code = NotFound desc = could not find container \"1c7d653f81fa1e71828fdb1eaff2feeb554e88bf38c9e22c4f90fbfcde3f2659\": container with ID starting with 1c7d653f81fa1e71828fdb1eaff2feeb554e88bf38c9e22c4f90fbfcde3f2659 not found: ID does not exist" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.493726 4890 scope.go:117] "RemoveContainer" containerID="f387b29da54c75ad0c6513c82dbb7b3fa2bceb03bcf5e8cf39deb409c2f611a9" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.501785 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f387b29da54c75ad0c6513c82dbb7b3fa2bceb03bcf5e8cf39deb409c2f611a9"} err="failed to get container status \"f387b29da54c75ad0c6513c82dbb7b3fa2bceb03bcf5e8cf39deb409c2f611a9\": rpc error: code = NotFound desc = could not find container \"f387b29da54c75ad0c6513c82dbb7b3fa2bceb03bcf5e8cf39deb409c2f611a9\": container with ID starting with f387b29da54c75ad0c6513c82dbb7b3fa2bceb03bcf5e8cf39deb409c2f611a9 not found: ID does not exist" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.501976 4890 scope.go:117] "RemoveContainer" containerID="1c7d653f81fa1e71828fdb1eaff2feeb554e88bf38c9e22c4f90fbfcde3f2659" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.512266 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c7d653f81fa1e71828fdb1eaff2feeb554e88bf38c9e22c4f90fbfcde3f2659"} err="failed to get container status \"1c7d653f81fa1e71828fdb1eaff2feeb554e88bf38c9e22c4f90fbfcde3f2659\": rpc error: code = NotFound desc = could not find container \"1c7d653f81fa1e71828fdb1eaff2feeb554e88bf38c9e22c4f90fbfcde3f2659\": container with ID starting with 1c7d653f81fa1e71828fdb1eaff2feeb554e88bf38c9e22c4f90fbfcde3f2659 not found: ID does not exist" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.513580 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f-config-data" (OuterVolumeSpecName: "config-data") pod "1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f" (UID: "1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.513658 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.541986 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/59f62350-c15f-4893-ba59-79cca5353cb0-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"59f62350-c15f-4893-ba59-79cca5353cb0\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.542237 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59f62350-c15f-4893-ba59-79cca5353cb0-config-data\") pod \"glance-default-internal-api-0\" (UID: \"59f62350-c15f-4893-ba59-79cca5353cb0\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.542407 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59f62350-c15f-4893-ba59-79cca5353cb0-scripts\") pod \"glance-default-internal-api-0\" (UID: \"59f62350-c15f-4893-ba59-79cca5353cb0\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.542491 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbl7m\" (UniqueName: \"kubernetes.io/projected/59f62350-c15f-4893-ba59-79cca5353cb0-kube-api-access-hbl7m\") pod \"glance-default-internal-api-0\" (UID: \"59f62350-c15f-4893-ba59-79cca5353cb0\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.542692 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59f62350-c15f-4893-ba59-79cca5353cb0-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"59f62350-c15f-4893-ba59-79cca5353cb0\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.542790 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59f62350-c15f-4893-ba59-79cca5353cb0-logs\") pod \"glance-default-internal-api-0\" (UID: \"59f62350-c15f-4893-ba59-79cca5353cb0\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.542878 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"59f62350-c15f-4893-ba59-79cca5353cb0\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.542968 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/59f62350-c15f-4893-ba59-79cca5353cb0-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"59f62350-c15f-4893-ba59-79cca5353cb0\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.543269 4890 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.543339 4890 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.542551 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/59f62350-c15f-4893-ba59-79cca5353cb0-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"59f62350-c15f-4893-ba59-79cca5353cb0\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.551191 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59f62350-c15f-4893-ba59-79cca5353cb0-logs\") pod \"glance-default-internal-api-0\" (UID: \"59f62350-c15f-4893-ba59-79cca5353cb0\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.551683 4890 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"59f62350-c15f-4893-ba59-79cca5353cb0\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-internal-api-0" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.581119 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f" (UID: "1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.582351 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59f62350-c15f-4893-ba59-79cca5353cb0-scripts\") pod \"glance-default-internal-api-0\" (UID: \"59f62350-c15f-4893-ba59-79cca5353cb0\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.582963 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59f62350-c15f-4893-ba59-79cca5353cb0-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"59f62350-c15f-4893-ba59-79cca5353cb0\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.585843 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/59f62350-c15f-4893-ba59-79cca5353cb0-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"59f62350-c15f-4893-ba59-79cca5353cb0\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.587097 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbl7m\" (UniqueName: \"kubernetes.io/projected/59f62350-c15f-4893-ba59-79cca5353cb0-kube-api-access-hbl7m\") pod \"glance-default-internal-api-0\" (UID: \"59f62350-c15f-4893-ba59-79cca5353cb0\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.590905 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59f62350-c15f-4893-ba59-79cca5353cb0-config-data\") pod \"glance-default-internal-api-0\" (UID: \"59f62350-c15f-4893-ba59-79cca5353cb0\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.632102 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f" (UID: "1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.646578 4890 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.646622 4890 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.669002 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"59f62350-c15f-4893-ba59-79cca5353cb0\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:43 crc kubenswrapper[4890]: I1125 15:21:43.760888 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 15:21:44 crc kubenswrapper[4890]: I1125 15:21:44.202903 4890 generic.go:334] "Generic (PLEG): container finished" podID="47f833ef-ad6e-4a03-a1f0-9507d2d1f11e" containerID="25f9a26993bce3c5084a666d32531c3c295f8fe62fd7b42825348d6e24d2c230" exitCode=143 Nov 25 15:21:44 crc kubenswrapper[4890]: I1125 15:21:44.203467 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-556858788d-s2ppl" Nov 25 15:21:44 crc kubenswrapper[4890]: I1125 15:21:44.206330 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4d53cba-d83e-4b1f-afe5-b6f21b939fb0" path="/var/lib/kubelet/pods/b4d53cba-d83e-4b1f-afe5-b6f21b939fb0/volumes" Nov 25 15:21:44 crc kubenswrapper[4890]: I1125 15:21:44.207100 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="bd0c2b2d-c230-487c-ae4f-f231e973e1c3" containerName="cinder-api-log" containerID="cri-o://fc645d4225419eb44aac652080ca3a27fbcc00627a803e750a4c82f85299871a" gracePeriod=30 Nov 25 15:21:44 crc kubenswrapper[4890]: I1125 15:21:44.209409 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="bd0c2b2d-c230-487c-ae4f-f231e973e1c3" containerName="cinder-api" containerID="cri-o://13f2e42ea2f12408dbaed3897ff7cee43f4e235b1ab146c7f90fa9f8b7efefd0" gracePeriod=30 Nov 25 15:21:44 crc kubenswrapper[4890]: I1125 15:21:44.221128 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.6939325419999998 podStartE2EDuration="9.221112211s" podCreationTimestamp="2025-11-25 15:21:35 +0000 UTC" firstStartedPulling="2025-11-25 15:21:37.345951141 +0000 UTC m=+1155.788413751" lastFinishedPulling="2025-11-25 15:21:42.87313081 +0000 UTC m=+1161.315593420" observedRunningTime="2025-11-25 15:21:44.214591977 +0000 UTC m=+1162.657054587" watchObservedRunningTime="2025-11-25 15:21:44.221112211 +0000 UTC m=+1162.663574821" Nov 25 15:21:44 crc kubenswrapper[4890]: I1125 15:21:44.223718 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 25 15:21:44 crc kubenswrapper[4890]: I1125 15:21:44.223745 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f1ae99b8-5004-434a-812d-0ab8043bd031","Type":"ContainerStarted","Data":"f3f7fdb1eac70f15ae9dbb3f74123477d82ac50e5c44c0d7c36eccb1e03f2ed7"} Nov 25 15:21:44 crc kubenswrapper[4890]: I1125 15:21:44.223762 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"47f833ef-ad6e-4a03-a1f0-9507d2d1f11e","Type":"ContainerDied","Data":"25f9a26993bce3c5084a666d32531c3c295f8fe62fd7b42825348d6e24d2c230"} Nov 25 15:21:44 crc kubenswrapper[4890]: I1125 15:21:44.281781 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-556858788d-s2ppl"] Nov 25 15:21:44 crc kubenswrapper[4890]: I1125 15:21:44.295483 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-556858788d-s2ppl"] Nov 25 15:21:44 crc kubenswrapper[4890]: I1125 15:21:44.502465 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 15:21:44 crc kubenswrapper[4890]: W1125 15:21:44.511792 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod59f62350_c15f_4893_ba59_79cca5353cb0.slice/crio-b64868a75f02ebf8ff528cbfa6523e692098b40ba196625945d65b60842aa015 WatchSource:0}: Error finding container b64868a75f02ebf8ff528cbfa6523e692098b40ba196625945d65b60842aa015: Status 404 returned error can't find the container with id b64868a75f02ebf8ff528cbfa6523e692098b40ba196625945d65b60842aa015 Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.105797 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.176642 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd0c2b2d-c230-487c-ae4f-f231e973e1c3-combined-ca-bundle\") pod \"bd0c2b2d-c230-487c-ae4f-f231e973e1c3\" (UID: \"bd0c2b2d-c230-487c-ae4f-f231e973e1c3\") " Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.176686 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bd0c2b2d-c230-487c-ae4f-f231e973e1c3-config-data-custom\") pod \"bd0c2b2d-c230-487c-ae4f-f231e973e1c3\" (UID: \"bd0c2b2d-c230-487c-ae4f-f231e973e1c3\") " Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.176773 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6w2g7\" (UniqueName: \"kubernetes.io/projected/bd0c2b2d-c230-487c-ae4f-f231e973e1c3-kube-api-access-6w2g7\") pod \"bd0c2b2d-c230-487c-ae4f-f231e973e1c3\" (UID: \"bd0c2b2d-c230-487c-ae4f-f231e973e1c3\") " Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.176801 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd0c2b2d-c230-487c-ae4f-f231e973e1c3-config-data\") pod \"bd0c2b2d-c230-487c-ae4f-f231e973e1c3\" (UID: \"bd0c2b2d-c230-487c-ae4f-f231e973e1c3\") " Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.176829 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd0c2b2d-c230-487c-ae4f-f231e973e1c3-logs\") pod \"bd0c2b2d-c230-487c-ae4f-f231e973e1c3\" (UID: \"bd0c2b2d-c230-487c-ae4f-f231e973e1c3\") " Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.176862 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bd0c2b2d-c230-487c-ae4f-f231e973e1c3-etc-machine-id\") pod \"bd0c2b2d-c230-487c-ae4f-f231e973e1c3\" (UID: \"bd0c2b2d-c230-487c-ae4f-f231e973e1c3\") " Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.176985 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd0c2b2d-c230-487c-ae4f-f231e973e1c3-scripts\") pod \"bd0c2b2d-c230-487c-ae4f-f231e973e1c3\" (UID: \"bd0c2b2d-c230-487c-ae4f-f231e973e1c3\") " Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.177292 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bd0c2b2d-c230-487c-ae4f-f231e973e1c3-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "bd0c2b2d-c230-487c-ae4f-f231e973e1c3" (UID: "bd0c2b2d-c230-487c-ae4f-f231e973e1c3"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.177396 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd0c2b2d-c230-487c-ae4f-f231e973e1c3-logs" (OuterVolumeSpecName: "logs") pod "bd0c2b2d-c230-487c-ae4f-f231e973e1c3" (UID: "bd0c2b2d-c230-487c-ae4f-f231e973e1c3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.177513 4890 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bd0c2b2d-c230-487c-ae4f-f231e973e1c3-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.183310 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd0c2b2d-c230-487c-ae4f-f231e973e1c3-scripts" (OuterVolumeSpecName: "scripts") pod "bd0c2b2d-c230-487c-ae4f-f231e973e1c3" (UID: "bd0c2b2d-c230-487c-ae4f-f231e973e1c3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.183352 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd0c2b2d-c230-487c-ae4f-f231e973e1c3-kube-api-access-6w2g7" (OuterVolumeSpecName: "kube-api-access-6w2g7") pod "bd0c2b2d-c230-487c-ae4f-f231e973e1c3" (UID: "bd0c2b2d-c230-487c-ae4f-f231e973e1c3"). InnerVolumeSpecName "kube-api-access-6w2g7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.186057 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd0c2b2d-c230-487c-ae4f-f231e973e1c3-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "bd0c2b2d-c230-487c-ae4f-f231e973e1c3" (UID: "bd0c2b2d-c230-487c-ae4f-f231e973e1c3"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.216194 4890 generic.go:334] "Generic (PLEG): container finished" podID="47f833ef-ad6e-4a03-a1f0-9507d2d1f11e" containerID="5acb64d6ccbff760c46190787766719f1c75efd8e9c67af19facdafca534d0e7" exitCode=0 Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.216270 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"47f833ef-ad6e-4a03-a1f0-9507d2d1f11e","Type":"ContainerDied","Data":"5acb64d6ccbff760c46190787766719f1c75efd8e9c67af19facdafca534d0e7"} Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.217973 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"59f62350-c15f-4893-ba59-79cca5353cb0","Type":"ContainerStarted","Data":"b64868a75f02ebf8ff528cbfa6523e692098b40ba196625945d65b60842aa015"} Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.225534 4890 generic.go:334] "Generic (PLEG): container finished" podID="bd0c2b2d-c230-487c-ae4f-f231e973e1c3" containerID="13f2e42ea2f12408dbaed3897ff7cee43f4e235b1ab146c7f90fa9f8b7efefd0" exitCode=0 Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.225595 4890 generic.go:334] "Generic (PLEG): container finished" podID="bd0c2b2d-c230-487c-ae4f-f231e973e1c3" containerID="fc645d4225419eb44aac652080ca3a27fbcc00627a803e750a4c82f85299871a" exitCode=143 Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.226127 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.228079 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"bd0c2b2d-c230-487c-ae4f-f231e973e1c3","Type":"ContainerDied","Data":"13f2e42ea2f12408dbaed3897ff7cee43f4e235b1ab146c7f90fa9f8b7efefd0"} Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.228112 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"bd0c2b2d-c230-487c-ae4f-f231e973e1c3","Type":"ContainerDied","Data":"fc645d4225419eb44aac652080ca3a27fbcc00627a803e750a4c82f85299871a"} Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.228123 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"bd0c2b2d-c230-487c-ae4f-f231e973e1c3","Type":"ContainerDied","Data":"77baf01d5d4a5ee3d68e83d21894d2fa1f336e9768fcd79e2b239493021bf205"} Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.228138 4890 scope.go:117] "RemoveContainer" containerID="13f2e42ea2f12408dbaed3897ff7cee43f4e235b1ab146c7f90fa9f8b7efefd0" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.268401 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd0c2b2d-c230-487c-ae4f-f231e973e1c3-config-data" (OuterVolumeSpecName: "config-data") pod "bd0c2b2d-c230-487c-ae4f-f231e973e1c3" (UID: "bd0c2b2d-c230-487c-ae4f-f231e973e1c3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.269643 4890 scope.go:117] "RemoveContainer" containerID="fc645d4225419eb44aac652080ca3a27fbcc00627a803e750a4c82f85299871a" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.274633 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7d4dcdcc94-g9drz" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.281256 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6w2g7\" (UniqueName: \"kubernetes.io/projected/bd0c2b2d-c230-487c-ae4f-f231e973e1c3-kube-api-access-6w2g7\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.281280 4890 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd0c2b2d-c230-487c-ae4f-f231e973e1c3-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.281289 4890 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd0c2b2d-c230-487c-ae4f-f231e973e1c3-logs\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.281297 4890 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd0c2b2d-c230-487c-ae4f-f231e973e1c3-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.281305 4890 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bd0c2b2d-c230-487c-ae4f-f231e973e1c3-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.314292 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd0c2b2d-c230-487c-ae4f-f231e973e1c3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bd0c2b2d-c230-487c-ae4f-f231e973e1c3" (UID: "bd0c2b2d-c230-487c-ae4f-f231e973e1c3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.323226 4890 scope.go:117] "RemoveContainer" containerID="13f2e42ea2f12408dbaed3897ff7cee43f4e235b1ab146c7f90fa9f8b7efefd0" Nov 25 15:21:45 crc kubenswrapper[4890]: E1125 15:21:45.327115 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13f2e42ea2f12408dbaed3897ff7cee43f4e235b1ab146c7f90fa9f8b7efefd0\": container with ID starting with 13f2e42ea2f12408dbaed3897ff7cee43f4e235b1ab146c7f90fa9f8b7efefd0 not found: ID does not exist" containerID="13f2e42ea2f12408dbaed3897ff7cee43f4e235b1ab146c7f90fa9f8b7efefd0" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.327172 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13f2e42ea2f12408dbaed3897ff7cee43f4e235b1ab146c7f90fa9f8b7efefd0"} err="failed to get container status \"13f2e42ea2f12408dbaed3897ff7cee43f4e235b1ab146c7f90fa9f8b7efefd0\": rpc error: code = NotFound desc = could not find container \"13f2e42ea2f12408dbaed3897ff7cee43f4e235b1ab146c7f90fa9f8b7efefd0\": container with ID starting with 13f2e42ea2f12408dbaed3897ff7cee43f4e235b1ab146c7f90fa9f8b7efefd0 not found: ID does not exist" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.327202 4890 scope.go:117] "RemoveContainer" containerID="fc645d4225419eb44aac652080ca3a27fbcc00627a803e750a4c82f85299871a" Nov 25 15:21:45 crc kubenswrapper[4890]: E1125 15:21:45.332614 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc645d4225419eb44aac652080ca3a27fbcc00627a803e750a4c82f85299871a\": container with ID starting with fc645d4225419eb44aac652080ca3a27fbcc00627a803e750a4c82f85299871a not found: ID does not exist" containerID="fc645d4225419eb44aac652080ca3a27fbcc00627a803e750a4c82f85299871a" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.334208 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc645d4225419eb44aac652080ca3a27fbcc00627a803e750a4c82f85299871a"} err="failed to get container status \"fc645d4225419eb44aac652080ca3a27fbcc00627a803e750a4c82f85299871a\": rpc error: code = NotFound desc = could not find container \"fc645d4225419eb44aac652080ca3a27fbcc00627a803e750a4c82f85299871a\": container with ID starting with fc645d4225419eb44aac652080ca3a27fbcc00627a803e750a4c82f85299871a not found: ID does not exist" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.334329 4890 scope.go:117] "RemoveContainer" containerID="13f2e42ea2f12408dbaed3897ff7cee43f4e235b1ab146c7f90fa9f8b7efefd0" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.339897 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13f2e42ea2f12408dbaed3897ff7cee43f4e235b1ab146c7f90fa9f8b7efefd0"} err="failed to get container status \"13f2e42ea2f12408dbaed3897ff7cee43f4e235b1ab146c7f90fa9f8b7efefd0\": rpc error: code = NotFound desc = could not find container \"13f2e42ea2f12408dbaed3897ff7cee43f4e235b1ab146c7f90fa9f8b7efefd0\": container with ID starting with 13f2e42ea2f12408dbaed3897ff7cee43f4e235b1ab146c7f90fa9f8b7efefd0 not found: ID does not exist" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.340225 4890 scope.go:117] "RemoveContainer" containerID="fc645d4225419eb44aac652080ca3a27fbcc00627a803e750a4c82f85299871a" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.345599 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc645d4225419eb44aac652080ca3a27fbcc00627a803e750a4c82f85299871a"} err="failed to get container status \"fc645d4225419eb44aac652080ca3a27fbcc00627a803e750a4c82f85299871a\": rpc error: code = NotFound desc = could not find container \"fc645d4225419eb44aac652080ca3a27fbcc00627a803e750a4c82f85299871a\": container with ID starting with fc645d4225419eb44aac652080ca3a27fbcc00627a803e750a4c82f85299871a not found: ID does not exist" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.376390 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.384905 4890 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd0c2b2d-c230-487c-ae4f-f231e973e1c3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.487150 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47f833ef-ad6e-4a03-a1f0-9507d2d1f11e-logs\") pod \"47f833ef-ad6e-4a03-a1f0-9507d2d1f11e\" (UID: \"47f833ef-ad6e-4a03-a1f0-9507d2d1f11e\") " Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.487299 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/47f833ef-ad6e-4a03-a1f0-9507d2d1f11e-httpd-run\") pod \"47f833ef-ad6e-4a03-a1f0-9507d2d1f11e\" (UID: \"47f833ef-ad6e-4a03-a1f0-9507d2d1f11e\") " Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.487397 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47f833ef-ad6e-4a03-a1f0-9507d2d1f11e-scripts\") pod \"47f833ef-ad6e-4a03-a1f0-9507d2d1f11e\" (UID: \"47f833ef-ad6e-4a03-a1f0-9507d2d1f11e\") " Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.487532 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rhvw9\" (UniqueName: \"kubernetes.io/projected/47f833ef-ad6e-4a03-a1f0-9507d2d1f11e-kube-api-access-rhvw9\") pod \"47f833ef-ad6e-4a03-a1f0-9507d2d1f11e\" (UID: \"47f833ef-ad6e-4a03-a1f0-9507d2d1f11e\") " Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.487662 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47f833ef-ad6e-4a03-a1f0-9507d2d1f11e-combined-ca-bundle\") pod \"47f833ef-ad6e-4a03-a1f0-9507d2d1f11e\" (UID: \"47f833ef-ad6e-4a03-a1f0-9507d2d1f11e\") " Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.487742 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47f833ef-ad6e-4a03-a1f0-9507d2d1f11e-config-data\") pod \"47f833ef-ad6e-4a03-a1f0-9507d2d1f11e\" (UID: \"47f833ef-ad6e-4a03-a1f0-9507d2d1f11e\") " Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.487786 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47f833ef-ad6e-4a03-a1f0-9507d2d1f11e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "47f833ef-ad6e-4a03-a1f0-9507d2d1f11e" (UID: "47f833ef-ad6e-4a03-a1f0-9507d2d1f11e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.487883 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"47f833ef-ad6e-4a03-a1f0-9507d2d1f11e\" (UID: \"47f833ef-ad6e-4a03-a1f0-9507d2d1f11e\") " Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.488021 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47f833ef-ad6e-4a03-a1f0-9507d2d1f11e-logs" (OuterVolumeSpecName: "logs") pod "47f833ef-ad6e-4a03-a1f0-9507d2d1f11e" (UID: "47f833ef-ad6e-4a03-a1f0-9507d2d1f11e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.488704 4890 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47f833ef-ad6e-4a03-a1f0-9507d2d1f11e-logs\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.488732 4890 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/47f833ef-ad6e-4a03-a1f0-9507d2d1f11e-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.492291 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47f833ef-ad6e-4a03-a1f0-9507d2d1f11e-scripts" (OuterVolumeSpecName: "scripts") pod "47f833ef-ad6e-4a03-a1f0-9507d2d1f11e" (UID: "47f833ef-ad6e-4a03-a1f0-9507d2d1f11e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.492309 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47f833ef-ad6e-4a03-a1f0-9507d2d1f11e-kube-api-access-rhvw9" (OuterVolumeSpecName: "kube-api-access-rhvw9") pod "47f833ef-ad6e-4a03-a1f0-9507d2d1f11e" (UID: "47f833ef-ad6e-4a03-a1f0-9507d2d1f11e"). InnerVolumeSpecName "kube-api-access-rhvw9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.499457 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "47f833ef-ad6e-4a03-a1f0-9507d2d1f11e" (UID: "47f833ef-ad6e-4a03-a1f0-9507d2d1f11e"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.546467 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47f833ef-ad6e-4a03-a1f0-9507d2d1f11e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "47f833ef-ad6e-4a03-a1f0-9507d2d1f11e" (UID: "47f833ef-ad6e-4a03-a1f0-9507d2d1f11e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.568969 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.589760 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.590975 4890 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47f833ef-ad6e-4a03-a1f0-9507d2d1f11e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.591027 4890 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.591047 4890 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47f833ef-ad6e-4a03-a1f0-9507d2d1f11e-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.591061 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rhvw9\" (UniqueName: \"kubernetes.io/projected/47f833ef-ad6e-4a03-a1f0-9507d2d1f11e-kube-api-access-rhvw9\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.642777 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 25 15:21:45 crc kubenswrapper[4890]: E1125 15:21:45.643213 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd0c2b2d-c230-487c-ae4f-f231e973e1c3" containerName="cinder-api" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.643224 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd0c2b2d-c230-487c-ae4f-f231e973e1c3" containerName="cinder-api" Nov 25 15:21:45 crc kubenswrapper[4890]: E1125 15:21:45.643243 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47f833ef-ad6e-4a03-a1f0-9507d2d1f11e" containerName="glance-log" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.643249 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="47f833ef-ad6e-4a03-a1f0-9507d2d1f11e" containerName="glance-log" Nov 25 15:21:45 crc kubenswrapper[4890]: E1125 15:21:45.643274 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd0c2b2d-c230-487c-ae4f-f231e973e1c3" containerName="cinder-api-log" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.643280 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd0c2b2d-c230-487c-ae4f-f231e973e1c3" containerName="cinder-api-log" Nov 25 15:21:45 crc kubenswrapper[4890]: E1125 15:21:45.643289 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47f833ef-ad6e-4a03-a1f0-9507d2d1f11e" containerName="glance-httpd" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.643296 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="47f833ef-ad6e-4a03-a1f0-9507d2d1f11e" containerName="glance-httpd" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.643455 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd0c2b2d-c230-487c-ae4f-f231e973e1c3" containerName="cinder-api-log" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.643470 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd0c2b2d-c230-487c-ae4f-f231e973e1c3" containerName="cinder-api" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.643490 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="47f833ef-ad6e-4a03-a1f0-9507d2d1f11e" containerName="glance-log" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.643504 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="47f833ef-ad6e-4a03-a1f0-9507d2d1f11e" containerName="glance-httpd" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.644406 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47f833ef-ad6e-4a03-a1f0-9507d2d1f11e-config-data" (OuterVolumeSpecName: "config-data") pod "47f833ef-ad6e-4a03-a1f0-9507d2d1f11e" (UID: "47f833ef-ad6e-4a03-a1f0-9507d2d1f11e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.644496 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.648523 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.648784 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.648933 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.665331 4890 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.670686 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.694422 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/309052f6-3a50-436c-89e0-b97d5454b935-public-tls-certs\") pod \"cinder-api-0\" (UID: \"309052f6-3a50-436c-89e0-b97d5454b935\") " pod="openstack/cinder-api-0" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.694802 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/309052f6-3a50-436c-89e0-b97d5454b935-config-data\") pod \"cinder-api-0\" (UID: \"309052f6-3a50-436c-89e0-b97d5454b935\") " pod="openstack/cinder-api-0" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.694846 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/309052f6-3a50-436c-89e0-b97d5454b935-config-data-custom\") pod \"cinder-api-0\" (UID: \"309052f6-3a50-436c-89e0-b97d5454b935\") " pod="openstack/cinder-api-0" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.694927 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/309052f6-3a50-436c-89e0-b97d5454b935-logs\") pod \"cinder-api-0\" (UID: \"309052f6-3a50-436c-89e0-b97d5454b935\") " pod="openstack/cinder-api-0" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.694983 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/309052f6-3a50-436c-89e0-b97d5454b935-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"309052f6-3a50-436c-89e0-b97d5454b935\") " pod="openstack/cinder-api-0" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.695050 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/309052f6-3a50-436c-89e0-b97d5454b935-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"309052f6-3a50-436c-89e0-b97d5454b935\") " pod="openstack/cinder-api-0" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.695091 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/309052f6-3a50-436c-89e0-b97d5454b935-etc-machine-id\") pod \"cinder-api-0\" (UID: \"309052f6-3a50-436c-89e0-b97d5454b935\") " pod="openstack/cinder-api-0" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.695129 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dc9nh\" (UniqueName: \"kubernetes.io/projected/309052f6-3a50-436c-89e0-b97d5454b935-kube-api-access-dc9nh\") pod \"cinder-api-0\" (UID: \"309052f6-3a50-436c-89e0-b97d5454b935\") " pod="openstack/cinder-api-0" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.695151 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/309052f6-3a50-436c-89e0-b97d5454b935-scripts\") pod \"cinder-api-0\" (UID: \"309052f6-3a50-436c-89e0-b97d5454b935\") " pod="openstack/cinder-api-0" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.695248 4890 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47f833ef-ad6e-4a03-a1f0-9507d2d1f11e-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.695264 4890 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.796030 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/309052f6-3a50-436c-89e0-b97d5454b935-logs\") pod \"cinder-api-0\" (UID: \"309052f6-3a50-436c-89e0-b97d5454b935\") " pod="openstack/cinder-api-0" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.796089 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/309052f6-3a50-436c-89e0-b97d5454b935-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"309052f6-3a50-436c-89e0-b97d5454b935\") " pod="openstack/cinder-api-0" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.796134 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/309052f6-3a50-436c-89e0-b97d5454b935-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"309052f6-3a50-436c-89e0-b97d5454b935\") " pod="openstack/cinder-api-0" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.796176 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/309052f6-3a50-436c-89e0-b97d5454b935-etc-machine-id\") pod \"cinder-api-0\" (UID: \"309052f6-3a50-436c-89e0-b97d5454b935\") " pod="openstack/cinder-api-0" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.796200 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dc9nh\" (UniqueName: \"kubernetes.io/projected/309052f6-3a50-436c-89e0-b97d5454b935-kube-api-access-dc9nh\") pod \"cinder-api-0\" (UID: \"309052f6-3a50-436c-89e0-b97d5454b935\") " pod="openstack/cinder-api-0" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.796214 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/309052f6-3a50-436c-89e0-b97d5454b935-scripts\") pod \"cinder-api-0\" (UID: \"309052f6-3a50-436c-89e0-b97d5454b935\") " pod="openstack/cinder-api-0" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.796245 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/309052f6-3a50-436c-89e0-b97d5454b935-public-tls-certs\") pod \"cinder-api-0\" (UID: \"309052f6-3a50-436c-89e0-b97d5454b935\") " pod="openstack/cinder-api-0" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.796259 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/309052f6-3a50-436c-89e0-b97d5454b935-config-data\") pod \"cinder-api-0\" (UID: \"309052f6-3a50-436c-89e0-b97d5454b935\") " pod="openstack/cinder-api-0" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.796280 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/309052f6-3a50-436c-89e0-b97d5454b935-config-data-custom\") pod \"cinder-api-0\" (UID: \"309052f6-3a50-436c-89e0-b97d5454b935\") " pod="openstack/cinder-api-0" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.797322 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/309052f6-3a50-436c-89e0-b97d5454b935-logs\") pod \"cinder-api-0\" (UID: \"309052f6-3a50-436c-89e0-b97d5454b935\") " pod="openstack/cinder-api-0" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.799983 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/309052f6-3a50-436c-89e0-b97d5454b935-etc-machine-id\") pod \"cinder-api-0\" (UID: \"309052f6-3a50-436c-89e0-b97d5454b935\") " pod="openstack/cinder-api-0" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.801504 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/309052f6-3a50-436c-89e0-b97d5454b935-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"309052f6-3a50-436c-89e0-b97d5454b935\") " pod="openstack/cinder-api-0" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.805583 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/309052f6-3a50-436c-89e0-b97d5454b935-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"309052f6-3a50-436c-89e0-b97d5454b935\") " pod="openstack/cinder-api-0" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.805854 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/309052f6-3a50-436c-89e0-b97d5454b935-config-data-custom\") pod \"cinder-api-0\" (UID: \"309052f6-3a50-436c-89e0-b97d5454b935\") " pod="openstack/cinder-api-0" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.806200 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/309052f6-3a50-436c-89e0-b97d5454b935-scripts\") pod \"cinder-api-0\" (UID: \"309052f6-3a50-436c-89e0-b97d5454b935\") " pod="openstack/cinder-api-0" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.808906 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/309052f6-3a50-436c-89e0-b97d5454b935-config-data\") pod \"cinder-api-0\" (UID: \"309052f6-3a50-436c-89e0-b97d5454b935\") " pod="openstack/cinder-api-0" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.830757 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dc9nh\" (UniqueName: \"kubernetes.io/projected/309052f6-3a50-436c-89e0-b97d5454b935-kube-api-access-dc9nh\") pod \"cinder-api-0\" (UID: \"309052f6-3a50-436c-89e0-b97d5454b935\") " pod="openstack/cinder-api-0" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.835406 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/309052f6-3a50-436c-89e0-b97d5454b935-public-tls-certs\") pod \"cinder-api-0\" (UID: \"309052f6-3a50-436c-89e0-b97d5454b935\") " pod="openstack/cinder-api-0" Nov 25 15:21:45 crc kubenswrapper[4890]: I1125 15:21:45.946055 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7d4dcdcc94-g9drz" Nov 25 15:21:46 crc kubenswrapper[4890]: I1125 15:21:46.003765 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-8f5fb9544-r49m4"] Nov 25 15:21:46 crc kubenswrapper[4890]: I1125 15:21:46.003984 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-8f5fb9544-r49m4" podUID="2f6fb686-a04a-40d5-a67a-caddba7e8d4d" containerName="barbican-api-log" containerID="cri-o://a1b729f7f24bb522a987cb0b04c7b9a5b52df51799327aa3e4d15458ec94d95c" gracePeriod=30 Nov 25 15:21:46 crc kubenswrapper[4890]: I1125 15:21:46.010618 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-8f5fb9544-r49m4" podUID="2f6fb686-a04a-40d5-a67a-caddba7e8d4d" containerName="barbican-api" containerID="cri-o://b1e404d377b43303ab6cf27c3866fde1baa533c9810625bcbf24c2695dfc0d89" gracePeriod=30 Nov 25 15:21:46 crc kubenswrapper[4890]: I1125 15:21:46.078358 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 25 15:21:46 crc kubenswrapper[4890]: I1125 15:21:46.217896 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f" path="/var/lib/kubelet/pods/1ed31a6e-d5bf-4e5f-9e28-cb0e29aad74f/volumes" Nov 25 15:21:46 crc kubenswrapper[4890]: I1125 15:21:46.218721 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd0c2b2d-c230-487c-ae4f-f231e973e1c3" path="/var/lib/kubelet/pods/bd0c2b2d-c230-487c-ae4f-f231e973e1c3/volumes" Nov 25 15:21:46 crc kubenswrapper[4890]: I1125 15:21:46.268920 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"47f833ef-ad6e-4a03-a1f0-9507d2d1f11e","Type":"ContainerDied","Data":"c16d0c26b1e388fe63a5e78a81af8a7c8d2c6df38dbaa7a1788c51fe961f6235"} Nov 25 15:21:46 crc kubenswrapper[4890]: I1125 15:21:46.268979 4890 scope.go:117] "RemoveContainer" containerID="5acb64d6ccbff760c46190787766719f1c75efd8e9c67af19facdafca534d0e7" Nov 25 15:21:46 crc kubenswrapper[4890]: I1125 15:21:46.269134 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 15:21:46 crc kubenswrapper[4890]: I1125 15:21:46.298009 4890 generic.go:334] "Generic (PLEG): container finished" podID="2f6fb686-a04a-40d5-a67a-caddba7e8d4d" containerID="a1b729f7f24bb522a987cb0b04c7b9a5b52df51799327aa3e4d15458ec94d95c" exitCode=143 Nov 25 15:21:46 crc kubenswrapper[4890]: I1125 15:21:46.298108 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8f5fb9544-r49m4" event={"ID":"2f6fb686-a04a-40d5-a67a-caddba7e8d4d","Type":"ContainerDied","Data":"a1b729f7f24bb522a987cb0b04c7b9a5b52df51799327aa3e4d15458ec94d95c"} Nov 25 15:21:46 crc kubenswrapper[4890]: I1125 15:21:46.310139 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 15:21:46 crc kubenswrapper[4890]: I1125 15:21:46.318695 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"59f62350-c15f-4893-ba59-79cca5353cb0","Type":"ContainerStarted","Data":"6c3312722a7e81ef35562577700748b4a13d43c509d835aca0dc24f388927ae9"} Nov 25 15:21:46 crc kubenswrapper[4890]: I1125 15:21:46.327256 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 15:21:46 crc kubenswrapper[4890]: I1125 15:21:46.338126 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 15:21:46 crc kubenswrapper[4890]: I1125 15:21:46.339963 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 15:21:46 crc kubenswrapper[4890]: I1125 15:21:46.346649 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 15:21:46 crc kubenswrapper[4890]: I1125 15:21:46.350328 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 25 15:21:46 crc kubenswrapper[4890]: I1125 15:21:46.350368 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 25 15:21:46 crc kubenswrapper[4890]: I1125 15:21:46.355473 4890 scope.go:117] "RemoveContainer" containerID="25f9a26993bce3c5084a666d32531c3c295f8fe62fd7b42825348d6e24d2c230" Nov 25 15:21:46 crc kubenswrapper[4890]: I1125 15:21:46.426349 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c617481-223e-4a84-8ecc-087c98dbd459-config-data\") pod \"glance-default-external-api-0\" (UID: \"9c617481-223e-4a84-8ecc-087c98dbd459\") " pod="openstack/glance-default-external-api-0" Nov 25 15:21:46 crc kubenswrapper[4890]: I1125 15:21:46.426455 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c617481-223e-4a84-8ecc-087c98dbd459-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9c617481-223e-4a84-8ecc-087c98dbd459\") " pod="openstack/glance-default-external-api-0" Nov 25 15:21:46 crc kubenswrapper[4890]: I1125 15:21:46.426494 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"9c617481-223e-4a84-8ecc-087c98dbd459\") " pod="openstack/glance-default-external-api-0" Nov 25 15:21:46 crc kubenswrapper[4890]: I1125 15:21:46.426545 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c617481-223e-4a84-8ecc-087c98dbd459-logs\") pod \"glance-default-external-api-0\" (UID: \"9c617481-223e-4a84-8ecc-087c98dbd459\") " pod="openstack/glance-default-external-api-0" Nov 25 15:21:46 crc kubenswrapper[4890]: I1125 15:21:46.426573 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcs9z\" (UniqueName: \"kubernetes.io/projected/9c617481-223e-4a84-8ecc-087c98dbd459-kube-api-access-pcs9z\") pod \"glance-default-external-api-0\" (UID: \"9c617481-223e-4a84-8ecc-087c98dbd459\") " pod="openstack/glance-default-external-api-0" Nov 25 15:21:46 crc kubenswrapper[4890]: I1125 15:21:46.426607 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c617481-223e-4a84-8ecc-087c98dbd459-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9c617481-223e-4a84-8ecc-087c98dbd459\") " pod="openstack/glance-default-external-api-0" Nov 25 15:21:46 crc kubenswrapper[4890]: I1125 15:21:46.426643 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9c617481-223e-4a84-8ecc-087c98dbd459-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9c617481-223e-4a84-8ecc-087c98dbd459\") " pod="openstack/glance-default-external-api-0" Nov 25 15:21:46 crc kubenswrapper[4890]: I1125 15:21:46.426682 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c617481-223e-4a84-8ecc-087c98dbd459-scripts\") pod \"glance-default-external-api-0\" (UID: \"9c617481-223e-4a84-8ecc-087c98dbd459\") " pod="openstack/glance-default-external-api-0" Nov 25 15:21:46 crc kubenswrapper[4890]: I1125 15:21:46.528701 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c617481-223e-4a84-8ecc-087c98dbd459-config-data\") pod \"glance-default-external-api-0\" (UID: \"9c617481-223e-4a84-8ecc-087c98dbd459\") " pod="openstack/glance-default-external-api-0" Nov 25 15:21:46 crc kubenswrapper[4890]: I1125 15:21:46.528789 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c617481-223e-4a84-8ecc-087c98dbd459-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9c617481-223e-4a84-8ecc-087c98dbd459\") " pod="openstack/glance-default-external-api-0" Nov 25 15:21:46 crc kubenswrapper[4890]: I1125 15:21:46.528829 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"9c617481-223e-4a84-8ecc-087c98dbd459\") " pod="openstack/glance-default-external-api-0" Nov 25 15:21:46 crc kubenswrapper[4890]: I1125 15:21:46.528866 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c617481-223e-4a84-8ecc-087c98dbd459-logs\") pod \"glance-default-external-api-0\" (UID: \"9c617481-223e-4a84-8ecc-087c98dbd459\") " pod="openstack/glance-default-external-api-0" Nov 25 15:21:46 crc kubenswrapper[4890]: I1125 15:21:46.528897 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcs9z\" (UniqueName: \"kubernetes.io/projected/9c617481-223e-4a84-8ecc-087c98dbd459-kube-api-access-pcs9z\") pod \"glance-default-external-api-0\" (UID: \"9c617481-223e-4a84-8ecc-087c98dbd459\") " pod="openstack/glance-default-external-api-0" Nov 25 15:21:46 crc kubenswrapper[4890]: I1125 15:21:46.528927 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c617481-223e-4a84-8ecc-087c98dbd459-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9c617481-223e-4a84-8ecc-087c98dbd459\") " pod="openstack/glance-default-external-api-0" Nov 25 15:21:46 crc kubenswrapper[4890]: I1125 15:21:46.528965 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9c617481-223e-4a84-8ecc-087c98dbd459-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9c617481-223e-4a84-8ecc-087c98dbd459\") " pod="openstack/glance-default-external-api-0" Nov 25 15:21:46 crc kubenswrapper[4890]: I1125 15:21:46.529007 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c617481-223e-4a84-8ecc-087c98dbd459-scripts\") pod \"glance-default-external-api-0\" (UID: \"9c617481-223e-4a84-8ecc-087c98dbd459\") " pod="openstack/glance-default-external-api-0" Nov 25 15:21:46 crc kubenswrapper[4890]: I1125 15:21:46.529781 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c617481-223e-4a84-8ecc-087c98dbd459-logs\") pod \"glance-default-external-api-0\" (UID: \"9c617481-223e-4a84-8ecc-087c98dbd459\") " pod="openstack/glance-default-external-api-0" Nov 25 15:21:46 crc kubenswrapper[4890]: I1125 15:21:46.530291 4890 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"9c617481-223e-4a84-8ecc-087c98dbd459\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-external-api-0" Nov 25 15:21:46 crc kubenswrapper[4890]: I1125 15:21:46.530693 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9c617481-223e-4a84-8ecc-087c98dbd459-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9c617481-223e-4a84-8ecc-087c98dbd459\") " pod="openstack/glance-default-external-api-0" Nov 25 15:21:46 crc kubenswrapper[4890]: I1125 15:21:46.546045 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c617481-223e-4a84-8ecc-087c98dbd459-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9c617481-223e-4a84-8ecc-087c98dbd459\") " pod="openstack/glance-default-external-api-0" Nov 25 15:21:46 crc kubenswrapper[4890]: I1125 15:21:46.564448 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c617481-223e-4a84-8ecc-087c98dbd459-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9c617481-223e-4a84-8ecc-087c98dbd459\") " pod="openstack/glance-default-external-api-0" Nov 25 15:21:46 crc kubenswrapper[4890]: I1125 15:21:46.572590 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c617481-223e-4a84-8ecc-087c98dbd459-config-data\") pod \"glance-default-external-api-0\" (UID: \"9c617481-223e-4a84-8ecc-087c98dbd459\") " pod="openstack/glance-default-external-api-0" Nov 25 15:21:46 crc kubenswrapper[4890]: I1125 15:21:46.575817 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c617481-223e-4a84-8ecc-087c98dbd459-scripts\") pod \"glance-default-external-api-0\" (UID: \"9c617481-223e-4a84-8ecc-087c98dbd459\") " pod="openstack/glance-default-external-api-0" Nov 25 15:21:46 crc kubenswrapper[4890]: I1125 15:21:46.580905 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcs9z\" (UniqueName: \"kubernetes.io/projected/9c617481-223e-4a84-8ecc-087c98dbd459-kube-api-access-pcs9z\") pod \"glance-default-external-api-0\" (UID: \"9c617481-223e-4a84-8ecc-087c98dbd459\") " pod="openstack/glance-default-external-api-0" Nov 25 15:21:46 crc kubenswrapper[4890]: I1125 15:21:46.599792 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"9c617481-223e-4a84-8ecc-087c98dbd459\") " pod="openstack/glance-default-external-api-0" Nov 25 15:21:46 crc kubenswrapper[4890]: I1125 15:21:46.657398 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 25 15:21:46 crc kubenswrapper[4890]: I1125 15:21:46.709653 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 15:21:46 crc kubenswrapper[4890]: I1125 15:21:46.730130 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 25 15:21:46 crc kubenswrapper[4890]: W1125 15:21:46.740922 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod309052f6_3a50_436c_89e0_b97d5454b935.slice/crio-4b710264e7c07b2f106b8eb5f3d1aa1d54a2525abf9889d19b0273267e7ea964 WatchSource:0}: Error finding container 4b710264e7c07b2f106b8eb5f3d1aa1d54a2525abf9889d19b0273267e7ea964: Status 404 returned error can't find the container with id 4b710264e7c07b2f106b8eb5f3d1aa1d54a2525abf9889d19b0273267e7ea964 Nov 25 15:21:46 crc kubenswrapper[4890]: I1125 15:21:46.916408 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-795f4db4bc-xwdbs" Nov 25 15:21:47 crc kubenswrapper[4890]: I1125 15:21:47.014306 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58957f86ff-jbsf6"] Nov 25 15:21:47 crc kubenswrapper[4890]: I1125 15:21:47.014533 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-58957f86ff-jbsf6" podUID="34bffacc-3ed2-4f5a-9338-34f1293b3776" containerName="dnsmasq-dns" containerID="cri-o://be14c60cab33ab2ded2e1d36df0016a66348977ed17f64bc0f141bc63f78da6c" gracePeriod=10 Nov 25 15:21:47 crc kubenswrapper[4890]: I1125 15:21:47.269640 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 25 15:21:47 crc kubenswrapper[4890]: I1125 15:21:47.384590 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"59f62350-c15f-4893-ba59-79cca5353cb0","Type":"ContainerStarted","Data":"ef0a80dd523cdb8b8412838670a24406e19b5458dcbc367eacf74e0e105e1ce4"} Nov 25 15:21:47 crc kubenswrapper[4890]: I1125 15:21:47.396766 4890 generic.go:334] "Generic (PLEG): container finished" podID="34bffacc-3ed2-4f5a-9338-34f1293b3776" containerID="be14c60cab33ab2ded2e1d36df0016a66348977ed17f64bc0f141bc63f78da6c" exitCode=0 Nov 25 15:21:47 crc kubenswrapper[4890]: I1125 15:21:47.396842 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58957f86ff-jbsf6" event={"ID":"34bffacc-3ed2-4f5a-9338-34f1293b3776","Type":"ContainerDied","Data":"be14c60cab33ab2ded2e1d36df0016a66348977ed17f64bc0f141bc63f78da6c"} Nov 25 15:21:47 crc kubenswrapper[4890]: I1125 15:21:47.408829 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"309052f6-3a50-436c-89e0-b97d5454b935","Type":"ContainerStarted","Data":"4b710264e7c07b2f106b8eb5f3d1aa1d54a2525abf9889d19b0273267e7ea964"} Nov 25 15:21:47 crc kubenswrapper[4890]: I1125 15:21:47.418016 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.417995042 podStartE2EDuration="4.417995042s" podCreationTimestamp="2025-11-25 15:21:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:21:47.407652632 +0000 UTC m=+1165.850115242" watchObservedRunningTime="2025-11-25 15:21:47.417995042 +0000 UTC m=+1165.860457652" Nov 25 15:21:47 crc kubenswrapper[4890]: I1125 15:21:47.469665 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 15:21:47 crc kubenswrapper[4890]: I1125 15:21:47.530964 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 15:21:47 crc kubenswrapper[4890]: I1125 15:21:47.631107 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58957f86ff-jbsf6" Nov 25 15:21:47 crc kubenswrapper[4890]: I1125 15:21:47.782437 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34bffacc-3ed2-4f5a-9338-34f1293b3776-config\") pod \"34bffacc-3ed2-4f5a-9338-34f1293b3776\" (UID: \"34bffacc-3ed2-4f5a-9338-34f1293b3776\") " Nov 25 15:21:47 crc kubenswrapper[4890]: I1125 15:21:47.782798 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/34bffacc-3ed2-4f5a-9338-34f1293b3776-dns-swift-storage-0\") pod \"34bffacc-3ed2-4f5a-9338-34f1293b3776\" (UID: \"34bffacc-3ed2-4f5a-9338-34f1293b3776\") " Nov 25 15:21:47 crc kubenswrapper[4890]: I1125 15:21:47.782972 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/34bffacc-3ed2-4f5a-9338-34f1293b3776-dns-svc\") pod \"34bffacc-3ed2-4f5a-9338-34f1293b3776\" (UID: \"34bffacc-3ed2-4f5a-9338-34f1293b3776\") " Nov 25 15:21:47 crc kubenswrapper[4890]: I1125 15:21:47.783040 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sfr6z\" (UniqueName: \"kubernetes.io/projected/34bffacc-3ed2-4f5a-9338-34f1293b3776-kube-api-access-sfr6z\") pod \"34bffacc-3ed2-4f5a-9338-34f1293b3776\" (UID: \"34bffacc-3ed2-4f5a-9338-34f1293b3776\") " Nov 25 15:21:47 crc kubenswrapper[4890]: I1125 15:21:47.783086 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/34bffacc-3ed2-4f5a-9338-34f1293b3776-ovsdbserver-nb\") pod \"34bffacc-3ed2-4f5a-9338-34f1293b3776\" (UID: \"34bffacc-3ed2-4f5a-9338-34f1293b3776\") " Nov 25 15:21:47 crc kubenswrapper[4890]: I1125 15:21:47.783143 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/34bffacc-3ed2-4f5a-9338-34f1293b3776-ovsdbserver-sb\") pod \"34bffacc-3ed2-4f5a-9338-34f1293b3776\" (UID: \"34bffacc-3ed2-4f5a-9338-34f1293b3776\") " Nov 25 15:21:47 crc kubenswrapper[4890]: I1125 15:21:47.810318 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34bffacc-3ed2-4f5a-9338-34f1293b3776-kube-api-access-sfr6z" (OuterVolumeSpecName: "kube-api-access-sfr6z") pod "34bffacc-3ed2-4f5a-9338-34f1293b3776" (UID: "34bffacc-3ed2-4f5a-9338-34f1293b3776"). InnerVolumeSpecName "kube-api-access-sfr6z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:21:47 crc kubenswrapper[4890]: I1125 15:21:47.886239 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sfr6z\" (UniqueName: \"kubernetes.io/projected/34bffacc-3ed2-4f5a-9338-34f1293b3776-kube-api-access-sfr6z\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:48 crc kubenswrapper[4890]: I1125 15:21:48.025033 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34bffacc-3ed2-4f5a-9338-34f1293b3776-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "34bffacc-3ed2-4f5a-9338-34f1293b3776" (UID: "34bffacc-3ed2-4f5a-9338-34f1293b3776"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:21:48 crc kubenswrapper[4890]: I1125 15:21:48.045587 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34bffacc-3ed2-4f5a-9338-34f1293b3776-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "34bffacc-3ed2-4f5a-9338-34f1293b3776" (UID: "34bffacc-3ed2-4f5a-9338-34f1293b3776"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:21:48 crc kubenswrapper[4890]: I1125 15:21:48.064274 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34bffacc-3ed2-4f5a-9338-34f1293b3776-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "34bffacc-3ed2-4f5a-9338-34f1293b3776" (UID: "34bffacc-3ed2-4f5a-9338-34f1293b3776"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:21:48 crc kubenswrapper[4890]: I1125 15:21:48.075004 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34bffacc-3ed2-4f5a-9338-34f1293b3776-config" (OuterVolumeSpecName: "config") pod "34bffacc-3ed2-4f5a-9338-34f1293b3776" (UID: "34bffacc-3ed2-4f5a-9338-34f1293b3776"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:21:48 crc kubenswrapper[4890]: I1125 15:21:48.101760 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34bffacc-3ed2-4f5a-9338-34f1293b3776-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "34bffacc-3ed2-4f5a-9338-34f1293b3776" (UID: "34bffacc-3ed2-4f5a-9338-34f1293b3776"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:21:48 crc kubenswrapper[4890]: I1125 15:21:48.103135 4890 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/34bffacc-3ed2-4f5a-9338-34f1293b3776-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:48 crc kubenswrapper[4890]: I1125 15:21:48.103186 4890 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/34bffacc-3ed2-4f5a-9338-34f1293b3776-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:48 crc kubenswrapper[4890]: I1125 15:21:48.103197 4890 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34bffacc-3ed2-4f5a-9338-34f1293b3776-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:48 crc kubenswrapper[4890]: I1125 15:21:48.103207 4890 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/34bffacc-3ed2-4f5a-9338-34f1293b3776-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:48 crc kubenswrapper[4890]: I1125 15:21:48.103217 4890 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/34bffacc-3ed2-4f5a-9338-34f1293b3776-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:48 crc kubenswrapper[4890]: I1125 15:21:48.190930 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47f833ef-ad6e-4a03-a1f0-9507d2d1f11e" path="/var/lib/kubelet/pods/47f833ef-ad6e-4a03-a1f0-9507d2d1f11e/volumes" Nov 25 15:21:48 crc kubenswrapper[4890]: I1125 15:21:48.433351 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58957f86ff-jbsf6" event={"ID":"34bffacc-3ed2-4f5a-9338-34f1293b3776","Type":"ContainerDied","Data":"20be58ca723b1cb56f2be0b27b47cdd9f8c3c7b5f5d8ae1c4588ecc2eb955cf1"} Nov 25 15:21:48 crc kubenswrapper[4890]: I1125 15:21:48.434582 4890 scope.go:117] "RemoveContainer" containerID="be14c60cab33ab2ded2e1d36df0016a66348977ed17f64bc0f141bc63f78da6c" Nov 25 15:21:48 crc kubenswrapper[4890]: I1125 15:21:48.436150 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58957f86ff-jbsf6" Nov 25 15:21:48 crc kubenswrapper[4890]: I1125 15:21:48.437062 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"309052f6-3a50-436c-89e0-b97d5454b935","Type":"ContainerStarted","Data":"49938bea770f6a06f61fb46ddc773fcafb02062ddc5d4d8a1b96bf8e387589ca"} Nov 25 15:21:48 crc kubenswrapper[4890]: I1125 15:21:48.455221 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9c617481-223e-4a84-8ecc-087c98dbd459","Type":"ContainerStarted","Data":"d9d8b5891c76a64c87b0fd714f93944ce2ed9cabc42ec7e8b0f58cbe0e85ab37"} Nov 25 15:21:48 crc kubenswrapper[4890]: I1125 15:21:48.455281 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9c617481-223e-4a84-8ecc-087c98dbd459","Type":"ContainerStarted","Data":"45cf7f441932fa6eea68ec0aa9debaec9863cf3b4281e1f4c6ab22034b596514"} Nov 25 15:21:48 crc kubenswrapper[4890]: I1125 15:21:48.455515 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="d65af5aa-095d-465f-8fd7-ba76b4ec08af" containerName="cinder-scheduler" containerID="cri-o://bd60a64ae0e4a9adde0fbb49c6acc7c6ce9c4b28a194e92a9791eacf02e81b52" gracePeriod=30 Nov 25 15:21:48 crc kubenswrapper[4890]: I1125 15:21:48.455904 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="d65af5aa-095d-465f-8fd7-ba76b4ec08af" containerName="probe" containerID="cri-o://0d6ac61429a65c1c7f8aaf11b9dbd4aa64709aa5aa05a214a384ec49ea22dd13" gracePeriod=30 Nov 25 15:21:48 crc kubenswrapper[4890]: I1125 15:21:48.492908 4890 scope.go:117] "RemoveContainer" containerID="c620d993985400412f84721f35e5c75fcad55b4c3d737d01500f6f2ad21a589b" Nov 25 15:21:48 crc kubenswrapper[4890]: I1125 15:21:48.497257 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58957f86ff-jbsf6"] Nov 25 15:21:48 crc kubenswrapper[4890]: I1125 15:21:48.526193 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-58957f86ff-jbsf6"] Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.058361 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-cwx2k"] Nov 25 15:21:49 crc kubenswrapper[4890]: E1125 15:21:49.059038 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34bffacc-3ed2-4f5a-9338-34f1293b3776" containerName="dnsmasq-dns" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.059055 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="34bffacc-3ed2-4f5a-9338-34f1293b3776" containerName="dnsmasq-dns" Nov 25 15:21:49 crc kubenswrapper[4890]: E1125 15:21:49.059069 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34bffacc-3ed2-4f5a-9338-34f1293b3776" containerName="init" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.059076 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="34bffacc-3ed2-4f5a-9338-34f1293b3776" containerName="init" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.059251 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="34bffacc-3ed2-4f5a-9338-34f1293b3776" containerName="dnsmasq-dns" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.059832 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-cwx2k" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.077097 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-cwx2k"] Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.160969 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-xmpnt"] Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.162036 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-xmpnt" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.176955 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-xmpnt"] Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.244312 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqg82\" (UniqueName: \"kubernetes.io/projected/0c71a662-55d3-46c8-8e7f-f8595bf24172-kube-api-access-rqg82\") pod \"nova-api-db-create-cwx2k\" (UID: \"0c71a662-55d3-46c8-8e7f-f8595bf24172\") " pod="openstack/nova-api-db-create-cwx2k" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.244405 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c71a662-55d3-46c8-8e7f-f8595bf24172-operator-scripts\") pod \"nova-api-db-create-cwx2k\" (UID: \"0c71a662-55d3-46c8-8e7f-f8595bf24172\") " pod="openstack/nova-api-db-create-cwx2k" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.280754 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-66a8-account-create-627gj"] Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.281860 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-66a8-account-create-627gj" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.285466 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.298810 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-66a8-account-create-627gj"] Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.347519 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60cb628f-f5b8-43ef-ab79-9753d39d4c04-operator-scripts\") pod \"nova-cell0-db-create-xmpnt\" (UID: \"60cb628f-f5b8-43ef-ab79-9753d39d4c04\") " pod="openstack/nova-cell0-db-create-xmpnt" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.347624 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqg82\" (UniqueName: \"kubernetes.io/projected/0c71a662-55d3-46c8-8e7f-f8595bf24172-kube-api-access-rqg82\") pod \"nova-api-db-create-cwx2k\" (UID: \"0c71a662-55d3-46c8-8e7f-f8595bf24172\") " pod="openstack/nova-api-db-create-cwx2k" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.347657 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bfecce7f-d8ed-44d9-bc97-aba2e5b19e8b-operator-scripts\") pod \"nova-api-66a8-account-create-627gj\" (UID: \"bfecce7f-d8ed-44d9-bc97-aba2e5b19e8b\") " pod="openstack/nova-api-66a8-account-create-627gj" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.347721 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbbg9\" (UniqueName: \"kubernetes.io/projected/60cb628f-f5b8-43ef-ab79-9753d39d4c04-kube-api-access-qbbg9\") pod \"nova-cell0-db-create-xmpnt\" (UID: \"60cb628f-f5b8-43ef-ab79-9753d39d4c04\") " pod="openstack/nova-cell0-db-create-xmpnt" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.347741 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c71a662-55d3-46c8-8e7f-f8595bf24172-operator-scripts\") pod \"nova-api-db-create-cwx2k\" (UID: \"0c71a662-55d3-46c8-8e7f-f8595bf24172\") " pod="openstack/nova-api-db-create-cwx2k" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.347778 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vqmd\" (UniqueName: \"kubernetes.io/projected/bfecce7f-d8ed-44d9-bc97-aba2e5b19e8b-kube-api-access-7vqmd\") pod \"nova-api-66a8-account-create-627gj\" (UID: \"bfecce7f-d8ed-44d9-bc97-aba2e5b19e8b\") " pod="openstack/nova-api-66a8-account-create-627gj" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.354625 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c71a662-55d3-46c8-8e7f-f8595bf24172-operator-scripts\") pod \"nova-api-db-create-cwx2k\" (UID: \"0c71a662-55d3-46c8-8e7f-f8595bf24172\") " pod="openstack/nova-api-db-create-cwx2k" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.382794 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqg82\" (UniqueName: \"kubernetes.io/projected/0c71a662-55d3-46c8-8e7f-f8595bf24172-kube-api-access-rqg82\") pod \"nova-api-db-create-cwx2k\" (UID: \"0c71a662-55d3-46c8-8e7f-f8595bf24172\") " pod="openstack/nova-api-db-create-cwx2k" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.446349 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-n5rj9"] Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.447946 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-n5rj9" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.449553 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bfecce7f-d8ed-44d9-bc97-aba2e5b19e8b-operator-scripts\") pod \"nova-api-66a8-account-create-627gj\" (UID: \"bfecce7f-d8ed-44d9-bc97-aba2e5b19e8b\") " pod="openstack/nova-api-66a8-account-create-627gj" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.449650 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbbg9\" (UniqueName: \"kubernetes.io/projected/60cb628f-f5b8-43ef-ab79-9753d39d4c04-kube-api-access-qbbg9\") pod \"nova-cell0-db-create-xmpnt\" (UID: \"60cb628f-f5b8-43ef-ab79-9753d39d4c04\") " pod="openstack/nova-cell0-db-create-xmpnt" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.449715 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vqmd\" (UniqueName: \"kubernetes.io/projected/bfecce7f-d8ed-44d9-bc97-aba2e5b19e8b-kube-api-access-7vqmd\") pod \"nova-api-66a8-account-create-627gj\" (UID: \"bfecce7f-d8ed-44d9-bc97-aba2e5b19e8b\") " pod="openstack/nova-api-66a8-account-create-627gj" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.449766 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60cb628f-f5b8-43ef-ab79-9753d39d4c04-operator-scripts\") pod \"nova-cell0-db-create-xmpnt\" (UID: \"60cb628f-f5b8-43ef-ab79-9753d39d4c04\") " pod="openstack/nova-cell0-db-create-xmpnt" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.450695 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60cb628f-f5b8-43ef-ab79-9753d39d4c04-operator-scripts\") pod \"nova-cell0-db-create-xmpnt\" (UID: \"60cb628f-f5b8-43ef-ab79-9753d39d4c04\") " pod="openstack/nova-cell0-db-create-xmpnt" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.452471 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bfecce7f-d8ed-44d9-bc97-aba2e5b19e8b-operator-scripts\") pod \"nova-api-66a8-account-create-627gj\" (UID: \"bfecce7f-d8ed-44d9-bc97-aba2e5b19e8b\") " pod="openstack/nova-api-66a8-account-create-627gj" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.482214 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-n5rj9"] Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.482381 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vqmd\" (UniqueName: \"kubernetes.io/projected/bfecce7f-d8ed-44d9-bc97-aba2e5b19e8b-kube-api-access-7vqmd\") pod \"nova-api-66a8-account-create-627gj\" (UID: \"bfecce7f-d8ed-44d9-bc97-aba2e5b19e8b\") " pod="openstack/nova-api-66a8-account-create-627gj" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.503984 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-66a8-account-create-627gj" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.509673 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-57c6-account-create-skvmk"] Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.511017 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-57c6-account-create-skvmk" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.513484 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.519203 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbbg9\" (UniqueName: \"kubernetes.io/projected/60cb628f-f5b8-43ef-ab79-9753d39d4c04-kube-api-access-qbbg9\") pod \"nova-cell0-db-create-xmpnt\" (UID: \"60cb628f-f5b8-43ef-ab79-9753d39d4c04\") " pod="openstack/nova-cell0-db-create-xmpnt" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.524172 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"309052f6-3a50-436c-89e0-b97d5454b935","Type":"ContainerStarted","Data":"2aa149409013c2e485cae439720b276edda14184399c65b3a309ffc19184264b"} Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.524933 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.537461 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-57c6-account-create-skvmk"] Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.548200 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9c617481-223e-4a84-8ecc-087c98dbd459","Type":"ContainerStarted","Data":"c4dd1483ae20cb8b9d34549a92ab50044ac7fb5a7e12deb52e257c1d9b63d209"} Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.551098 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dk275\" (UniqueName: \"kubernetes.io/projected/197b5965-ef39-4de5-9b7c-cb64c982c2a4-kube-api-access-dk275\") pod \"nova-cell1-db-create-n5rj9\" (UID: \"197b5965-ef39-4de5-9b7c-cb64c982c2a4\") " pod="openstack/nova-cell1-db-create-n5rj9" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.551242 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/197b5965-ef39-4de5-9b7c-cb64c982c2a4-operator-scripts\") pod \"nova-cell1-db-create-n5rj9\" (UID: \"197b5965-ef39-4de5-9b7c-cb64c982c2a4\") " pod="openstack/nova-cell1-db-create-n5rj9" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.571100 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.571077037 podStartE2EDuration="4.571077037s" podCreationTimestamp="2025-11-25 15:21:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:21:49.566577313 +0000 UTC m=+1168.009039923" watchObservedRunningTime="2025-11-25 15:21:49.571077037 +0000 UTC m=+1168.013539647" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.596236 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.59619783 podStartE2EDuration="3.59619783s" podCreationTimestamp="2025-11-25 15:21:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:21:49.586830404 +0000 UTC m=+1168.029293014" watchObservedRunningTime="2025-11-25 15:21:49.59619783 +0000 UTC m=+1168.038660430" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.596471 4890 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-8f5fb9544-r49m4" podUID="2f6fb686-a04a-40d5-a67a-caddba7e8d4d" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.156:9311/healthcheck\": read tcp 10.217.0.2:46436->10.217.0.156:9311: read: connection reset by peer" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.596404 4890 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-8f5fb9544-r49m4" podUID="2f6fb686-a04a-40d5-a67a-caddba7e8d4d" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.156:9311/healthcheck\": read tcp 10.217.0.2:46442->10.217.0.156:9311: read: connection reset by peer" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.652936 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/197b5965-ef39-4de5-9b7c-cb64c982c2a4-operator-scripts\") pod \"nova-cell1-db-create-n5rj9\" (UID: \"197b5965-ef39-4de5-9b7c-cb64c982c2a4\") " pod="openstack/nova-cell1-db-create-n5rj9" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.653329 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2d7p\" (UniqueName: \"kubernetes.io/projected/7bf4c0cd-f20d-4fd9-8077-e733e4196b62-kube-api-access-s2d7p\") pod \"nova-cell0-57c6-account-create-skvmk\" (UID: \"7bf4c0cd-f20d-4fd9-8077-e733e4196b62\") " pod="openstack/nova-cell0-57c6-account-create-skvmk" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.654380 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/197b5965-ef39-4de5-9b7c-cb64c982c2a4-operator-scripts\") pod \"nova-cell1-db-create-n5rj9\" (UID: \"197b5965-ef39-4de5-9b7c-cb64c982c2a4\") " pod="openstack/nova-cell1-db-create-n5rj9" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.654473 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7bf4c0cd-f20d-4fd9-8077-e733e4196b62-operator-scripts\") pod \"nova-cell0-57c6-account-create-skvmk\" (UID: \"7bf4c0cd-f20d-4fd9-8077-e733e4196b62\") " pod="openstack/nova-cell0-57c6-account-create-skvmk" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.654620 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dk275\" (UniqueName: \"kubernetes.io/projected/197b5965-ef39-4de5-9b7c-cb64c982c2a4-kube-api-access-dk275\") pod \"nova-cell1-db-create-n5rj9\" (UID: \"197b5965-ef39-4de5-9b7c-cb64c982c2a4\") " pod="openstack/nova-cell1-db-create-n5rj9" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.676421 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-d21a-account-create-nlvf4"] Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.678062 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-d21a-account-create-nlvf4" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.678515 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-cwx2k" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.683481 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.705117 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dk275\" (UniqueName: \"kubernetes.io/projected/197b5965-ef39-4de5-9b7c-cb64c982c2a4-kube-api-access-dk275\") pod \"nova-cell1-db-create-n5rj9\" (UID: \"197b5965-ef39-4de5-9b7c-cb64c982c2a4\") " pod="openstack/nova-cell1-db-create-n5rj9" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.725943 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-d21a-account-create-nlvf4"] Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.758076 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7bf4c0cd-f20d-4fd9-8077-e733e4196b62-operator-scripts\") pod \"nova-cell0-57c6-account-create-skvmk\" (UID: \"7bf4c0cd-f20d-4fd9-8077-e733e4196b62\") " pod="openstack/nova-cell0-57c6-account-create-skvmk" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.758254 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2d7p\" (UniqueName: \"kubernetes.io/projected/7bf4c0cd-f20d-4fd9-8077-e733e4196b62-kube-api-access-s2d7p\") pod \"nova-cell0-57c6-account-create-skvmk\" (UID: \"7bf4c0cd-f20d-4fd9-8077-e733e4196b62\") " pod="openstack/nova-cell0-57c6-account-create-skvmk" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.765762 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7bf4c0cd-f20d-4fd9-8077-e733e4196b62-operator-scripts\") pod \"nova-cell0-57c6-account-create-skvmk\" (UID: \"7bf4c0cd-f20d-4fd9-8077-e733e4196b62\") " pod="openstack/nova-cell0-57c6-account-create-skvmk" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.789486 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-xmpnt" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.797992 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2d7p\" (UniqueName: \"kubernetes.io/projected/7bf4c0cd-f20d-4fd9-8077-e733e4196b62-kube-api-access-s2d7p\") pod \"nova-cell0-57c6-account-create-skvmk\" (UID: \"7bf4c0cd-f20d-4fd9-8077-e733e4196b62\") " pod="openstack/nova-cell0-57c6-account-create-skvmk" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.852872 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-57c6-account-create-skvmk" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.856902 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-n5rj9" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.861444 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fea0bc8f-0181-4f5d-8e81-47bd6d89c4fb-operator-scripts\") pod \"nova-cell1-d21a-account-create-nlvf4\" (UID: \"fea0bc8f-0181-4f5d-8e81-47bd6d89c4fb\") " pod="openstack/nova-cell1-d21a-account-create-nlvf4" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.861519 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bm9rk\" (UniqueName: \"kubernetes.io/projected/fea0bc8f-0181-4f5d-8e81-47bd6d89c4fb-kube-api-access-bm9rk\") pod \"nova-cell1-d21a-account-create-nlvf4\" (UID: \"fea0bc8f-0181-4f5d-8e81-47bd6d89c4fb\") " pod="openstack/nova-cell1-d21a-account-create-nlvf4" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.965223 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fea0bc8f-0181-4f5d-8e81-47bd6d89c4fb-operator-scripts\") pod \"nova-cell1-d21a-account-create-nlvf4\" (UID: \"fea0bc8f-0181-4f5d-8e81-47bd6d89c4fb\") " pod="openstack/nova-cell1-d21a-account-create-nlvf4" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.965281 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bm9rk\" (UniqueName: \"kubernetes.io/projected/fea0bc8f-0181-4f5d-8e81-47bd6d89c4fb-kube-api-access-bm9rk\") pod \"nova-cell1-d21a-account-create-nlvf4\" (UID: \"fea0bc8f-0181-4f5d-8e81-47bd6d89c4fb\") " pod="openstack/nova-cell1-d21a-account-create-nlvf4" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.966484 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fea0bc8f-0181-4f5d-8e81-47bd6d89c4fb-operator-scripts\") pod \"nova-cell1-d21a-account-create-nlvf4\" (UID: \"fea0bc8f-0181-4f5d-8e81-47bd6d89c4fb\") " pod="openstack/nova-cell1-d21a-account-create-nlvf4" Nov 25 15:21:49 crc kubenswrapper[4890]: I1125 15:21:49.987379 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bm9rk\" (UniqueName: \"kubernetes.io/projected/fea0bc8f-0181-4f5d-8e81-47bd6d89c4fb-kube-api-access-bm9rk\") pod \"nova-cell1-d21a-account-create-nlvf4\" (UID: \"fea0bc8f-0181-4f5d-8e81-47bd6d89c4fb\") " pod="openstack/nova-cell1-d21a-account-create-nlvf4" Nov 25 15:21:50 crc kubenswrapper[4890]: I1125 15:21:50.171755 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-66a8-account-create-627gj"] Nov 25 15:21:50 crc kubenswrapper[4890]: W1125 15:21:50.200889 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbfecce7f_d8ed_44d9_bc97_aba2e5b19e8b.slice/crio-443c0761849e2c50bec2d6c41489a102946d7825b23b4b43fac15da1637758c6 WatchSource:0}: Error finding container 443c0761849e2c50bec2d6c41489a102946d7825b23b4b43fac15da1637758c6: Status 404 returned error can't find the container with id 443c0761849e2c50bec2d6c41489a102946d7825b23b4b43fac15da1637758c6 Nov 25 15:21:50 crc kubenswrapper[4890]: I1125 15:21:50.211012 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-d21a-account-create-nlvf4" Nov 25 15:21:50 crc kubenswrapper[4890]: I1125 15:21:50.214290 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34bffacc-3ed2-4f5a-9338-34f1293b3776" path="/var/lib/kubelet/pods/34bffacc-3ed2-4f5a-9338-34f1293b3776/volumes" Nov 25 15:21:50 crc kubenswrapper[4890]: I1125 15:21:50.341215 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-cwx2k"] Nov 25 15:21:50 crc kubenswrapper[4890]: I1125 15:21:50.476319 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-8f5fb9544-r49m4" Nov 25 15:21:50 crc kubenswrapper[4890]: I1125 15:21:50.566938 4890 generic.go:334] "Generic (PLEG): container finished" podID="2f6fb686-a04a-40d5-a67a-caddba7e8d4d" containerID="b1e404d377b43303ab6cf27c3866fde1baa533c9810625bcbf24c2695dfc0d89" exitCode=0 Nov 25 15:21:50 crc kubenswrapper[4890]: I1125 15:21:50.566978 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8f5fb9544-r49m4" event={"ID":"2f6fb686-a04a-40d5-a67a-caddba7e8d4d","Type":"ContainerDied","Data":"b1e404d377b43303ab6cf27c3866fde1baa533c9810625bcbf24c2695dfc0d89"} Nov 25 15:21:50 crc kubenswrapper[4890]: I1125 15:21:50.567037 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8f5fb9544-r49m4" event={"ID":"2f6fb686-a04a-40d5-a67a-caddba7e8d4d","Type":"ContainerDied","Data":"02082ee4a3ea4ab77848fb789b7ba14148aa57252500150eaf8c8ae2605bccf1"} Nov 25 15:21:50 crc kubenswrapper[4890]: I1125 15:21:50.567037 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-8f5fb9544-r49m4" Nov 25 15:21:50 crc kubenswrapper[4890]: I1125 15:21:50.567057 4890 scope.go:117] "RemoveContainer" containerID="b1e404d377b43303ab6cf27c3866fde1baa533c9810625bcbf24c2695dfc0d89" Nov 25 15:21:50 crc kubenswrapper[4890]: I1125 15:21:50.581756 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"1d4df90d-315a-4483-a6e5-c6e5b322db54","Type":"ContainerStarted","Data":"4fe3e37a9a408113316c5420215c491da8e87a16acb9f41bd76b96a43ce83974"} Nov 25 15:21:50 crc kubenswrapper[4890]: I1125 15:21:50.588505 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-66a8-account-create-627gj" event={"ID":"bfecce7f-d8ed-44d9-bc97-aba2e5b19e8b","Type":"ContainerStarted","Data":"443c0761849e2c50bec2d6c41489a102946d7825b23b4b43fac15da1637758c6"} Nov 25 15:21:50 crc kubenswrapper[4890]: I1125 15:21:50.603219 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-cwx2k" event={"ID":"0c71a662-55d3-46c8-8e7f-f8595bf24172","Type":"ContainerStarted","Data":"aee3266eaaff71ab4c9b4e697d6fd1e12e92228b97a0e8ca41959ca807c38d84"} Nov 25 15:21:50 crc kubenswrapper[4890]: I1125 15:21:50.612199 4890 generic.go:334] "Generic (PLEG): container finished" podID="d65af5aa-095d-465f-8fd7-ba76b4ec08af" containerID="0d6ac61429a65c1c7f8aaf11b9dbd4aa64709aa5aa05a214a384ec49ea22dd13" exitCode=0 Nov 25 15:21:50 crc kubenswrapper[4890]: I1125 15:21:50.612651 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d65af5aa-095d-465f-8fd7-ba76b4ec08af","Type":"ContainerDied","Data":"0d6ac61429a65c1c7f8aaf11b9dbd4aa64709aa5aa05a214a384ec49ea22dd13"} Nov 25 15:21:50 crc kubenswrapper[4890]: I1125 15:21:50.620370 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2f6fb686-a04a-40d5-a67a-caddba7e8d4d-config-data-custom\") pod \"2f6fb686-a04a-40d5-a67a-caddba7e8d4d\" (UID: \"2f6fb686-a04a-40d5-a67a-caddba7e8d4d\") " Nov 25 15:21:50 crc kubenswrapper[4890]: I1125 15:21:50.625632 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2v8rh\" (UniqueName: \"kubernetes.io/projected/2f6fb686-a04a-40d5-a67a-caddba7e8d4d-kube-api-access-2v8rh\") pod \"2f6fb686-a04a-40d5-a67a-caddba7e8d4d\" (UID: \"2f6fb686-a04a-40d5-a67a-caddba7e8d4d\") " Nov 25 15:21:50 crc kubenswrapper[4890]: I1125 15:21:50.625775 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f6fb686-a04a-40d5-a67a-caddba7e8d4d-config-data\") pod \"2f6fb686-a04a-40d5-a67a-caddba7e8d4d\" (UID: \"2f6fb686-a04a-40d5-a67a-caddba7e8d4d\") " Nov 25 15:21:50 crc kubenswrapper[4890]: I1125 15:21:50.625889 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f6fb686-a04a-40d5-a67a-caddba7e8d4d-logs\") pod \"2f6fb686-a04a-40d5-a67a-caddba7e8d4d\" (UID: \"2f6fb686-a04a-40d5-a67a-caddba7e8d4d\") " Nov 25 15:21:50 crc kubenswrapper[4890]: I1125 15:21:50.625965 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f6fb686-a04a-40d5-a67a-caddba7e8d4d-combined-ca-bundle\") pod \"2f6fb686-a04a-40d5-a67a-caddba7e8d4d\" (UID: \"2f6fb686-a04a-40d5-a67a-caddba7e8d4d\") " Nov 25 15:21:50 crc kubenswrapper[4890]: I1125 15:21:50.627732 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f6fb686-a04a-40d5-a67a-caddba7e8d4d-logs" (OuterVolumeSpecName: "logs") pod "2f6fb686-a04a-40d5-a67a-caddba7e8d4d" (UID: "2f6fb686-a04a-40d5-a67a-caddba7e8d4d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:21:50 crc kubenswrapper[4890]: I1125 15:21:50.632685 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f6fb686-a04a-40d5-a67a-caddba7e8d4d-kube-api-access-2v8rh" (OuterVolumeSpecName: "kube-api-access-2v8rh") pod "2f6fb686-a04a-40d5-a67a-caddba7e8d4d" (UID: "2f6fb686-a04a-40d5-a67a-caddba7e8d4d"). InnerVolumeSpecName "kube-api-access-2v8rh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:21:50 crc kubenswrapper[4890]: I1125 15:21:50.647398 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f6fb686-a04a-40d5-a67a-caddba7e8d4d-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "2f6fb686-a04a-40d5-a67a-caddba7e8d4d" (UID: "2f6fb686-a04a-40d5-a67a-caddba7e8d4d"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:21:50 crc kubenswrapper[4890]: I1125 15:21:50.648700 4890 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f6fb686-a04a-40d5-a67a-caddba7e8d4d-logs\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:50 crc kubenswrapper[4890]: I1125 15:21:50.648801 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-xmpnt"] Nov 25 15:21:50 crc kubenswrapper[4890]: I1125 15:21:50.649404 4890 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2f6fb686-a04a-40d5-a67a-caddba7e8d4d-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:50 crc kubenswrapper[4890]: I1125 15:21:50.649499 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2v8rh\" (UniqueName: \"kubernetes.io/projected/2f6fb686-a04a-40d5-a67a-caddba7e8d4d-kube-api-access-2v8rh\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:50 crc kubenswrapper[4890]: I1125 15:21:50.667849 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.693930583 podStartE2EDuration="35.667832617s" podCreationTimestamp="2025-11-25 15:21:15 +0000 UTC" firstStartedPulling="2025-11-25 15:21:16.268608465 +0000 UTC m=+1134.711071075" lastFinishedPulling="2025-11-25 15:21:49.242510499 +0000 UTC m=+1167.684973109" observedRunningTime="2025-11-25 15:21:50.610553254 +0000 UTC m=+1169.053015864" watchObservedRunningTime="2025-11-25 15:21:50.667832617 +0000 UTC m=+1169.110295227" Nov 25 15:21:50 crc kubenswrapper[4890]: I1125 15:21:50.680964 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f6fb686-a04a-40d5-a67a-caddba7e8d4d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2f6fb686-a04a-40d5-a67a-caddba7e8d4d" (UID: "2f6fb686-a04a-40d5-a67a-caddba7e8d4d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:21:50 crc kubenswrapper[4890]: I1125 15:21:50.690977 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-57c6-account-create-skvmk"] Nov 25 15:21:50 crc kubenswrapper[4890]: I1125 15:21:50.713153 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f6fb686-a04a-40d5-a67a-caddba7e8d4d-config-data" (OuterVolumeSpecName: "config-data") pod "2f6fb686-a04a-40d5-a67a-caddba7e8d4d" (UID: "2f6fb686-a04a-40d5-a67a-caddba7e8d4d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:21:50 crc kubenswrapper[4890]: I1125 15:21:50.754765 4890 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f6fb686-a04a-40d5-a67a-caddba7e8d4d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:50 crc kubenswrapper[4890]: I1125 15:21:50.754789 4890 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f6fb686-a04a-40d5-a67a-caddba7e8d4d-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:50 crc kubenswrapper[4890]: I1125 15:21:50.755864 4890 scope.go:117] "RemoveContainer" containerID="a1b729f7f24bb522a987cb0b04c7b9a5b52df51799327aa3e4d15458ec94d95c" Nov 25 15:21:50 crc kubenswrapper[4890]: I1125 15:21:50.757418 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-n5rj9"] Nov 25 15:21:50 crc kubenswrapper[4890]: W1125 15:21:50.820423 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod197b5965_ef39_4de5_9b7c_cb64c982c2a4.slice/crio-66a95703f88e757681686fa024ce41052d98dfacb6596b261f5d13aed02229d5 WatchSource:0}: Error finding container 66a95703f88e757681686fa024ce41052d98dfacb6596b261f5d13aed02229d5: Status 404 returned error can't find the container with id 66a95703f88e757681686fa024ce41052d98dfacb6596b261f5d13aed02229d5 Nov 25 15:21:50 crc kubenswrapper[4890]: I1125 15:21:50.895636 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-d21a-account-create-nlvf4"] Nov 25 15:21:50 crc kubenswrapper[4890]: I1125 15:21:50.916548 4890 scope.go:117] "RemoveContainer" containerID="b1e404d377b43303ab6cf27c3866fde1baa533c9810625bcbf24c2695dfc0d89" Nov 25 15:21:50 crc kubenswrapper[4890]: E1125 15:21:50.919812 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1e404d377b43303ab6cf27c3866fde1baa533c9810625bcbf24c2695dfc0d89\": container with ID starting with b1e404d377b43303ab6cf27c3866fde1baa533c9810625bcbf24c2695dfc0d89 not found: ID does not exist" containerID="b1e404d377b43303ab6cf27c3866fde1baa533c9810625bcbf24c2695dfc0d89" Nov 25 15:21:50 crc kubenswrapper[4890]: I1125 15:21:50.919864 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1e404d377b43303ab6cf27c3866fde1baa533c9810625bcbf24c2695dfc0d89"} err="failed to get container status \"b1e404d377b43303ab6cf27c3866fde1baa533c9810625bcbf24c2695dfc0d89\": rpc error: code = NotFound desc = could not find container \"b1e404d377b43303ab6cf27c3866fde1baa533c9810625bcbf24c2695dfc0d89\": container with ID starting with b1e404d377b43303ab6cf27c3866fde1baa533c9810625bcbf24c2695dfc0d89 not found: ID does not exist" Nov 25 15:21:50 crc kubenswrapper[4890]: I1125 15:21:50.919893 4890 scope.go:117] "RemoveContainer" containerID="a1b729f7f24bb522a987cb0b04c7b9a5b52df51799327aa3e4d15458ec94d95c" Nov 25 15:21:50 crc kubenswrapper[4890]: E1125 15:21:50.920632 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1b729f7f24bb522a987cb0b04c7b9a5b52df51799327aa3e4d15458ec94d95c\": container with ID starting with a1b729f7f24bb522a987cb0b04c7b9a5b52df51799327aa3e4d15458ec94d95c not found: ID does not exist" containerID="a1b729f7f24bb522a987cb0b04c7b9a5b52df51799327aa3e4d15458ec94d95c" Nov 25 15:21:50 crc kubenswrapper[4890]: I1125 15:21:50.920746 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1b729f7f24bb522a987cb0b04c7b9a5b52df51799327aa3e4d15458ec94d95c"} err="failed to get container status \"a1b729f7f24bb522a987cb0b04c7b9a5b52df51799327aa3e4d15458ec94d95c\": rpc error: code = NotFound desc = could not find container \"a1b729f7f24bb522a987cb0b04c7b9a5b52df51799327aa3e4d15458ec94d95c\": container with ID starting with a1b729f7f24bb522a987cb0b04c7b9a5b52df51799327aa3e4d15458ec94d95c not found: ID does not exist" Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.039905 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-8f5fb9544-r49m4"] Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.085442 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-8f5fb9544-r49m4"] Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.152717 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.279879 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d65af5aa-095d-465f-8fd7-ba76b4ec08af-etc-machine-id\") pod \"d65af5aa-095d-465f-8fd7-ba76b4ec08af\" (UID: \"d65af5aa-095d-465f-8fd7-ba76b4ec08af\") " Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.279947 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q8fkv\" (UniqueName: \"kubernetes.io/projected/d65af5aa-095d-465f-8fd7-ba76b4ec08af-kube-api-access-q8fkv\") pod \"d65af5aa-095d-465f-8fd7-ba76b4ec08af\" (UID: \"d65af5aa-095d-465f-8fd7-ba76b4ec08af\") " Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.279982 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d65af5aa-095d-465f-8fd7-ba76b4ec08af-config-data\") pod \"d65af5aa-095d-465f-8fd7-ba76b4ec08af\" (UID: \"d65af5aa-095d-465f-8fd7-ba76b4ec08af\") " Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.280017 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d65af5aa-095d-465f-8fd7-ba76b4ec08af-combined-ca-bundle\") pod \"d65af5aa-095d-465f-8fd7-ba76b4ec08af\" (UID: \"d65af5aa-095d-465f-8fd7-ba76b4ec08af\") " Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.280093 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d65af5aa-095d-465f-8fd7-ba76b4ec08af-scripts\") pod \"d65af5aa-095d-465f-8fd7-ba76b4ec08af\" (UID: \"d65af5aa-095d-465f-8fd7-ba76b4ec08af\") " Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.280198 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d65af5aa-095d-465f-8fd7-ba76b4ec08af-config-data-custom\") pod \"d65af5aa-095d-465f-8fd7-ba76b4ec08af\" (UID: \"d65af5aa-095d-465f-8fd7-ba76b4ec08af\") " Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.281479 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d65af5aa-095d-465f-8fd7-ba76b4ec08af-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "d65af5aa-095d-465f-8fd7-ba76b4ec08af" (UID: "d65af5aa-095d-465f-8fd7-ba76b4ec08af"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.288819 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d65af5aa-095d-465f-8fd7-ba76b4ec08af-kube-api-access-q8fkv" (OuterVolumeSpecName: "kube-api-access-q8fkv") pod "d65af5aa-095d-465f-8fd7-ba76b4ec08af" (UID: "d65af5aa-095d-465f-8fd7-ba76b4ec08af"). InnerVolumeSpecName "kube-api-access-q8fkv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.297369 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d65af5aa-095d-465f-8fd7-ba76b4ec08af-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d65af5aa-095d-465f-8fd7-ba76b4ec08af" (UID: "d65af5aa-095d-465f-8fd7-ba76b4ec08af"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.307770 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d65af5aa-095d-465f-8fd7-ba76b4ec08af-scripts" (OuterVolumeSpecName: "scripts") pod "d65af5aa-095d-465f-8fd7-ba76b4ec08af" (UID: "d65af5aa-095d-465f-8fd7-ba76b4ec08af"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.371368 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d65af5aa-095d-465f-8fd7-ba76b4ec08af-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d65af5aa-095d-465f-8fd7-ba76b4ec08af" (UID: "d65af5aa-095d-465f-8fd7-ba76b4ec08af"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.382487 4890 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d65af5aa-095d-465f-8fd7-ba76b4ec08af-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.382523 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q8fkv\" (UniqueName: \"kubernetes.io/projected/d65af5aa-095d-465f-8fd7-ba76b4ec08af-kube-api-access-q8fkv\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.382534 4890 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d65af5aa-095d-465f-8fd7-ba76b4ec08af-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.382543 4890 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d65af5aa-095d-465f-8fd7-ba76b4ec08af-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.382553 4890 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d65af5aa-095d-465f-8fd7-ba76b4ec08af-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.432266 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d65af5aa-095d-465f-8fd7-ba76b4ec08af-config-data" (OuterVolumeSpecName: "config-data") pod "d65af5aa-095d-465f-8fd7-ba76b4ec08af" (UID: "d65af5aa-095d-465f-8fd7-ba76b4ec08af"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.484082 4890 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d65af5aa-095d-465f-8fd7-ba76b4ec08af-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.621336 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-xmpnt" event={"ID":"60cb628f-f5b8-43ef-ab79-9753d39d4c04","Type":"ContainerStarted","Data":"968f71594fbf195538b3495381442d82ebb8bc7cbd156a37ede4596b9ed58f09"} Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.621379 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-xmpnt" event={"ID":"60cb628f-f5b8-43ef-ab79-9753d39d4c04","Type":"ContainerStarted","Data":"df04e86f5afad6d7afd7166efd78efe982eb3be0ce3bf1f1e5b327a1cebbf11c"} Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.622473 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-cwx2k" event={"ID":"0c71a662-55d3-46c8-8e7f-f8595bf24172","Type":"ContainerStarted","Data":"649bba0742590fe6749edc35bcbbb56579c0de557a05e2c389b08b8f5102f2d6"} Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.625145 4890 generic.go:334] "Generic (PLEG): container finished" podID="d65af5aa-095d-465f-8fd7-ba76b4ec08af" containerID="bd60a64ae0e4a9adde0fbb49c6acc7c6ce9c4b28a194e92a9791eacf02e81b52" exitCode=0 Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.625208 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d65af5aa-095d-465f-8fd7-ba76b4ec08af","Type":"ContainerDied","Data":"bd60a64ae0e4a9adde0fbb49c6acc7c6ce9c4b28a194e92a9791eacf02e81b52"} Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.625250 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d65af5aa-095d-465f-8fd7-ba76b4ec08af","Type":"ContainerDied","Data":"87a8b8c8de7fa3af02a2134b3a7e4fbc1de8ac4f257df8e52aad1d9f76d366a4"} Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.625273 4890 scope.go:117] "RemoveContainer" containerID="0d6ac61429a65c1c7f8aaf11b9dbd4aa64709aa5aa05a214a384ec49ea22dd13" Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.625248 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.631027 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-d21a-account-create-nlvf4" event={"ID":"fea0bc8f-0181-4f5d-8e81-47bd6d89c4fb","Type":"ContainerStarted","Data":"3341eafe5523ba14f01d0cde11d921109ba1d96008b5d6c483a72bdd45b3f430"} Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.631077 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-d21a-account-create-nlvf4" event={"ID":"fea0bc8f-0181-4f5d-8e81-47bd6d89c4fb","Type":"ContainerStarted","Data":"764803b46f953511703c6456d7d9809eae6daf79a370b150ab2c90c4c88d0ae5"} Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.632842 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-66a8-account-create-627gj" event={"ID":"bfecce7f-d8ed-44d9-bc97-aba2e5b19e8b","Type":"ContainerStarted","Data":"6e7f48c14398dcfdf4962151c72cc993d0cff6a4d0863894b039abf0f5055f69"} Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.637262 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-n5rj9" event={"ID":"197b5965-ef39-4de5-9b7c-cb64c982c2a4","Type":"ContainerStarted","Data":"27b4254c2dd7539a7c6185ff497de399703ce0e26dee37b3760a3ea080dd76a5"} Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.637607 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-n5rj9" event={"ID":"197b5965-ef39-4de5-9b7c-cb64c982c2a4","Type":"ContainerStarted","Data":"66a95703f88e757681686fa024ce41052d98dfacb6596b261f5d13aed02229d5"} Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.650417 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-xmpnt" podStartSLOduration=2.650400721 podStartE2EDuration="2.650400721s" podCreationTimestamp="2025-11-25 15:21:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:21:51.639415695 +0000 UTC m=+1170.081878305" watchObservedRunningTime="2025-11-25 15:21:51.650400721 +0000 UTC m=+1170.092863331" Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.656557 4890 scope.go:117] "RemoveContainer" containerID="bd60a64ae0e4a9adde0fbb49c6acc7c6ce9c4b28a194e92a9791eacf02e81b52" Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.658100 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-57c6-account-create-skvmk" event={"ID":"7bf4c0cd-f20d-4fd9-8077-e733e4196b62","Type":"ContainerStarted","Data":"eb87dca0ce22763a819779b239a159b8c0b39406bd4a1afc6dbea87a86ff28c9"} Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.658146 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-57c6-account-create-skvmk" event={"ID":"7bf4c0cd-f20d-4fd9-8077-e733e4196b62","Type":"ContainerStarted","Data":"73cf59b4c0c9be16f79a3ae6dbc9af01360e0a2933c223a9d0b349a457ed06ab"} Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.682296 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-cwx2k" podStartSLOduration=2.682274482 podStartE2EDuration="2.682274482s" podCreationTimestamp="2025-11-25 15:21:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:21:51.652001171 +0000 UTC m=+1170.094463781" watchObservedRunningTime="2025-11-25 15:21:51.682274482 +0000 UTC m=+1170.124737092" Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.689216 4890 scope.go:117] "RemoveContainer" containerID="0d6ac61429a65c1c7f8aaf11b9dbd4aa64709aa5aa05a214a384ec49ea22dd13" Nov 25 15:21:51 crc kubenswrapper[4890]: E1125 15:21:51.690070 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d6ac61429a65c1c7f8aaf11b9dbd4aa64709aa5aa05a214a384ec49ea22dd13\": container with ID starting with 0d6ac61429a65c1c7f8aaf11b9dbd4aa64709aa5aa05a214a384ec49ea22dd13 not found: ID does not exist" containerID="0d6ac61429a65c1c7f8aaf11b9dbd4aa64709aa5aa05a214a384ec49ea22dd13" Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.690105 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d6ac61429a65c1c7f8aaf11b9dbd4aa64709aa5aa05a214a384ec49ea22dd13"} err="failed to get container status \"0d6ac61429a65c1c7f8aaf11b9dbd4aa64709aa5aa05a214a384ec49ea22dd13\": rpc error: code = NotFound desc = could not find container \"0d6ac61429a65c1c7f8aaf11b9dbd4aa64709aa5aa05a214a384ec49ea22dd13\": container with ID starting with 0d6ac61429a65c1c7f8aaf11b9dbd4aa64709aa5aa05a214a384ec49ea22dd13 not found: ID does not exist" Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.690129 4890 scope.go:117] "RemoveContainer" containerID="bd60a64ae0e4a9adde0fbb49c6acc7c6ce9c4b28a194e92a9791eacf02e81b52" Nov 25 15:21:51 crc kubenswrapper[4890]: E1125 15:21:51.690487 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd60a64ae0e4a9adde0fbb49c6acc7c6ce9c4b28a194e92a9791eacf02e81b52\": container with ID starting with bd60a64ae0e4a9adde0fbb49c6acc7c6ce9c4b28a194e92a9791eacf02e81b52 not found: ID does not exist" containerID="bd60a64ae0e4a9adde0fbb49c6acc7c6ce9c4b28a194e92a9791eacf02e81b52" Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.690546 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd60a64ae0e4a9adde0fbb49c6acc7c6ce9c4b28a194e92a9791eacf02e81b52"} err="failed to get container status \"bd60a64ae0e4a9adde0fbb49c6acc7c6ce9c4b28a194e92a9791eacf02e81b52\": rpc error: code = NotFound desc = could not find container \"bd60a64ae0e4a9adde0fbb49c6acc7c6ce9c4b28a194e92a9791eacf02e81b52\": container with ID starting with bd60a64ae0e4a9adde0fbb49c6acc7c6ce9c4b28a194e92a9791eacf02e81b52 not found: ID does not exist" Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.692564 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-66a8-account-create-627gj" podStartSLOduration=2.69254988 podStartE2EDuration="2.69254988s" podCreationTimestamp="2025-11-25 15:21:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:21:51.6722501 +0000 UTC m=+1170.114712720" watchObservedRunningTime="2025-11-25 15:21:51.69254988 +0000 UTC m=+1170.135012490" Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.694351 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-n5rj9" podStartSLOduration=2.694340335 podStartE2EDuration="2.694340335s" podCreationTimestamp="2025-11-25 15:21:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:21:51.686301813 +0000 UTC m=+1170.128764423" watchObservedRunningTime="2025-11-25 15:21:51.694340335 +0000 UTC m=+1170.136802945" Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.715388 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-d21a-account-create-nlvf4" podStartSLOduration=2.715369173 podStartE2EDuration="2.715369173s" podCreationTimestamp="2025-11-25 15:21:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:21:51.702712605 +0000 UTC m=+1170.145175225" watchObservedRunningTime="2025-11-25 15:21:51.715369173 +0000 UTC m=+1170.157831783" Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.770809 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.778930 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.785063 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-57c6-account-create-skvmk" podStartSLOduration=2.785040354 podStartE2EDuration="2.785040354s" podCreationTimestamp="2025-11-25 15:21:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:21:51.748993758 +0000 UTC m=+1170.191456368" watchObservedRunningTime="2025-11-25 15:21:51.785040354 +0000 UTC m=+1170.227502964" Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.798856 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 15:21:51 crc kubenswrapper[4890]: E1125 15:21:51.799386 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d65af5aa-095d-465f-8fd7-ba76b4ec08af" containerName="probe" Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.799411 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="d65af5aa-095d-465f-8fd7-ba76b4ec08af" containerName="probe" Nov 25 15:21:51 crc kubenswrapper[4890]: E1125 15:21:51.799437 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f6fb686-a04a-40d5-a67a-caddba7e8d4d" containerName="barbican-api" Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.799447 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f6fb686-a04a-40d5-a67a-caddba7e8d4d" containerName="barbican-api" Nov 25 15:21:51 crc kubenswrapper[4890]: E1125 15:21:51.799465 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f6fb686-a04a-40d5-a67a-caddba7e8d4d" containerName="barbican-api-log" Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.799475 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f6fb686-a04a-40d5-a67a-caddba7e8d4d" containerName="barbican-api-log" Nov 25 15:21:51 crc kubenswrapper[4890]: E1125 15:21:51.799492 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d65af5aa-095d-465f-8fd7-ba76b4ec08af" containerName="cinder-scheduler" Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.799501 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="d65af5aa-095d-465f-8fd7-ba76b4ec08af" containerName="cinder-scheduler" Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.806611 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="d65af5aa-095d-465f-8fd7-ba76b4ec08af" containerName="cinder-scheduler" Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.806662 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f6fb686-a04a-40d5-a67a-caddba7e8d4d" containerName="barbican-api-log" Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.806706 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f6fb686-a04a-40d5-a67a-caddba7e8d4d" containerName="barbican-api" Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.806722 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="d65af5aa-095d-465f-8fd7-ba76b4ec08af" containerName="probe" Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.807943 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.808701 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.810097 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.996589 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb86ab7c-e51d-46ad-a1d6-6a7f5ebabc41-config-data\") pod \"cinder-scheduler-0\" (UID: \"bb86ab7c-e51d-46ad-a1d6-6a7f5ebabc41\") " pod="openstack/cinder-scheduler-0" Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.996863 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8m444\" (UniqueName: \"kubernetes.io/projected/bb86ab7c-e51d-46ad-a1d6-6a7f5ebabc41-kube-api-access-8m444\") pod \"cinder-scheduler-0\" (UID: \"bb86ab7c-e51d-46ad-a1d6-6a7f5ebabc41\") " pod="openstack/cinder-scheduler-0" Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.996990 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bb86ab7c-e51d-46ad-a1d6-6a7f5ebabc41-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"bb86ab7c-e51d-46ad-a1d6-6a7f5ebabc41\") " pod="openstack/cinder-scheduler-0" Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.997110 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bb86ab7c-e51d-46ad-a1d6-6a7f5ebabc41-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"bb86ab7c-e51d-46ad-a1d6-6a7f5ebabc41\") " pod="openstack/cinder-scheduler-0" Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.997231 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb86ab7c-e51d-46ad-a1d6-6a7f5ebabc41-scripts\") pod \"cinder-scheduler-0\" (UID: \"bb86ab7c-e51d-46ad-a1d6-6a7f5ebabc41\") " pod="openstack/cinder-scheduler-0" Nov 25 15:21:51 crc kubenswrapper[4890]: I1125 15:21:51.997368 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb86ab7c-e51d-46ad-a1d6-6a7f5ebabc41-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"bb86ab7c-e51d-46ad-a1d6-6a7f5ebabc41\") " pod="openstack/cinder-scheduler-0" Nov 25 15:21:52 crc kubenswrapper[4890]: I1125 15:21:52.099542 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bb86ab7c-e51d-46ad-a1d6-6a7f5ebabc41-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"bb86ab7c-e51d-46ad-a1d6-6a7f5ebabc41\") " pod="openstack/cinder-scheduler-0" Nov 25 15:21:52 crc kubenswrapper[4890]: I1125 15:21:52.099596 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb86ab7c-e51d-46ad-a1d6-6a7f5ebabc41-scripts\") pod \"cinder-scheduler-0\" (UID: \"bb86ab7c-e51d-46ad-a1d6-6a7f5ebabc41\") " pod="openstack/cinder-scheduler-0" Nov 25 15:21:52 crc kubenswrapper[4890]: I1125 15:21:52.099623 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb86ab7c-e51d-46ad-a1d6-6a7f5ebabc41-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"bb86ab7c-e51d-46ad-a1d6-6a7f5ebabc41\") " pod="openstack/cinder-scheduler-0" Nov 25 15:21:52 crc kubenswrapper[4890]: I1125 15:21:52.099689 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bb86ab7c-e51d-46ad-a1d6-6a7f5ebabc41-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"bb86ab7c-e51d-46ad-a1d6-6a7f5ebabc41\") " pod="openstack/cinder-scheduler-0" Nov 25 15:21:52 crc kubenswrapper[4890]: I1125 15:21:52.099791 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb86ab7c-e51d-46ad-a1d6-6a7f5ebabc41-config-data\") pod \"cinder-scheduler-0\" (UID: \"bb86ab7c-e51d-46ad-a1d6-6a7f5ebabc41\") " pod="openstack/cinder-scheduler-0" Nov 25 15:21:52 crc kubenswrapper[4890]: I1125 15:21:52.099949 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8m444\" (UniqueName: \"kubernetes.io/projected/bb86ab7c-e51d-46ad-a1d6-6a7f5ebabc41-kube-api-access-8m444\") pod \"cinder-scheduler-0\" (UID: \"bb86ab7c-e51d-46ad-a1d6-6a7f5ebabc41\") " pod="openstack/cinder-scheduler-0" Nov 25 15:21:52 crc kubenswrapper[4890]: I1125 15:21:52.100039 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bb86ab7c-e51d-46ad-a1d6-6a7f5ebabc41-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"bb86ab7c-e51d-46ad-a1d6-6a7f5ebabc41\") " pod="openstack/cinder-scheduler-0" Nov 25 15:21:52 crc kubenswrapper[4890]: I1125 15:21:52.103855 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb86ab7c-e51d-46ad-a1d6-6a7f5ebabc41-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"bb86ab7c-e51d-46ad-a1d6-6a7f5ebabc41\") " pod="openstack/cinder-scheduler-0" Nov 25 15:21:52 crc kubenswrapper[4890]: I1125 15:21:52.104685 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bb86ab7c-e51d-46ad-a1d6-6a7f5ebabc41-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"bb86ab7c-e51d-46ad-a1d6-6a7f5ebabc41\") " pod="openstack/cinder-scheduler-0" Nov 25 15:21:52 crc kubenswrapper[4890]: I1125 15:21:52.105408 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb86ab7c-e51d-46ad-a1d6-6a7f5ebabc41-config-data\") pod \"cinder-scheduler-0\" (UID: \"bb86ab7c-e51d-46ad-a1d6-6a7f5ebabc41\") " pod="openstack/cinder-scheduler-0" Nov 25 15:21:52 crc kubenswrapper[4890]: I1125 15:21:52.106645 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb86ab7c-e51d-46ad-a1d6-6a7f5ebabc41-scripts\") pod \"cinder-scheduler-0\" (UID: \"bb86ab7c-e51d-46ad-a1d6-6a7f5ebabc41\") " pod="openstack/cinder-scheduler-0" Nov 25 15:21:52 crc kubenswrapper[4890]: I1125 15:21:52.124510 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8m444\" (UniqueName: \"kubernetes.io/projected/bb86ab7c-e51d-46ad-a1d6-6a7f5ebabc41-kube-api-access-8m444\") pod \"cinder-scheduler-0\" (UID: \"bb86ab7c-e51d-46ad-a1d6-6a7f5ebabc41\") " pod="openstack/cinder-scheduler-0" Nov 25 15:21:52 crc kubenswrapper[4890]: I1125 15:21:52.149787 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 25 15:21:52 crc kubenswrapper[4890]: I1125 15:21:52.274151 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f6fb686-a04a-40d5-a67a-caddba7e8d4d" path="/var/lib/kubelet/pods/2f6fb686-a04a-40d5-a67a-caddba7e8d4d/volumes" Nov 25 15:21:52 crc kubenswrapper[4890]: I1125 15:21:52.274732 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d65af5aa-095d-465f-8fd7-ba76b4ec08af" path="/var/lib/kubelet/pods/d65af5aa-095d-465f-8fd7-ba76b4ec08af/volumes" Nov 25 15:21:52 crc kubenswrapper[4890]: I1125 15:21:52.465548 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 15:21:52 crc kubenswrapper[4890]: I1125 15:21:52.466441 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="59f62350-c15f-4893-ba59-79cca5353cb0" containerName="glance-httpd" containerID="cri-o://ef0a80dd523cdb8b8412838670a24406e19b5458dcbc367eacf74e0e105e1ce4" gracePeriod=30 Nov 25 15:21:52 crc kubenswrapper[4890]: I1125 15:21:52.466341 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="59f62350-c15f-4893-ba59-79cca5353cb0" containerName="glance-log" containerID="cri-o://6c3312722a7e81ef35562577700748b4a13d43c509d835aca0dc24f388927ae9" gracePeriod=30 Nov 25 15:21:52 crc kubenswrapper[4890]: I1125 15:21:52.694241 4890 generic.go:334] "Generic (PLEG): container finished" podID="59f62350-c15f-4893-ba59-79cca5353cb0" containerID="6c3312722a7e81ef35562577700748b4a13d43c509d835aca0dc24f388927ae9" exitCode=143 Nov 25 15:21:52 crc kubenswrapper[4890]: I1125 15:21:52.694369 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"59f62350-c15f-4893-ba59-79cca5353cb0","Type":"ContainerDied","Data":"6c3312722a7e81ef35562577700748b4a13d43c509d835aca0dc24f388927ae9"} Nov 25 15:21:52 crc kubenswrapper[4890]: I1125 15:21:52.715144 4890 generic.go:334] "Generic (PLEG): container finished" podID="fea0bc8f-0181-4f5d-8e81-47bd6d89c4fb" containerID="3341eafe5523ba14f01d0cde11d921109ba1d96008b5d6c483a72bdd45b3f430" exitCode=0 Nov 25 15:21:52 crc kubenswrapper[4890]: I1125 15:21:52.715729 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-d21a-account-create-nlvf4" event={"ID":"fea0bc8f-0181-4f5d-8e81-47bd6d89c4fb","Type":"ContainerDied","Data":"3341eafe5523ba14f01d0cde11d921109ba1d96008b5d6c483a72bdd45b3f430"} Nov 25 15:21:52 crc kubenswrapper[4890]: I1125 15:21:52.724015 4890 generic.go:334] "Generic (PLEG): container finished" podID="bfecce7f-d8ed-44d9-bc97-aba2e5b19e8b" containerID="6e7f48c14398dcfdf4962151c72cc993d0cff6a4d0863894b039abf0f5055f69" exitCode=0 Nov 25 15:21:52 crc kubenswrapper[4890]: I1125 15:21:52.724144 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-66a8-account-create-627gj" event={"ID":"bfecce7f-d8ed-44d9-bc97-aba2e5b19e8b","Type":"ContainerDied","Data":"6e7f48c14398dcfdf4962151c72cc993d0cff6a4d0863894b039abf0f5055f69"} Nov 25 15:21:52 crc kubenswrapper[4890]: I1125 15:21:52.727648 4890 generic.go:334] "Generic (PLEG): container finished" podID="197b5965-ef39-4de5-9b7c-cb64c982c2a4" containerID="27b4254c2dd7539a7c6185ff497de399703ce0e26dee37b3760a3ea080dd76a5" exitCode=0 Nov 25 15:21:52 crc kubenswrapper[4890]: I1125 15:21:52.727734 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-n5rj9" event={"ID":"197b5965-ef39-4de5-9b7c-cb64c982c2a4","Type":"ContainerDied","Data":"27b4254c2dd7539a7c6185ff497de399703ce0e26dee37b3760a3ea080dd76a5"} Nov 25 15:21:52 crc kubenswrapper[4890]: I1125 15:21:52.741809 4890 generic.go:334] "Generic (PLEG): container finished" podID="0c71a662-55d3-46c8-8e7f-f8595bf24172" containerID="649bba0742590fe6749edc35bcbbb56579c0de557a05e2c389b08b8f5102f2d6" exitCode=0 Nov 25 15:21:52 crc kubenswrapper[4890]: I1125 15:21:52.742013 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-cwx2k" event={"ID":"0c71a662-55d3-46c8-8e7f-f8595bf24172","Type":"ContainerDied","Data":"649bba0742590fe6749edc35bcbbb56579c0de557a05e2c389b08b8f5102f2d6"} Nov 25 15:21:52 crc kubenswrapper[4890]: I1125 15:21:52.907585 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 15:21:52 crc kubenswrapper[4890]: W1125 15:21:52.946401 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbb86ab7c_e51d_46ad_a1d6_6a7f5ebabc41.slice/crio-6249d334c5b63d5389ca40d2856d52aff8bb49d2e3d8edbe1081787db03f63c4 WatchSource:0}: Error finding container 6249d334c5b63d5389ca40d2856d52aff8bb49d2e3d8edbe1081787db03f63c4: Status 404 returned error can't find the container with id 6249d334c5b63d5389ca40d2856d52aff8bb49d2e3d8edbe1081787db03f63c4 Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.288108 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.449495 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"59f62350-c15f-4893-ba59-79cca5353cb0\" (UID: \"59f62350-c15f-4893-ba59-79cca5353cb0\") " Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.449597 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/59f62350-c15f-4893-ba59-79cca5353cb0-httpd-run\") pod \"59f62350-c15f-4893-ba59-79cca5353cb0\" (UID: \"59f62350-c15f-4893-ba59-79cca5353cb0\") " Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.449627 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59f62350-c15f-4893-ba59-79cca5353cb0-combined-ca-bundle\") pod \"59f62350-c15f-4893-ba59-79cca5353cb0\" (UID: \"59f62350-c15f-4893-ba59-79cca5353cb0\") " Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.449734 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59f62350-c15f-4893-ba59-79cca5353cb0-scripts\") pod \"59f62350-c15f-4893-ba59-79cca5353cb0\" (UID: \"59f62350-c15f-4893-ba59-79cca5353cb0\") " Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.449761 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hbl7m\" (UniqueName: \"kubernetes.io/projected/59f62350-c15f-4893-ba59-79cca5353cb0-kube-api-access-hbl7m\") pod \"59f62350-c15f-4893-ba59-79cca5353cb0\" (UID: \"59f62350-c15f-4893-ba59-79cca5353cb0\") " Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.449820 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59f62350-c15f-4893-ba59-79cca5353cb0-logs\") pod \"59f62350-c15f-4893-ba59-79cca5353cb0\" (UID: \"59f62350-c15f-4893-ba59-79cca5353cb0\") " Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.449861 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/59f62350-c15f-4893-ba59-79cca5353cb0-internal-tls-certs\") pod \"59f62350-c15f-4893-ba59-79cca5353cb0\" (UID: \"59f62350-c15f-4893-ba59-79cca5353cb0\") " Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.449954 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59f62350-c15f-4893-ba59-79cca5353cb0-config-data\") pod \"59f62350-c15f-4893-ba59-79cca5353cb0\" (UID: \"59f62350-c15f-4893-ba59-79cca5353cb0\") " Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.450263 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59f62350-c15f-4893-ba59-79cca5353cb0-logs" (OuterVolumeSpecName: "logs") pod "59f62350-c15f-4893-ba59-79cca5353cb0" (UID: "59f62350-c15f-4893-ba59-79cca5353cb0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.450963 4890 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59f62350-c15f-4893-ba59-79cca5353cb0-logs\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.451759 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59f62350-c15f-4893-ba59-79cca5353cb0-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "59f62350-c15f-4893-ba59-79cca5353cb0" (UID: "59f62350-c15f-4893-ba59-79cca5353cb0"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.455618 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "59f62350-c15f-4893-ba59-79cca5353cb0" (UID: "59f62350-c15f-4893-ba59-79cca5353cb0"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.457542 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59f62350-c15f-4893-ba59-79cca5353cb0-kube-api-access-hbl7m" (OuterVolumeSpecName: "kube-api-access-hbl7m") pod "59f62350-c15f-4893-ba59-79cca5353cb0" (UID: "59f62350-c15f-4893-ba59-79cca5353cb0"). InnerVolumeSpecName "kube-api-access-hbl7m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.458254 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59f62350-c15f-4893-ba59-79cca5353cb0-scripts" (OuterVolumeSpecName: "scripts") pod "59f62350-c15f-4893-ba59-79cca5353cb0" (UID: "59f62350-c15f-4893-ba59-79cca5353cb0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.479716 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59f62350-c15f-4893-ba59-79cca5353cb0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "59f62350-c15f-4893-ba59-79cca5353cb0" (UID: "59f62350-c15f-4893-ba59-79cca5353cb0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.505449 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59f62350-c15f-4893-ba59-79cca5353cb0-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "59f62350-c15f-4893-ba59-79cca5353cb0" (UID: "59f62350-c15f-4893-ba59-79cca5353cb0"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.509070 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59f62350-c15f-4893-ba59-79cca5353cb0-config-data" (OuterVolumeSpecName: "config-data") pod "59f62350-c15f-4893-ba59-79cca5353cb0" (UID: "59f62350-c15f-4893-ba59-79cca5353cb0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.553327 4890 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/59f62350-c15f-4893-ba59-79cca5353cb0-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.553367 4890 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59f62350-c15f-4893-ba59-79cca5353cb0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.553380 4890 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59f62350-c15f-4893-ba59-79cca5353cb0-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.553392 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hbl7m\" (UniqueName: \"kubernetes.io/projected/59f62350-c15f-4893-ba59-79cca5353cb0-kube-api-access-hbl7m\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.553403 4890 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/59f62350-c15f-4893-ba59-79cca5353cb0-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.553412 4890 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59f62350-c15f-4893-ba59-79cca5353cb0-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.553449 4890 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.573943 4890 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.656994 4890 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.751886 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bb86ab7c-e51d-46ad-a1d6-6a7f5ebabc41","Type":"ContainerStarted","Data":"6249d334c5b63d5389ca40d2856d52aff8bb49d2e3d8edbe1081787db03f63c4"} Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.755533 4890 generic.go:334] "Generic (PLEG): container finished" podID="59f62350-c15f-4893-ba59-79cca5353cb0" containerID="ef0a80dd523cdb8b8412838670a24406e19b5458dcbc367eacf74e0e105e1ce4" exitCode=0 Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.755601 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"59f62350-c15f-4893-ba59-79cca5353cb0","Type":"ContainerDied","Data":"ef0a80dd523cdb8b8412838670a24406e19b5458dcbc367eacf74e0e105e1ce4"} Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.755877 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"59f62350-c15f-4893-ba59-79cca5353cb0","Type":"ContainerDied","Data":"b64868a75f02ebf8ff528cbfa6523e692098b40ba196625945d65b60842aa015"} Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.755896 4890 scope.go:117] "RemoveContainer" containerID="ef0a80dd523cdb8b8412838670a24406e19b5458dcbc367eacf74e0e105e1ce4" Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.755644 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.759217 4890 generic.go:334] "Generic (PLEG): container finished" podID="7bf4c0cd-f20d-4fd9-8077-e733e4196b62" containerID="eb87dca0ce22763a819779b239a159b8c0b39406bd4a1afc6dbea87a86ff28c9" exitCode=0 Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.759299 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-57c6-account-create-skvmk" event={"ID":"7bf4c0cd-f20d-4fd9-8077-e733e4196b62","Type":"ContainerDied","Data":"eb87dca0ce22763a819779b239a159b8c0b39406bd4a1afc6dbea87a86ff28c9"} Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.762958 4890 generic.go:334] "Generic (PLEG): container finished" podID="60cb628f-f5b8-43ef-ab79-9753d39d4c04" containerID="968f71594fbf195538b3495381442d82ebb8bc7cbd156a37ede4596b9ed58f09" exitCode=0 Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.763146 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-xmpnt" event={"ID":"60cb628f-f5b8-43ef-ab79-9753d39d4c04","Type":"ContainerDied","Data":"968f71594fbf195538b3495381442d82ebb8bc7cbd156a37ede4596b9ed58f09"} Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.823957 4890 scope.go:117] "RemoveContainer" containerID="6c3312722a7e81ef35562577700748b4a13d43c509d835aca0dc24f388927ae9" Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.846266 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.859432 4890 scope.go:117] "RemoveContainer" containerID="ef0a80dd523cdb8b8412838670a24406e19b5458dcbc367eacf74e0e105e1ce4" Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.859562 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 15:21:53 crc kubenswrapper[4890]: E1125 15:21:53.860538 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef0a80dd523cdb8b8412838670a24406e19b5458dcbc367eacf74e0e105e1ce4\": container with ID starting with ef0a80dd523cdb8b8412838670a24406e19b5458dcbc367eacf74e0e105e1ce4 not found: ID does not exist" containerID="ef0a80dd523cdb8b8412838670a24406e19b5458dcbc367eacf74e0e105e1ce4" Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.860574 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef0a80dd523cdb8b8412838670a24406e19b5458dcbc367eacf74e0e105e1ce4"} err="failed to get container status \"ef0a80dd523cdb8b8412838670a24406e19b5458dcbc367eacf74e0e105e1ce4\": rpc error: code = NotFound desc = could not find container \"ef0a80dd523cdb8b8412838670a24406e19b5458dcbc367eacf74e0e105e1ce4\": container with ID starting with ef0a80dd523cdb8b8412838670a24406e19b5458dcbc367eacf74e0e105e1ce4 not found: ID does not exist" Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.860601 4890 scope.go:117] "RemoveContainer" containerID="6c3312722a7e81ef35562577700748b4a13d43c509d835aca0dc24f388927ae9" Nov 25 15:21:53 crc kubenswrapper[4890]: E1125 15:21:53.860899 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c3312722a7e81ef35562577700748b4a13d43c509d835aca0dc24f388927ae9\": container with ID starting with 6c3312722a7e81ef35562577700748b4a13d43c509d835aca0dc24f388927ae9 not found: ID does not exist" containerID="6c3312722a7e81ef35562577700748b4a13d43c509d835aca0dc24f388927ae9" Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.860928 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c3312722a7e81ef35562577700748b4a13d43c509d835aca0dc24f388927ae9"} err="failed to get container status \"6c3312722a7e81ef35562577700748b4a13d43c509d835aca0dc24f388927ae9\": rpc error: code = NotFound desc = could not find container \"6c3312722a7e81ef35562577700748b4a13d43c509d835aca0dc24f388927ae9\": container with ID starting with 6c3312722a7e81ef35562577700748b4a13d43c509d835aca0dc24f388927ae9 not found: ID does not exist" Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.867178 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 15:21:53 crc kubenswrapper[4890]: E1125 15:21:53.867656 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59f62350-c15f-4893-ba59-79cca5353cb0" containerName="glance-log" Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.867675 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="59f62350-c15f-4893-ba59-79cca5353cb0" containerName="glance-log" Nov 25 15:21:53 crc kubenswrapper[4890]: E1125 15:21:53.867704 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59f62350-c15f-4893-ba59-79cca5353cb0" containerName="glance-httpd" Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.867712 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="59f62350-c15f-4893-ba59-79cca5353cb0" containerName="glance-httpd" Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.867923 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="59f62350-c15f-4893-ba59-79cca5353cb0" containerName="glance-httpd" Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.868530 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="59f62350-c15f-4893-ba59-79cca5353cb0" containerName="glance-log" Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.869759 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.874626 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.880964 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 25 15:21:53 crc kubenswrapper[4890]: I1125 15:21:53.881179 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.090135 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/31454d1f-69db-4da1-a219-8c8ded147188-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"31454d1f-69db-4da1-a219-8c8ded147188\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.090291 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwvd5\" (UniqueName: \"kubernetes.io/projected/31454d1f-69db-4da1-a219-8c8ded147188-kube-api-access-vwvd5\") pod \"glance-default-internal-api-0\" (UID: \"31454d1f-69db-4da1-a219-8c8ded147188\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.090322 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31454d1f-69db-4da1-a219-8c8ded147188-scripts\") pod \"glance-default-internal-api-0\" (UID: \"31454d1f-69db-4da1-a219-8c8ded147188\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.090359 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31454d1f-69db-4da1-a219-8c8ded147188-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"31454d1f-69db-4da1-a219-8c8ded147188\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.090421 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/31454d1f-69db-4da1-a219-8c8ded147188-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"31454d1f-69db-4da1-a219-8c8ded147188\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.090445 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31454d1f-69db-4da1-a219-8c8ded147188-config-data\") pod \"glance-default-internal-api-0\" (UID: \"31454d1f-69db-4da1-a219-8c8ded147188\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.090469 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"31454d1f-69db-4da1-a219-8c8ded147188\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.090503 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/31454d1f-69db-4da1-a219-8c8ded147188-logs\") pod \"glance-default-internal-api-0\" (UID: \"31454d1f-69db-4da1-a219-8c8ded147188\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.196369 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/31454d1f-69db-4da1-a219-8c8ded147188-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"31454d1f-69db-4da1-a219-8c8ded147188\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.196883 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwvd5\" (UniqueName: \"kubernetes.io/projected/31454d1f-69db-4da1-a219-8c8ded147188-kube-api-access-vwvd5\") pod \"glance-default-internal-api-0\" (UID: \"31454d1f-69db-4da1-a219-8c8ded147188\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.196909 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31454d1f-69db-4da1-a219-8c8ded147188-scripts\") pod \"glance-default-internal-api-0\" (UID: \"31454d1f-69db-4da1-a219-8c8ded147188\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.196937 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31454d1f-69db-4da1-a219-8c8ded147188-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"31454d1f-69db-4da1-a219-8c8ded147188\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.196998 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/31454d1f-69db-4da1-a219-8c8ded147188-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"31454d1f-69db-4da1-a219-8c8ded147188\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.197026 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31454d1f-69db-4da1-a219-8c8ded147188-config-data\") pod \"glance-default-internal-api-0\" (UID: \"31454d1f-69db-4da1-a219-8c8ded147188\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.197052 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"31454d1f-69db-4da1-a219-8c8ded147188\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.197093 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/31454d1f-69db-4da1-a219-8c8ded147188-logs\") pod \"glance-default-internal-api-0\" (UID: \"31454d1f-69db-4da1-a219-8c8ded147188\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.198109 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/31454d1f-69db-4da1-a219-8c8ded147188-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"31454d1f-69db-4da1-a219-8c8ded147188\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.198525 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/31454d1f-69db-4da1-a219-8c8ded147188-logs\") pod \"glance-default-internal-api-0\" (UID: \"31454d1f-69db-4da1-a219-8c8ded147188\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.200958 4890 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"31454d1f-69db-4da1-a219-8c8ded147188\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-internal-api-0" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.206874 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59f62350-c15f-4893-ba59-79cca5353cb0" path="/var/lib/kubelet/pods/59f62350-c15f-4893-ba59-79cca5353cb0/volumes" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.218780 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31454d1f-69db-4da1-a219-8c8ded147188-config-data\") pod \"glance-default-internal-api-0\" (UID: \"31454d1f-69db-4da1-a219-8c8ded147188\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.223729 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/31454d1f-69db-4da1-a219-8c8ded147188-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"31454d1f-69db-4da1-a219-8c8ded147188\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.224975 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31454d1f-69db-4da1-a219-8c8ded147188-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"31454d1f-69db-4da1-a219-8c8ded147188\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.225714 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31454d1f-69db-4da1-a219-8c8ded147188-scripts\") pod \"glance-default-internal-api-0\" (UID: \"31454d1f-69db-4da1-a219-8c8ded147188\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.226030 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwvd5\" (UniqueName: \"kubernetes.io/projected/31454d1f-69db-4da1-a219-8c8ded147188-kube-api-access-vwvd5\") pod \"glance-default-internal-api-0\" (UID: \"31454d1f-69db-4da1-a219-8c8ded147188\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.324478 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"31454d1f-69db-4da1-a219-8c8ded147188\") " pod="openstack/glance-default-internal-api-0" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.393363 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-66a8-account-create-627gj" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.505595 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-d21a-account-create-nlvf4" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.509664 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bfecce7f-d8ed-44d9-bc97-aba2e5b19e8b-operator-scripts\") pod \"bfecce7f-d8ed-44d9-bc97-aba2e5b19e8b\" (UID: \"bfecce7f-d8ed-44d9-bc97-aba2e5b19e8b\") " Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.509850 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vqmd\" (UniqueName: \"kubernetes.io/projected/bfecce7f-d8ed-44d9-bc97-aba2e5b19e8b-kube-api-access-7vqmd\") pod \"bfecce7f-d8ed-44d9-bc97-aba2e5b19e8b\" (UID: \"bfecce7f-d8ed-44d9-bc97-aba2e5b19e8b\") " Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.511588 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bfecce7f-d8ed-44d9-bc97-aba2e5b19e8b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bfecce7f-d8ed-44d9-bc97-aba2e5b19e8b" (UID: "bfecce7f-d8ed-44d9-bc97-aba2e5b19e8b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.515450 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfecce7f-d8ed-44d9-bc97-aba2e5b19e8b-kube-api-access-7vqmd" (OuterVolumeSpecName: "kube-api-access-7vqmd") pod "bfecce7f-d8ed-44d9-bc97-aba2e5b19e8b" (UID: "bfecce7f-d8ed-44d9-bc97-aba2e5b19e8b"). InnerVolumeSpecName "kube-api-access-7vqmd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.518867 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-n5rj9" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.527291 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-cwx2k" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.608907 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.611782 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c71a662-55d3-46c8-8e7f-f8595bf24172-operator-scripts\") pod \"0c71a662-55d3-46c8-8e7f-f8595bf24172\" (UID: \"0c71a662-55d3-46c8-8e7f-f8595bf24172\") " Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.611858 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rqg82\" (UniqueName: \"kubernetes.io/projected/0c71a662-55d3-46c8-8e7f-f8595bf24172-kube-api-access-rqg82\") pod \"0c71a662-55d3-46c8-8e7f-f8595bf24172\" (UID: \"0c71a662-55d3-46c8-8e7f-f8595bf24172\") " Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.611887 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fea0bc8f-0181-4f5d-8e81-47bd6d89c4fb-operator-scripts\") pod \"fea0bc8f-0181-4f5d-8e81-47bd6d89c4fb\" (UID: \"fea0bc8f-0181-4f5d-8e81-47bd6d89c4fb\") " Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.612044 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bm9rk\" (UniqueName: \"kubernetes.io/projected/fea0bc8f-0181-4f5d-8e81-47bd6d89c4fb-kube-api-access-bm9rk\") pod \"fea0bc8f-0181-4f5d-8e81-47bd6d89c4fb\" (UID: \"fea0bc8f-0181-4f5d-8e81-47bd6d89c4fb\") " Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.612122 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/197b5965-ef39-4de5-9b7c-cb64c982c2a4-operator-scripts\") pod \"197b5965-ef39-4de5-9b7c-cb64c982c2a4\" (UID: \"197b5965-ef39-4de5-9b7c-cb64c982c2a4\") " Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.612145 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dk275\" (UniqueName: \"kubernetes.io/projected/197b5965-ef39-4de5-9b7c-cb64c982c2a4-kube-api-access-dk275\") pod \"197b5965-ef39-4de5-9b7c-cb64c982c2a4\" (UID: \"197b5965-ef39-4de5-9b7c-cb64c982c2a4\") " Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.612441 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fea0bc8f-0181-4f5d-8e81-47bd6d89c4fb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fea0bc8f-0181-4f5d-8e81-47bd6d89c4fb" (UID: "fea0bc8f-0181-4f5d-8e81-47bd6d89c4fb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.612551 4890 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bfecce7f-d8ed-44d9-bc97-aba2e5b19e8b-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.612564 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vqmd\" (UniqueName: \"kubernetes.io/projected/bfecce7f-d8ed-44d9-bc97-aba2e5b19e8b-kube-api-access-7vqmd\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.612575 4890 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fea0bc8f-0181-4f5d-8e81-47bd6d89c4fb-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.613152 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/197b5965-ef39-4de5-9b7c-cb64c982c2a4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "197b5965-ef39-4de5-9b7c-cb64c982c2a4" (UID: "197b5965-ef39-4de5-9b7c-cb64c982c2a4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.615599 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fea0bc8f-0181-4f5d-8e81-47bd6d89c4fb-kube-api-access-bm9rk" (OuterVolumeSpecName: "kube-api-access-bm9rk") pod "fea0bc8f-0181-4f5d-8e81-47bd6d89c4fb" (UID: "fea0bc8f-0181-4f5d-8e81-47bd6d89c4fb"). InnerVolumeSpecName "kube-api-access-bm9rk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.615634 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/197b5965-ef39-4de5-9b7c-cb64c982c2a4-kube-api-access-dk275" (OuterVolumeSpecName: "kube-api-access-dk275") pod "197b5965-ef39-4de5-9b7c-cb64c982c2a4" (UID: "197b5965-ef39-4de5-9b7c-cb64c982c2a4"). InnerVolumeSpecName "kube-api-access-dk275". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.616144 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c71a662-55d3-46c8-8e7f-f8595bf24172-kube-api-access-rqg82" (OuterVolumeSpecName: "kube-api-access-rqg82") pod "0c71a662-55d3-46c8-8e7f-f8595bf24172" (UID: "0c71a662-55d3-46c8-8e7f-f8595bf24172"). InnerVolumeSpecName "kube-api-access-rqg82". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.616906 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c71a662-55d3-46c8-8e7f-f8595bf24172-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0c71a662-55d3-46c8-8e7f-f8595bf24172" (UID: "0c71a662-55d3-46c8-8e7f-f8595bf24172"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.714135 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bm9rk\" (UniqueName: \"kubernetes.io/projected/fea0bc8f-0181-4f5d-8e81-47bd6d89c4fb-kube-api-access-bm9rk\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.714401 4890 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/197b5965-ef39-4de5-9b7c-cb64c982c2a4-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.714411 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dk275\" (UniqueName: \"kubernetes.io/projected/197b5965-ef39-4de5-9b7c-cb64c982c2a4-kube-api-access-dk275\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.714420 4890 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c71a662-55d3-46c8-8e7f-f8595bf24172-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.714428 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rqg82\" (UniqueName: \"kubernetes.io/projected/0c71a662-55d3-46c8-8e7f-f8595bf24172-kube-api-access-rqg82\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.846144 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-d21a-account-create-nlvf4" event={"ID":"fea0bc8f-0181-4f5d-8e81-47bd6d89c4fb","Type":"ContainerDied","Data":"764803b46f953511703c6456d7d9809eae6daf79a370b150ab2c90c4c88d0ae5"} Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.846194 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="764803b46f953511703c6456d7d9809eae6daf79a370b150ab2c90c4c88d0ae5" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.846251 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-d21a-account-create-nlvf4" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.854325 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-66a8-account-create-627gj" event={"ID":"bfecce7f-d8ed-44d9-bc97-aba2e5b19e8b","Type":"ContainerDied","Data":"443c0761849e2c50bec2d6c41489a102946d7825b23b4b43fac15da1637758c6"} Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.854367 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="443c0761849e2c50bec2d6c41489a102946d7825b23b4b43fac15da1637758c6" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.854430 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-66a8-account-create-627gj" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.865596 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-n5rj9" event={"ID":"197b5965-ef39-4de5-9b7c-cb64c982c2a4","Type":"ContainerDied","Data":"66a95703f88e757681686fa024ce41052d98dfacb6596b261f5d13aed02229d5"} Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.865637 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="66a95703f88e757681686fa024ce41052d98dfacb6596b261f5d13aed02229d5" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.865706 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-n5rj9" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.897363 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-cwx2k" event={"ID":"0c71a662-55d3-46c8-8e7f-f8595bf24172","Type":"ContainerDied","Data":"aee3266eaaff71ab4c9b4e697d6fd1e12e92228b97a0e8ca41959ca807c38d84"} Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.897414 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aee3266eaaff71ab4c9b4e697d6fd1e12e92228b97a0e8ca41959ca807c38d84" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.897488 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-cwx2k" Nov 25 15:21:54 crc kubenswrapper[4890]: I1125 15:21:54.910452 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bb86ab7c-e51d-46ad-a1d6-6a7f5ebabc41","Type":"ContainerStarted","Data":"45c514d454dbf35af3492692d04c881f069aa61f0850f9fc49218def2758ceff"} Nov 25 15:21:55 crc kubenswrapper[4890]: I1125 15:21:55.415953 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 15:21:55 crc kubenswrapper[4890]: I1125 15:21:55.590567 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-xmpnt" Nov 25 15:21:55 crc kubenswrapper[4890]: I1125 15:21:55.602520 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-57c6-account-create-skvmk" Nov 25 15:21:55 crc kubenswrapper[4890]: I1125 15:21:55.752071 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60cb628f-f5b8-43ef-ab79-9753d39d4c04-operator-scripts\") pod \"60cb628f-f5b8-43ef-ab79-9753d39d4c04\" (UID: \"60cb628f-f5b8-43ef-ab79-9753d39d4c04\") " Nov 25 15:21:55 crc kubenswrapper[4890]: I1125 15:21:55.752172 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbbg9\" (UniqueName: \"kubernetes.io/projected/60cb628f-f5b8-43ef-ab79-9753d39d4c04-kube-api-access-qbbg9\") pod \"60cb628f-f5b8-43ef-ab79-9753d39d4c04\" (UID: \"60cb628f-f5b8-43ef-ab79-9753d39d4c04\") " Nov 25 15:21:55 crc kubenswrapper[4890]: I1125 15:21:55.752256 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2d7p\" (UniqueName: \"kubernetes.io/projected/7bf4c0cd-f20d-4fd9-8077-e733e4196b62-kube-api-access-s2d7p\") pod \"7bf4c0cd-f20d-4fd9-8077-e733e4196b62\" (UID: \"7bf4c0cd-f20d-4fd9-8077-e733e4196b62\") " Nov 25 15:21:55 crc kubenswrapper[4890]: I1125 15:21:55.752300 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7bf4c0cd-f20d-4fd9-8077-e733e4196b62-operator-scripts\") pod \"7bf4c0cd-f20d-4fd9-8077-e733e4196b62\" (UID: \"7bf4c0cd-f20d-4fd9-8077-e733e4196b62\") " Nov 25 15:21:55 crc kubenswrapper[4890]: I1125 15:21:55.753114 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60cb628f-f5b8-43ef-ab79-9753d39d4c04-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "60cb628f-f5b8-43ef-ab79-9753d39d4c04" (UID: "60cb628f-f5b8-43ef-ab79-9753d39d4c04"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:21:55 crc kubenswrapper[4890]: I1125 15:21:55.753468 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bf4c0cd-f20d-4fd9-8077-e733e4196b62-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7bf4c0cd-f20d-4fd9-8077-e733e4196b62" (UID: "7bf4c0cd-f20d-4fd9-8077-e733e4196b62"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:21:55 crc kubenswrapper[4890]: I1125 15:21:55.772956 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60cb628f-f5b8-43ef-ab79-9753d39d4c04-kube-api-access-qbbg9" (OuterVolumeSpecName: "kube-api-access-qbbg9") pod "60cb628f-f5b8-43ef-ab79-9753d39d4c04" (UID: "60cb628f-f5b8-43ef-ab79-9753d39d4c04"). InnerVolumeSpecName "kube-api-access-qbbg9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:21:55 crc kubenswrapper[4890]: I1125 15:21:55.783433 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bf4c0cd-f20d-4fd9-8077-e733e4196b62-kube-api-access-s2d7p" (OuterVolumeSpecName: "kube-api-access-s2d7p") pod "7bf4c0cd-f20d-4fd9-8077-e733e4196b62" (UID: "7bf4c0cd-f20d-4fd9-8077-e733e4196b62"). InnerVolumeSpecName "kube-api-access-s2d7p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:21:55 crc kubenswrapper[4890]: I1125 15:21:55.854620 4890 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60cb628f-f5b8-43ef-ab79-9753d39d4c04-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:55 crc kubenswrapper[4890]: I1125 15:21:55.854668 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbbg9\" (UniqueName: \"kubernetes.io/projected/60cb628f-f5b8-43ef-ab79-9753d39d4c04-kube-api-access-qbbg9\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:55 crc kubenswrapper[4890]: I1125 15:21:55.854683 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2d7p\" (UniqueName: \"kubernetes.io/projected/7bf4c0cd-f20d-4fd9-8077-e733e4196b62-kube-api-access-s2d7p\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:55 crc kubenswrapper[4890]: I1125 15:21:55.854694 4890 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7bf4c0cd-f20d-4fd9-8077-e733e4196b62-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:55 crc kubenswrapper[4890]: I1125 15:21:55.961126 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"31454d1f-69db-4da1-a219-8c8ded147188","Type":"ContainerStarted","Data":"dfb11a9d40ea1f84b406e77bbc1fb3943c170215d7dcaa47d71818a59610d53b"} Nov 25 15:21:55 crc kubenswrapper[4890]: I1125 15:21:55.969715 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-57c6-account-create-skvmk" event={"ID":"7bf4c0cd-f20d-4fd9-8077-e733e4196b62","Type":"ContainerDied","Data":"73cf59b4c0c9be16f79a3ae6dbc9af01360e0a2933c223a9d0b349a457ed06ab"} Nov 25 15:21:55 crc kubenswrapper[4890]: I1125 15:21:55.969752 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="73cf59b4c0c9be16f79a3ae6dbc9af01360e0a2933c223a9d0b349a457ed06ab" Nov 25 15:21:55 crc kubenswrapper[4890]: I1125 15:21:55.969804 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-57c6-account-create-skvmk" Nov 25 15:21:55 crc kubenswrapper[4890]: I1125 15:21:55.986821 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-xmpnt" event={"ID":"60cb628f-f5b8-43ef-ab79-9753d39d4c04","Type":"ContainerDied","Data":"df04e86f5afad6d7afd7166efd78efe982eb3be0ce3bf1f1e5b327a1cebbf11c"} Nov 25 15:21:55 crc kubenswrapper[4890]: I1125 15:21:55.986861 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-xmpnt" Nov 25 15:21:55 crc kubenswrapper[4890]: I1125 15:21:55.986892 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df04e86f5afad6d7afd7166efd78efe982eb3be0ce3bf1f1e5b327a1cebbf11c" Nov 25 15:21:55 crc kubenswrapper[4890]: I1125 15:21:55.992941 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bb86ab7c-e51d-46ad-a1d6-6a7f5ebabc41","Type":"ContainerStarted","Data":"9bf35ecbef906e05c78a5e982f18a9b9d6a9e688c6ed6052eb94a4d2ad4c334f"} Nov 25 15:21:56 crc kubenswrapper[4890]: I1125 15:21:56.022137 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=5.022118005 podStartE2EDuration="5.022118005s" podCreationTimestamp="2025-11-25 15:21:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:21:56.019125479 +0000 UTC m=+1174.461588099" watchObservedRunningTime="2025-11-25 15:21:56.022118005 +0000 UTC m=+1174.464580615" Nov 25 15:21:56 crc kubenswrapper[4890]: I1125 15:21:56.711314 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 25 15:21:56 crc kubenswrapper[4890]: I1125 15:21:56.711665 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 25 15:21:56 crc kubenswrapper[4890]: I1125 15:21:56.780232 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 25 15:21:56 crc kubenswrapper[4890]: I1125 15:21:56.800771 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 25 15:21:57 crc kubenswrapper[4890]: I1125 15:21:57.004423 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"31454d1f-69db-4da1-a219-8c8ded147188","Type":"ContainerStarted","Data":"b750ac3618e2f4845ffb63c22a5cc0904dc7a2eb18bda66181f2a2929dd6622d"} Nov 25 15:21:57 crc kubenswrapper[4890]: I1125 15:21:57.004472 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"31454d1f-69db-4da1-a219-8c8ded147188","Type":"ContainerStarted","Data":"0b94ff7c679c2a4d1de2877541430b3e5c25dae01c62789612ada22bca31e0c8"} Nov 25 15:21:57 crc kubenswrapper[4890]: I1125 15:21:57.004685 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 25 15:21:57 crc kubenswrapper[4890]: I1125 15:21:57.004719 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 25 15:21:57 crc kubenswrapper[4890]: I1125 15:21:57.031238 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.031215465 podStartE2EDuration="4.031215465s" podCreationTimestamp="2025-11-25 15:21:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:21:57.024918776 +0000 UTC m=+1175.467381386" watchObservedRunningTime="2025-11-25 15:21:57.031215465 +0000 UTC m=+1175.473678075" Nov 25 15:21:57 crc kubenswrapper[4890]: I1125 15:21:57.108100 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:21:57 crc kubenswrapper[4890]: I1125 15:21:57.108431 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f1ae99b8-5004-434a-812d-0ab8043bd031" containerName="sg-core" containerID="cri-o://ed409c3fa7896d5038584fa75afca3e6decc87f9db9f82aa2e9515bd2784a68f" gracePeriod=30 Nov 25 15:21:57 crc kubenswrapper[4890]: I1125 15:21:57.108463 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f1ae99b8-5004-434a-812d-0ab8043bd031" containerName="proxy-httpd" containerID="cri-o://f3f7fdb1eac70f15ae9dbb3f74123477d82ac50e5c44c0d7c36eccb1e03f2ed7" gracePeriod=30 Nov 25 15:21:57 crc kubenswrapper[4890]: I1125 15:21:57.108499 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f1ae99b8-5004-434a-812d-0ab8043bd031" containerName="ceilometer-notification-agent" containerID="cri-o://89212db8a499c692049b06aa414e256a32d28f2a42604f084d5c0376ffb54df2" gracePeriod=30 Nov 25 15:21:57 crc kubenswrapper[4890]: I1125 15:21:57.108958 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f1ae99b8-5004-434a-812d-0ab8043bd031" containerName="ceilometer-central-agent" containerID="cri-o://05260f40f216ef0403254bbafc9d90e1f11f05da2c4e948c656caf0e28f2129c" gracePeriod=30 Nov 25 15:21:57 crc kubenswrapper[4890]: I1125 15:21:57.118412 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 25 15:21:57 crc kubenswrapper[4890]: I1125 15:21:57.153321 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 25 15:21:58 crc kubenswrapper[4890]: I1125 15:21:58.015054 4890 generic.go:334] "Generic (PLEG): container finished" podID="f1ae99b8-5004-434a-812d-0ab8043bd031" containerID="f3f7fdb1eac70f15ae9dbb3f74123477d82ac50e5c44c0d7c36eccb1e03f2ed7" exitCode=0 Nov 25 15:21:58 crc kubenswrapper[4890]: I1125 15:21:58.015450 4890 generic.go:334] "Generic (PLEG): container finished" podID="f1ae99b8-5004-434a-812d-0ab8043bd031" containerID="ed409c3fa7896d5038584fa75afca3e6decc87f9db9f82aa2e9515bd2784a68f" exitCode=2 Nov 25 15:21:58 crc kubenswrapper[4890]: I1125 15:21:58.015459 4890 generic.go:334] "Generic (PLEG): container finished" podID="f1ae99b8-5004-434a-812d-0ab8043bd031" containerID="89212db8a499c692049b06aa414e256a32d28f2a42604f084d5c0376ffb54df2" exitCode=0 Nov 25 15:21:58 crc kubenswrapper[4890]: I1125 15:21:58.015467 4890 generic.go:334] "Generic (PLEG): container finished" podID="f1ae99b8-5004-434a-812d-0ab8043bd031" containerID="05260f40f216ef0403254bbafc9d90e1f11f05da2c4e948c656caf0e28f2129c" exitCode=0 Nov 25 15:21:58 crc kubenswrapper[4890]: I1125 15:21:58.015135 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f1ae99b8-5004-434a-812d-0ab8043bd031","Type":"ContainerDied","Data":"f3f7fdb1eac70f15ae9dbb3f74123477d82ac50e5c44c0d7c36eccb1e03f2ed7"} Nov 25 15:21:58 crc kubenswrapper[4890]: I1125 15:21:58.015555 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f1ae99b8-5004-434a-812d-0ab8043bd031","Type":"ContainerDied","Data":"ed409c3fa7896d5038584fa75afca3e6decc87f9db9f82aa2e9515bd2784a68f"} Nov 25 15:21:58 crc kubenswrapper[4890]: I1125 15:21:58.015571 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f1ae99b8-5004-434a-812d-0ab8043bd031","Type":"ContainerDied","Data":"89212db8a499c692049b06aa414e256a32d28f2a42604f084d5c0376ffb54df2"} Nov 25 15:21:58 crc kubenswrapper[4890]: I1125 15:21:58.015582 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f1ae99b8-5004-434a-812d-0ab8043bd031","Type":"ContainerDied","Data":"05260f40f216ef0403254bbafc9d90e1f11f05da2c4e948c656caf0e28f2129c"} Nov 25 15:21:58 crc kubenswrapper[4890]: I1125 15:21:58.530746 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 15:21:58 crc kubenswrapper[4890]: I1125 15:21:58.605875 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 25 15:21:58 crc kubenswrapper[4890]: I1125 15:21:58.609763 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fvg6v\" (UniqueName: \"kubernetes.io/projected/f1ae99b8-5004-434a-812d-0ab8043bd031-kube-api-access-fvg6v\") pod \"f1ae99b8-5004-434a-812d-0ab8043bd031\" (UID: \"f1ae99b8-5004-434a-812d-0ab8043bd031\") " Nov 25 15:21:58 crc kubenswrapper[4890]: I1125 15:21:58.609817 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f1ae99b8-5004-434a-812d-0ab8043bd031-run-httpd\") pod \"f1ae99b8-5004-434a-812d-0ab8043bd031\" (UID: \"f1ae99b8-5004-434a-812d-0ab8043bd031\") " Nov 25 15:21:58 crc kubenswrapper[4890]: I1125 15:21:58.609884 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1ae99b8-5004-434a-812d-0ab8043bd031-combined-ca-bundle\") pod \"f1ae99b8-5004-434a-812d-0ab8043bd031\" (UID: \"f1ae99b8-5004-434a-812d-0ab8043bd031\") " Nov 25 15:21:58 crc kubenswrapper[4890]: I1125 15:21:58.609965 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f1ae99b8-5004-434a-812d-0ab8043bd031-log-httpd\") pod \"f1ae99b8-5004-434a-812d-0ab8043bd031\" (UID: \"f1ae99b8-5004-434a-812d-0ab8043bd031\") " Nov 25 15:21:58 crc kubenswrapper[4890]: I1125 15:21:58.609992 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1ae99b8-5004-434a-812d-0ab8043bd031-config-data\") pod \"f1ae99b8-5004-434a-812d-0ab8043bd031\" (UID: \"f1ae99b8-5004-434a-812d-0ab8043bd031\") " Nov 25 15:21:58 crc kubenswrapper[4890]: I1125 15:21:58.610115 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f1ae99b8-5004-434a-812d-0ab8043bd031-scripts\") pod \"f1ae99b8-5004-434a-812d-0ab8043bd031\" (UID: \"f1ae99b8-5004-434a-812d-0ab8043bd031\") " Nov 25 15:21:58 crc kubenswrapper[4890]: I1125 15:21:58.610155 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f1ae99b8-5004-434a-812d-0ab8043bd031-sg-core-conf-yaml\") pod \"f1ae99b8-5004-434a-812d-0ab8043bd031\" (UID: \"f1ae99b8-5004-434a-812d-0ab8043bd031\") " Nov 25 15:21:58 crc kubenswrapper[4890]: I1125 15:21:58.610531 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1ae99b8-5004-434a-812d-0ab8043bd031-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f1ae99b8-5004-434a-812d-0ab8043bd031" (UID: "f1ae99b8-5004-434a-812d-0ab8043bd031"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:21:58 crc kubenswrapper[4890]: I1125 15:21:58.610638 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1ae99b8-5004-434a-812d-0ab8043bd031-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f1ae99b8-5004-434a-812d-0ab8043bd031" (UID: "f1ae99b8-5004-434a-812d-0ab8043bd031"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:21:58 crc kubenswrapper[4890]: I1125 15:21:58.610802 4890 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f1ae99b8-5004-434a-812d-0ab8043bd031-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:58 crc kubenswrapper[4890]: I1125 15:21:58.619542 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1ae99b8-5004-434a-812d-0ab8043bd031-scripts" (OuterVolumeSpecName: "scripts") pod "f1ae99b8-5004-434a-812d-0ab8043bd031" (UID: "f1ae99b8-5004-434a-812d-0ab8043bd031"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:21:58 crc kubenswrapper[4890]: I1125 15:21:58.620195 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1ae99b8-5004-434a-812d-0ab8043bd031-kube-api-access-fvg6v" (OuterVolumeSpecName: "kube-api-access-fvg6v") pod "f1ae99b8-5004-434a-812d-0ab8043bd031" (UID: "f1ae99b8-5004-434a-812d-0ab8043bd031"). InnerVolumeSpecName "kube-api-access-fvg6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:21:58 crc kubenswrapper[4890]: I1125 15:21:58.707354 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1ae99b8-5004-434a-812d-0ab8043bd031-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f1ae99b8-5004-434a-812d-0ab8043bd031" (UID: "f1ae99b8-5004-434a-812d-0ab8043bd031"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:21:58 crc kubenswrapper[4890]: I1125 15:21:58.713130 4890 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f1ae99b8-5004-434a-812d-0ab8043bd031-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:58 crc kubenswrapper[4890]: I1125 15:21:58.713181 4890 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f1ae99b8-5004-434a-812d-0ab8043bd031-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:58 crc kubenswrapper[4890]: I1125 15:21:58.713196 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fvg6v\" (UniqueName: \"kubernetes.io/projected/f1ae99b8-5004-434a-812d-0ab8043bd031-kube-api-access-fvg6v\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:58 crc kubenswrapper[4890]: I1125 15:21:58.713208 4890 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f1ae99b8-5004-434a-812d-0ab8043bd031-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:58 crc kubenswrapper[4890]: I1125 15:21:58.733482 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1ae99b8-5004-434a-812d-0ab8043bd031-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f1ae99b8-5004-434a-812d-0ab8043bd031" (UID: "f1ae99b8-5004-434a-812d-0ab8043bd031"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:21:58 crc kubenswrapper[4890]: I1125 15:21:58.789403 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1ae99b8-5004-434a-812d-0ab8043bd031-config-data" (OuterVolumeSpecName: "config-data") pod "f1ae99b8-5004-434a-812d-0ab8043bd031" (UID: "f1ae99b8-5004-434a-812d-0ab8043bd031"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:21:58 crc kubenswrapper[4890]: I1125 15:21:58.814668 4890 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1ae99b8-5004-434a-812d-0ab8043bd031-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:58 crc kubenswrapper[4890]: I1125 15:21:58.814699 4890 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1ae99b8-5004-434a-812d-0ab8043bd031-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.027018 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f1ae99b8-5004-434a-812d-0ab8043bd031","Type":"ContainerDied","Data":"ccc2122b33c3cad963740d47b2d74385bd34b4e270f6fae8f66ebc5de53f3a3a"} Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.027074 4890 scope.go:117] "RemoveContainer" containerID="f3f7fdb1eac70f15ae9dbb3f74123477d82ac50e5c44c0d7c36eccb1e03f2ed7" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.027073 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.029248 4890 generic.go:334] "Generic (PLEG): container finished" podID="13af51fd-267a-4b92-b08b-49db0d79a866" containerID="57a8dfaf17074cb0e7b19059e430746740989c7d1433da6ed075a05661853b0a" exitCode=0 Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.029286 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-mg8z4" event={"ID":"13af51fd-267a-4b92-b08b-49db0d79a866","Type":"ContainerDied","Data":"57a8dfaf17074cb0e7b19059e430746740989c7d1433da6ed075a05661853b0a"} Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.076198 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.079491 4890 scope.go:117] "RemoveContainer" containerID="ed409c3fa7896d5038584fa75afca3e6decc87f9db9f82aa2e9515bd2784a68f" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.087670 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.112318 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:21:59 crc kubenswrapper[4890]: E1125 15:21:59.112783 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1ae99b8-5004-434a-812d-0ab8043bd031" containerName="sg-core" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.112805 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1ae99b8-5004-434a-812d-0ab8043bd031" containerName="sg-core" Nov 25 15:21:59 crc kubenswrapper[4890]: E1125 15:21:59.112825 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1ae99b8-5004-434a-812d-0ab8043bd031" containerName="proxy-httpd" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.112833 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1ae99b8-5004-434a-812d-0ab8043bd031" containerName="proxy-httpd" Nov 25 15:21:59 crc kubenswrapper[4890]: E1125 15:21:59.112850 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bf4c0cd-f20d-4fd9-8077-e733e4196b62" containerName="mariadb-account-create" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.112857 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bf4c0cd-f20d-4fd9-8077-e733e4196b62" containerName="mariadb-account-create" Nov 25 15:21:59 crc kubenswrapper[4890]: E1125 15:21:59.112873 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60cb628f-f5b8-43ef-ab79-9753d39d4c04" containerName="mariadb-database-create" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.112882 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="60cb628f-f5b8-43ef-ab79-9753d39d4c04" containerName="mariadb-database-create" Nov 25 15:21:59 crc kubenswrapper[4890]: E1125 15:21:59.112897 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c71a662-55d3-46c8-8e7f-f8595bf24172" containerName="mariadb-database-create" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.112905 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c71a662-55d3-46c8-8e7f-f8595bf24172" containerName="mariadb-database-create" Nov 25 15:21:59 crc kubenswrapper[4890]: E1125 15:21:59.112918 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1ae99b8-5004-434a-812d-0ab8043bd031" containerName="ceilometer-central-agent" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.112927 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1ae99b8-5004-434a-812d-0ab8043bd031" containerName="ceilometer-central-agent" Nov 25 15:21:59 crc kubenswrapper[4890]: E1125 15:21:59.112938 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1ae99b8-5004-434a-812d-0ab8043bd031" containerName="ceilometer-notification-agent" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.112948 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1ae99b8-5004-434a-812d-0ab8043bd031" containerName="ceilometer-notification-agent" Nov 25 15:21:59 crc kubenswrapper[4890]: E1125 15:21:59.112961 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fea0bc8f-0181-4f5d-8e81-47bd6d89c4fb" containerName="mariadb-account-create" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.112969 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="fea0bc8f-0181-4f5d-8e81-47bd6d89c4fb" containerName="mariadb-account-create" Nov 25 15:21:59 crc kubenswrapper[4890]: E1125 15:21:59.112987 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="197b5965-ef39-4de5-9b7c-cb64c982c2a4" containerName="mariadb-database-create" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.112994 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="197b5965-ef39-4de5-9b7c-cb64c982c2a4" containerName="mariadb-database-create" Nov 25 15:21:59 crc kubenswrapper[4890]: E1125 15:21:59.113007 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfecce7f-d8ed-44d9-bc97-aba2e5b19e8b" containerName="mariadb-account-create" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.113015 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfecce7f-d8ed-44d9-bc97-aba2e5b19e8b" containerName="mariadb-account-create" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.113643 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bf4c0cd-f20d-4fd9-8077-e733e4196b62" containerName="mariadb-account-create" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.113669 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1ae99b8-5004-434a-812d-0ab8043bd031" containerName="ceilometer-central-agent" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.113686 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="197b5965-ef39-4de5-9b7c-cb64c982c2a4" containerName="mariadb-database-create" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.113707 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="60cb628f-f5b8-43ef-ab79-9753d39d4c04" containerName="mariadb-database-create" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.113721 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfecce7f-d8ed-44d9-bc97-aba2e5b19e8b" containerName="mariadb-account-create" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.113744 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1ae99b8-5004-434a-812d-0ab8043bd031" containerName="proxy-httpd" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.113757 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="fea0bc8f-0181-4f5d-8e81-47bd6d89c4fb" containerName="mariadb-account-create" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.113769 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1ae99b8-5004-434a-812d-0ab8043bd031" containerName="sg-core" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.113782 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1ae99b8-5004-434a-812d-0ab8043bd031" containerName="ceilometer-notification-agent" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.113792 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c71a662-55d3-46c8-8e7f-f8595bf24172" containerName="mariadb-database-create" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.115746 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.116677 4890 scope.go:117] "RemoveContainer" containerID="89212db8a499c692049b06aa414e256a32d28f2a42604f084d5c0376ffb54df2" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.120822 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.121138 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.137175 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.166586 4890 scope.go:117] "RemoveContainer" containerID="05260f40f216ef0403254bbafc9d90e1f11f05da2c4e948c656caf0e28f2129c" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.227201 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9418e10d-9ec3-4b11-b3b9-5330acd24688-config-data\") pod \"ceilometer-0\" (UID: \"9418e10d-9ec3-4b11-b3b9-5330acd24688\") " pod="openstack/ceilometer-0" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.227960 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9418e10d-9ec3-4b11-b3b9-5330acd24688-scripts\") pod \"ceilometer-0\" (UID: \"9418e10d-9ec3-4b11-b3b9-5330acd24688\") " pod="openstack/ceilometer-0" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.228092 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9418e10d-9ec3-4b11-b3b9-5330acd24688-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9418e10d-9ec3-4b11-b3b9-5330acd24688\") " pod="openstack/ceilometer-0" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.228752 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9418e10d-9ec3-4b11-b3b9-5330acd24688-log-httpd\") pod \"ceilometer-0\" (UID: \"9418e10d-9ec3-4b11-b3b9-5330acd24688\") " pod="openstack/ceilometer-0" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.228967 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9418e10d-9ec3-4b11-b3b9-5330acd24688-run-httpd\") pod \"ceilometer-0\" (UID: \"9418e10d-9ec3-4b11-b3b9-5330acd24688\") " pod="openstack/ceilometer-0" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.229068 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9418e10d-9ec3-4b11-b3b9-5330acd24688-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9418e10d-9ec3-4b11-b3b9-5330acd24688\") " pod="openstack/ceilometer-0" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.229316 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2gs4\" (UniqueName: \"kubernetes.io/projected/9418e10d-9ec3-4b11-b3b9-5330acd24688-kube-api-access-v2gs4\") pod \"ceilometer-0\" (UID: \"9418e10d-9ec3-4b11-b3b9-5330acd24688\") " pod="openstack/ceilometer-0" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.331855 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9418e10d-9ec3-4b11-b3b9-5330acd24688-scripts\") pod \"ceilometer-0\" (UID: \"9418e10d-9ec3-4b11-b3b9-5330acd24688\") " pod="openstack/ceilometer-0" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.331921 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9418e10d-9ec3-4b11-b3b9-5330acd24688-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9418e10d-9ec3-4b11-b3b9-5330acd24688\") " pod="openstack/ceilometer-0" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.331985 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9418e10d-9ec3-4b11-b3b9-5330acd24688-log-httpd\") pod \"ceilometer-0\" (UID: \"9418e10d-9ec3-4b11-b3b9-5330acd24688\") " pod="openstack/ceilometer-0" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.332053 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9418e10d-9ec3-4b11-b3b9-5330acd24688-run-httpd\") pod \"ceilometer-0\" (UID: \"9418e10d-9ec3-4b11-b3b9-5330acd24688\") " pod="openstack/ceilometer-0" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.332095 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9418e10d-9ec3-4b11-b3b9-5330acd24688-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9418e10d-9ec3-4b11-b3b9-5330acd24688\") " pod="openstack/ceilometer-0" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.332124 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2gs4\" (UniqueName: \"kubernetes.io/projected/9418e10d-9ec3-4b11-b3b9-5330acd24688-kube-api-access-v2gs4\") pod \"ceilometer-0\" (UID: \"9418e10d-9ec3-4b11-b3b9-5330acd24688\") " pod="openstack/ceilometer-0" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.332184 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9418e10d-9ec3-4b11-b3b9-5330acd24688-config-data\") pod \"ceilometer-0\" (UID: \"9418e10d-9ec3-4b11-b3b9-5330acd24688\") " pod="openstack/ceilometer-0" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.333516 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9418e10d-9ec3-4b11-b3b9-5330acd24688-log-httpd\") pod \"ceilometer-0\" (UID: \"9418e10d-9ec3-4b11-b3b9-5330acd24688\") " pod="openstack/ceilometer-0" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.333951 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9418e10d-9ec3-4b11-b3b9-5330acd24688-run-httpd\") pod \"ceilometer-0\" (UID: \"9418e10d-9ec3-4b11-b3b9-5330acd24688\") " pod="openstack/ceilometer-0" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.340135 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9418e10d-9ec3-4b11-b3b9-5330acd24688-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9418e10d-9ec3-4b11-b3b9-5330acd24688\") " pod="openstack/ceilometer-0" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.340998 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9418e10d-9ec3-4b11-b3b9-5330acd24688-config-data\") pod \"ceilometer-0\" (UID: \"9418e10d-9ec3-4b11-b3b9-5330acd24688\") " pod="openstack/ceilometer-0" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.341440 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9418e10d-9ec3-4b11-b3b9-5330acd24688-scripts\") pod \"ceilometer-0\" (UID: \"9418e10d-9ec3-4b11-b3b9-5330acd24688\") " pod="openstack/ceilometer-0" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.341916 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9418e10d-9ec3-4b11-b3b9-5330acd24688-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9418e10d-9ec3-4b11-b3b9-5330acd24688\") " pod="openstack/ceilometer-0" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.352380 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2gs4\" (UniqueName: \"kubernetes.io/projected/9418e10d-9ec3-4b11-b3b9-5330acd24688-kube-api-access-v2gs4\") pod \"ceilometer-0\" (UID: \"9418e10d-9ec3-4b11-b3b9-5330acd24688\") " pod="openstack/ceilometer-0" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.456455 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.568379 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.568533 4890 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.811879 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-mlvqz"] Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.825800 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-mlvqz"] Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.825971 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.826090 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-mlvqz" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.828889 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.829070 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-qvvbs" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.829149 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.946288 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0-scripts\") pod \"nova-cell0-conductor-db-sync-mlvqz\" (UID: \"910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0\") " pod="openstack/nova-cell0-conductor-db-sync-mlvqz" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.946434 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbqv8\" (UniqueName: \"kubernetes.io/projected/910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0-kube-api-access-mbqv8\") pod \"nova-cell0-conductor-db-sync-mlvqz\" (UID: \"910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0\") " pod="openstack/nova-cell0-conductor-db-sync-mlvqz" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.946478 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-mlvqz\" (UID: \"910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0\") " pod="openstack/nova-cell0-conductor-db-sync-mlvqz" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.946506 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0-config-data\") pod \"nova-cell0-conductor-db-sync-mlvqz\" (UID: \"910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0\") " pod="openstack/nova-cell0-conductor-db-sync-mlvqz" Nov 25 15:21:59 crc kubenswrapper[4890]: I1125 15:21:59.954573 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:22:00 crc kubenswrapper[4890]: W1125 15:22:00.017365 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9418e10d_9ec3_4b11_b3b9_5330acd24688.slice/crio-8cc70c134ae977bfc0d007c279bf00a605bae4e3563035fc8cff49353c368a76 WatchSource:0}: Error finding container 8cc70c134ae977bfc0d007c279bf00a605bae4e3563035fc8cff49353c368a76: Status 404 returned error can't find the container with id 8cc70c134ae977bfc0d007c279bf00a605bae4e3563035fc8cff49353c368a76 Nov 25 15:22:00 crc kubenswrapper[4890]: I1125 15:22:00.043598 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9418e10d-9ec3-4b11-b3b9-5330acd24688","Type":"ContainerStarted","Data":"8cc70c134ae977bfc0d007c279bf00a605bae4e3563035fc8cff49353c368a76"} Nov 25 15:22:00 crc kubenswrapper[4890]: I1125 15:22:00.048531 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-mlvqz\" (UID: \"910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0\") " pod="openstack/nova-cell0-conductor-db-sync-mlvqz" Nov 25 15:22:00 crc kubenswrapper[4890]: I1125 15:22:00.048755 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0-config-data\") pod \"nova-cell0-conductor-db-sync-mlvqz\" (UID: \"910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0\") " pod="openstack/nova-cell0-conductor-db-sync-mlvqz" Nov 25 15:22:00 crc kubenswrapper[4890]: I1125 15:22:00.049018 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0-scripts\") pod \"nova-cell0-conductor-db-sync-mlvqz\" (UID: \"910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0\") " pod="openstack/nova-cell0-conductor-db-sync-mlvqz" Nov 25 15:22:00 crc kubenswrapper[4890]: I1125 15:22:00.049206 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbqv8\" (UniqueName: \"kubernetes.io/projected/910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0-kube-api-access-mbqv8\") pod \"nova-cell0-conductor-db-sync-mlvqz\" (UID: \"910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0\") " pod="openstack/nova-cell0-conductor-db-sync-mlvqz" Nov 25 15:22:00 crc kubenswrapper[4890]: I1125 15:22:00.055352 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-mlvqz\" (UID: \"910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0\") " pod="openstack/nova-cell0-conductor-db-sync-mlvqz" Nov 25 15:22:00 crc kubenswrapper[4890]: I1125 15:22:00.055862 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0-scripts\") pod \"nova-cell0-conductor-db-sync-mlvqz\" (UID: \"910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0\") " pod="openstack/nova-cell0-conductor-db-sync-mlvqz" Nov 25 15:22:00 crc kubenswrapper[4890]: I1125 15:22:00.074819 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbqv8\" (UniqueName: \"kubernetes.io/projected/910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0-kube-api-access-mbqv8\") pod \"nova-cell0-conductor-db-sync-mlvqz\" (UID: \"910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0\") " pod="openstack/nova-cell0-conductor-db-sync-mlvqz" Nov 25 15:22:00 crc kubenswrapper[4890]: I1125 15:22:00.078210 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0-config-data\") pod \"nova-cell0-conductor-db-sync-mlvqz\" (UID: \"910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0\") " pod="openstack/nova-cell0-conductor-db-sync-mlvqz" Nov 25 15:22:00 crc kubenswrapper[4890]: I1125 15:22:00.155431 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-mlvqz" Nov 25 15:22:00 crc kubenswrapper[4890]: I1125 15:22:00.207844 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1ae99b8-5004-434a-812d-0ab8043bd031" path="/var/lib/kubelet/pods/f1ae99b8-5004-434a-812d-0ab8043bd031/volumes" Nov 25 15:22:00 crc kubenswrapper[4890]: I1125 15:22:00.407802 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-mg8z4" Nov 25 15:22:00 crc kubenswrapper[4890]: I1125 15:22:00.557103 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13af51fd-267a-4b92-b08b-49db0d79a866-combined-ca-bundle\") pod \"13af51fd-267a-4b92-b08b-49db0d79a866\" (UID: \"13af51fd-267a-4b92-b08b-49db0d79a866\") " Nov 25 15:22:00 crc kubenswrapper[4890]: I1125 15:22:00.557212 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hj6k7\" (UniqueName: \"kubernetes.io/projected/13af51fd-267a-4b92-b08b-49db0d79a866-kube-api-access-hj6k7\") pod \"13af51fd-267a-4b92-b08b-49db0d79a866\" (UID: \"13af51fd-267a-4b92-b08b-49db0d79a866\") " Nov 25 15:22:00 crc kubenswrapper[4890]: I1125 15:22:00.557264 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/13af51fd-267a-4b92-b08b-49db0d79a866-config\") pod \"13af51fd-267a-4b92-b08b-49db0d79a866\" (UID: \"13af51fd-267a-4b92-b08b-49db0d79a866\") " Nov 25 15:22:00 crc kubenswrapper[4890]: I1125 15:22:00.578876 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13af51fd-267a-4b92-b08b-49db0d79a866-kube-api-access-hj6k7" (OuterVolumeSpecName: "kube-api-access-hj6k7") pod "13af51fd-267a-4b92-b08b-49db0d79a866" (UID: "13af51fd-267a-4b92-b08b-49db0d79a866"). InnerVolumeSpecName "kube-api-access-hj6k7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:22:00 crc kubenswrapper[4890]: I1125 15:22:00.601152 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13af51fd-267a-4b92-b08b-49db0d79a866-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "13af51fd-267a-4b92-b08b-49db0d79a866" (UID: "13af51fd-267a-4b92-b08b-49db0d79a866"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:22:00 crc kubenswrapper[4890]: I1125 15:22:00.625401 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13af51fd-267a-4b92-b08b-49db0d79a866-config" (OuterVolumeSpecName: "config") pod "13af51fd-267a-4b92-b08b-49db0d79a866" (UID: "13af51fd-267a-4b92-b08b-49db0d79a866"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:22:00 crc kubenswrapper[4890]: I1125 15:22:00.655334 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-mlvqz"] Nov 25 15:22:00 crc kubenswrapper[4890]: I1125 15:22:00.658759 4890 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13af51fd-267a-4b92-b08b-49db0d79a866-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:22:00 crc kubenswrapper[4890]: I1125 15:22:00.658791 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hj6k7\" (UniqueName: \"kubernetes.io/projected/13af51fd-267a-4b92-b08b-49db0d79a866-kube-api-access-hj6k7\") on node \"crc\" DevicePath \"\"" Nov 25 15:22:00 crc kubenswrapper[4890]: I1125 15:22:00.658804 4890 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/13af51fd-267a-4b92-b08b-49db0d79a866-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.061733 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-mg8z4" Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.061735 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-mg8z4" event={"ID":"13af51fd-267a-4b92-b08b-49db0d79a866","Type":"ContainerDied","Data":"49f5d63e0c337e5d81d61fd78d2701e12b7d8929b5365a29732cf565ade40ad1"} Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.062140 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="49f5d63e0c337e5d81d61fd78d2701e12b7d8929b5365a29732cf565ade40ad1" Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.063114 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-mlvqz" event={"ID":"910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0","Type":"ContainerStarted","Data":"15611717d8cf7ee7164c9af8f15896d018e145df2180ab4abda3052da41e4310"} Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.064689 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9418e10d-9ec3-4b11-b3b9-5330acd24688","Type":"ContainerStarted","Data":"97d333fd0742a2e95e17ba84bccccc4c4554e2c99a9131c924546cd829172754"} Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.246904 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-kl5cw"] Nov 25 15:22:01 crc kubenswrapper[4890]: E1125 15:22:01.247683 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13af51fd-267a-4b92-b08b-49db0d79a866" containerName="neutron-db-sync" Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.247707 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="13af51fd-267a-4b92-b08b-49db0d79a866" containerName="neutron-db-sync" Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.247923 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="13af51fd-267a-4b92-b08b-49db0d79a866" containerName="neutron-db-sync" Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.248867 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-kl5cw" Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.259787 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-kl5cw"] Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.364255 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-87c97d56d-x6xqf"] Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.381318 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-87c97d56d-x6xqf" Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.402820 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/690fc634-5a1a-4f30-bbba-264e5f3d0b5d-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-kl5cw\" (UID: \"690fc634-5a1a-4f30-bbba-264e5f3d0b5d\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kl5cw" Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.402939 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/690fc634-5a1a-4f30-bbba-264e5f3d0b5d-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-kl5cw\" (UID: \"690fc634-5a1a-4f30-bbba-264e5f3d0b5d\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kl5cw" Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.403013 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/690fc634-5a1a-4f30-bbba-264e5f3d0b5d-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-kl5cw\" (UID: \"690fc634-5a1a-4f30-bbba-264e5f3d0b5d\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kl5cw" Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.403081 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/690fc634-5a1a-4f30-bbba-264e5f3d0b5d-config\") pod \"dnsmasq-dns-5c9776ccc5-kl5cw\" (UID: \"690fc634-5a1a-4f30-bbba-264e5f3d0b5d\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kl5cw" Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.403117 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wr4s5\" (UniqueName: \"kubernetes.io/projected/690fc634-5a1a-4f30-bbba-264e5f3d0b5d-kube-api-access-wr4s5\") pod \"dnsmasq-dns-5c9776ccc5-kl5cw\" (UID: \"690fc634-5a1a-4f30-bbba-264e5f3d0b5d\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kl5cw" Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.403233 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/690fc634-5a1a-4f30-bbba-264e5f3d0b5d-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-kl5cw\" (UID: \"690fc634-5a1a-4f30-bbba-264e5f3d0b5d\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kl5cw" Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.405519 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.405714 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-xzflj" Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.408403 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.408583 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.406382 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-87c97d56d-x6xqf"] Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.507011 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/690fc634-5a1a-4f30-bbba-264e5f3d0b5d-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-kl5cw\" (UID: \"690fc634-5a1a-4f30-bbba-264e5f3d0b5d\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kl5cw" Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.507138 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbb94a13-5e59-489e-8f1a-3a75a1cbcece-ovndb-tls-certs\") pod \"neutron-87c97d56d-x6xqf\" (UID: \"dbb94a13-5e59-489e-8f1a-3a75a1cbcece\") " pod="openstack/neutron-87c97d56d-x6xqf" Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.507205 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/dbb94a13-5e59-489e-8f1a-3a75a1cbcece-httpd-config\") pod \"neutron-87c97d56d-x6xqf\" (UID: \"dbb94a13-5e59-489e-8f1a-3a75a1cbcece\") " pod="openstack/neutron-87c97d56d-x6xqf" Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.507257 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbb94a13-5e59-489e-8f1a-3a75a1cbcece-combined-ca-bundle\") pod \"neutron-87c97d56d-x6xqf\" (UID: \"dbb94a13-5e59-489e-8f1a-3a75a1cbcece\") " pod="openstack/neutron-87c97d56d-x6xqf" Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.507298 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kc6kb\" (UniqueName: \"kubernetes.io/projected/dbb94a13-5e59-489e-8f1a-3a75a1cbcece-kube-api-access-kc6kb\") pod \"neutron-87c97d56d-x6xqf\" (UID: \"dbb94a13-5e59-489e-8f1a-3a75a1cbcece\") " pod="openstack/neutron-87c97d56d-x6xqf" Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.507342 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/690fc634-5a1a-4f30-bbba-264e5f3d0b5d-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-kl5cw\" (UID: \"690fc634-5a1a-4f30-bbba-264e5f3d0b5d\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kl5cw" Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.507418 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/dbb94a13-5e59-489e-8f1a-3a75a1cbcece-config\") pod \"neutron-87c97d56d-x6xqf\" (UID: \"dbb94a13-5e59-489e-8f1a-3a75a1cbcece\") " pod="openstack/neutron-87c97d56d-x6xqf" Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.508022 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/690fc634-5a1a-4f30-bbba-264e5f3d0b5d-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-kl5cw\" (UID: \"690fc634-5a1a-4f30-bbba-264e5f3d0b5d\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kl5cw" Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.508182 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/690fc634-5a1a-4f30-bbba-264e5f3d0b5d-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-kl5cw\" (UID: \"690fc634-5a1a-4f30-bbba-264e5f3d0b5d\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kl5cw" Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.512295 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/690fc634-5a1a-4f30-bbba-264e5f3d0b5d-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-kl5cw\" (UID: \"690fc634-5a1a-4f30-bbba-264e5f3d0b5d\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kl5cw" Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.512473 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/690fc634-5a1a-4f30-bbba-264e5f3d0b5d-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-kl5cw\" (UID: \"690fc634-5a1a-4f30-bbba-264e5f3d0b5d\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kl5cw" Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.512620 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/690fc634-5a1a-4f30-bbba-264e5f3d0b5d-config\") pod \"dnsmasq-dns-5c9776ccc5-kl5cw\" (UID: \"690fc634-5a1a-4f30-bbba-264e5f3d0b5d\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kl5cw" Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.512664 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wr4s5\" (UniqueName: \"kubernetes.io/projected/690fc634-5a1a-4f30-bbba-264e5f3d0b5d-kube-api-access-wr4s5\") pod \"dnsmasq-dns-5c9776ccc5-kl5cw\" (UID: \"690fc634-5a1a-4f30-bbba-264e5f3d0b5d\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kl5cw" Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.516937 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/690fc634-5a1a-4f30-bbba-264e5f3d0b5d-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-kl5cw\" (UID: \"690fc634-5a1a-4f30-bbba-264e5f3d0b5d\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kl5cw" Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.517786 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/690fc634-5a1a-4f30-bbba-264e5f3d0b5d-config\") pod \"dnsmasq-dns-5c9776ccc5-kl5cw\" (UID: \"690fc634-5a1a-4f30-bbba-264e5f3d0b5d\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kl5cw" Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.518285 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/690fc634-5a1a-4f30-bbba-264e5f3d0b5d-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-kl5cw\" (UID: \"690fc634-5a1a-4f30-bbba-264e5f3d0b5d\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kl5cw" Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.550641 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wr4s5\" (UniqueName: \"kubernetes.io/projected/690fc634-5a1a-4f30-bbba-264e5f3d0b5d-kube-api-access-wr4s5\") pod \"dnsmasq-dns-5c9776ccc5-kl5cw\" (UID: \"690fc634-5a1a-4f30-bbba-264e5f3d0b5d\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kl5cw" Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.596552 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-kl5cw" Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.619251 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbb94a13-5e59-489e-8f1a-3a75a1cbcece-ovndb-tls-certs\") pod \"neutron-87c97d56d-x6xqf\" (UID: \"dbb94a13-5e59-489e-8f1a-3a75a1cbcece\") " pod="openstack/neutron-87c97d56d-x6xqf" Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.620254 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/dbb94a13-5e59-489e-8f1a-3a75a1cbcece-httpd-config\") pod \"neutron-87c97d56d-x6xqf\" (UID: \"dbb94a13-5e59-489e-8f1a-3a75a1cbcece\") " pod="openstack/neutron-87c97d56d-x6xqf" Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.620337 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbb94a13-5e59-489e-8f1a-3a75a1cbcece-combined-ca-bundle\") pod \"neutron-87c97d56d-x6xqf\" (UID: \"dbb94a13-5e59-489e-8f1a-3a75a1cbcece\") " pod="openstack/neutron-87c97d56d-x6xqf" Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.620412 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kc6kb\" (UniqueName: \"kubernetes.io/projected/dbb94a13-5e59-489e-8f1a-3a75a1cbcece-kube-api-access-kc6kb\") pod \"neutron-87c97d56d-x6xqf\" (UID: \"dbb94a13-5e59-489e-8f1a-3a75a1cbcece\") " pod="openstack/neutron-87c97d56d-x6xqf" Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.620562 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/dbb94a13-5e59-489e-8f1a-3a75a1cbcece-config\") pod \"neutron-87c97d56d-x6xqf\" (UID: \"dbb94a13-5e59-489e-8f1a-3a75a1cbcece\") " pod="openstack/neutron-87c97d56d-x6xqf" Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.626262 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbb94a13-5e59-489e-8f1a-3a75a1cbcece-ovndb-tls-certs\") pod \"neutron-87c97d56d-x6xqf\" (UID: \"dbb94a13-5e59-489e-8f1a-3a75a1cbcece\") " pod="openstack/neutron-87c97d56d-x6xqf" Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.626563 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbb94a13-5e59-489e-8f1a-3a75a1cbcece-combined-ca-bundle\") pod \"neutron-87c97d56d-x6xqf\" (UID: \"dbb94a13-5e59-489e-8f1a-3a75a1cbcece\") " pod="openstack/neutron-87c97d56d-x6xqf" Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.627731 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/dbb94a13-5e59-489e-8f1a-3a75a1cbcece-config\") pod \"neutron-87c97d56d-x6xqf\" (UID: \"dbb94a13-5e59-489e-8f1a-3a75a1cbcece\") " pod="openstack/neutron-87c97d56d-x6xqf" Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.636106 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/dbb94a13-5e59-489e-8f1a-3a75a1cbcece-httpd-config\") pod \"neutron-87c97d56d-x6xqf\" (UID: \"dbb94a13-5e59-489e-8f1a-3a75a1cbcece\") " pod="openstack/neutron-87c97d56d-x6xqf" Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.648536 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kc6kb\" (UniqueName: \"kubernetes.io/projected/dbb94a13-5e59-489e-8f1a-3a75a1cbcece-kube-api-access-kc6kb\") pod \"neutron-87c97d56d-x6xqf\" (UID: \"dbb94a13-5e59-489e-8f1a-3a75a1cbcece\") " pod="openstack/neutron-87c97d56d-x6xqf" Nov 25 15:22:01 crc kubenswrapper[4890]: I1125 15:22:01.818254 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-87c97d56d-x6xqf" Nov 25 15:22:02 crc kubenswrapper[4890]: I1125 15:22:02.101722 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9418e10d-9ec3-4b11-b3b9-5330acd24688","Type":"ContainerStarted","Data":"21a7bf5c99959d47a91b23c0cff11ab896c34faf7e6b72cb195b33a0416d9f21"} Nov 25 15:22:02 crc kubenswrapper[4890]: I1125 15:22:02.223140 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-kl5cw"] Nov 25 15:22:02 crc kubenswrapper[4890]: I1125 15:22:02.468771 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 25 15:22:02 crc kubenswrapper[4890]: I1125 15:22:02.556245 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-87c97d56d-x6xqf"] Nov 25 15:22:03 crc kubenswrapper[4890]: I1125 15:22:03.128178 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-87c97d56d-x6xqf" event={"ID":"dbb94a13-5e59-489e-8f1a-3a75a1cbcece","Type":"ContainerStarted","Data":"bbf330ee5fe4a96d267faaf6e14c5b54f1fa56e521e29cf61fa956850eb13c6e"} Nov 25 15:22:03 crc kubenswrapper[4890]: I1125 15:22:03.135029 4890 generic.go:334] "Generic (PLEG): container finished" podID="690fc634-5a1a-4f30-bbba-264e5f3d0b5d" containerID="00c78bdfe24e45137486eb82fefad0c72a5dacebc512a5e06667f61811b2ff97" exitCode=0 Nov 25 15:22:03 crc kubenswrapper[4890]: I1125 15:22:03.135098 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-kl5cw" event={"ID":"690fc634-5a1a-4f30-bbba-264e5f3d0b5d","Type":"ContainerDied","Data":"00c78bdfe24e45137486eb82fefad0c72a5dacebc512a5e06667f61811b2ff97"} Nov 25 15:22:03 crc kubenswrapper[4890]: I1125 15:22:03.135128 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-kl5cw" event={"ID":"690fc634-5a1a-4f30-bbba-264e5f3d0b5d","Type":"ContainerStarted","Data":"c84b3e7c06e146df097fa57c912f654d928baca2b5df02121f2da60796087d2c"} Nov 25 15:22:04 crc kubenswrapper[4890]: I1125 15:22:04.151325 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-kl5cw" event={"ID":"690fc634-5a1a-4f30-bbba-264e5f3d0b5d","Type":"ContainerStarted","Data":"07c84e1d442d2b461dc07c93f2e5691075954cc4f96c0f66bc25c474ca99f41b"} Nov 25 15:22:04 crc kubenswrapper[4890]: I1125 15:22:04.153014 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c9776ccc5-kl5cw" Nov 25 15:22:04 crc kubenswrapper[4890]: I1125 15:22:04.165392 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9418e10d-9ec3-4b11-b3b9-5330acd24688","Type":"ContainerStarted","Data":"17e6fd8207fd512b5bef6d0a90026b5eee1015135c00a178f7535a2837111a52"} Nov 25 15:22:04 crc kubenswrapper[4890]: I1125 15:22:04.180463 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c9776ccc5-kl5cw" podStartSLOduration=3.180438502 podStartE2EDuration="3.180438502s" podCreationTimestamp="2025-11-25 15:22:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:22:04.170193305 +0000 UTC m=+1182.612655915" watchObservedRunningTime="2025-11-25 15:22:04.180438502 +0000 UTC m=+1182.622901112" Nov 25 15:22:04 crc kubenswrapper[4890]: I1125 15:22:04.193185 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-87c97d56d-x6xqf" podStartSLOduration=3.193144492 podStartE2EDuration="3.193144492s" podCreationTimestamp="2025-11-25 15:22:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:22:04.190967737 +0000 UTC m=+1182.633430367" watchObservedRunningTime="2025-11-25 15:22:04.193144492 +0000 UTC m=+1182.635607102" Nov 25 15:22:04 crc kubenswrapper[4890]: I1125 15:22:04.204724 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-87c97d56d-x6xqf" event={"ID":"dbb94a13-5e59-489e-8f1a-3a75a1cbcece","Type":"ContainerStarted","Data":"c744f49d9533e645e19af2518ab1f8782de52653f5ee3e3e14434b00e721a8cd"} Nov 25 15:22:04 crc kubenswrapper[4890]: I1125 15:22:04.204821 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-87c97d56d-x6xqf" Nov 25 15:22:04 crc kubenswrapper[4890]: I1125 15:22:04.204846 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-87c97d56d-x6xqf" event={"ID":"dbb94a13-5e59-489e-8f1a-3a75a1cbcece","Type":"ContainerStarted","Data":"28d2ca6e727256b908ff796bafab8becc4b4af8ccfa2df60a6604a6701f0e265"} Nov 25 15:22:04 crc kubenswrapper[4890]: I1125 15:22:04.359888 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-59d57fcc85-nkqp6"] Nov 25 15:22:04 crc kubenswrapper[4890]: I1125 15:22:04.362724 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-59d57fcc85-nkqp6" Nov 25 15:22:04 crc kubenswrapper[4890]: I1125 15:22:04.390039 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Nov 25 15:22:04 crc kubenswrapper[4890]: I1125 15:22:04.390358 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Nov 25 15:22:04 crc kubenswrapper[4890]: I1125 15:22:04.427212 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-59d57fcc85-nkqp6"] Nov 25 15:22:04 crc kubenswrapper[4890]: I1125 15:22:04.494597 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5baff989-9642-4963-816c-83904cab7811-httpd-config\") pod \"neutron-59d57fcc85-nkqp6\" (UID: \"5baff989-9642-4963-816c-83904cab7811\") " pod="openstack/neutron-59d57fcc85-nkqp6" Nov 25 15:22:04 crc kubenswrapper[4890]: I1125 15:22:04.494652 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5baff989-9642-4963-816c-83904cab7811-internal-tls-certs\") pod \"neutron-59d57fcc85-nkqp6\" (UID: \"5baff989-9642-4963-816c-83904cab7811\") " pod="openstack/neutron-59d57fcc85-nkqp6" Nov 25 15:22:04 crc kubenswrapper[4890]: I1125 15:22:04.494851 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5baff989-9642-4963-816c-83904cab7811-ovndb-tls-certs\") pod \"neutron-59d57fcc85-nkqp6\" (UID: \"5baff989-9642-4963-816c-83904cab7811\") " pod="openstack/neutron-59d57fcc85-nkqp6" Nov 25 15:22:04 crc kubenswrapper[4890]: I1125 15:22:04.494952 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5baff989-9642-4963-816c-83904cab7811-public-tls-certs\") pod \"neutron-59d57fcc85-nkqp6\" (UID: \"5baff989-9642-4963-816c-83904cab7811\") " pod="openstack/neutron-59d57fcc85-nkqp6" Nov 25 15:22:04 crc kubenswrapper[4890]: I1125 15:22:04.495031 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5baff989-9642-4963-816c-83904cab7811-config\") pod \"neutron-59d57fcc85-nkqp6\" (UID: \"5baff989-9642-4963-816c-83904cab7811\") " pod="openstack/neutron-59d57fcc85-nkqp6" Nov 25 15:22:04 crc kubenswrapper[4890]: I1125 15:22:04.496567 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5baff989-9642-4963-816c-83904cab7811-combined-ca-bundle\") pod \"neutron-59d57fcc85-nkqp6\" (UID: \"5baff989-9642-4963-816c-83904cab7811\") " pod="openstack/neutron-59d57fcc85-nkqp6" Nov 25 15:22:04 crc kubenswrapper[4890]: I1125 15:22:04.496611 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcnjx\" (UniqueName: \"kubernetes.io/projected/5baff989-9642-4963-816c-83904cab7811-kube-api-access-mcnjx\") pod \"neutron-59d57fcc85-nkqp6\" (UID: \"5baff989-9642-4963-816c-83904cab7811\") " pod="openstack/neutron-59d57fcc85-nkqp6" Nov 25 15:22:04 crc kubenswrapper[4890]: I1125 15:22:04.597898 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5baff989-9642-4963-816c-83904cab7811-httpd-config\") pod \"neutron-59d57fcc85-nkqp6\" (UID: \"5baff989-9642-4963-816c-83904cab7811\") " pod="openstack/neutron-59d57fcc85-nkqp6" Nov 25 15:22:04 crc kubenswrapper[4890]: I1125 15:22:04.598173 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5baff989-9642-4963-816c-83904cab7811-internal-tls-certs\") pod \"neutron-59d57fcc85-nkqp6\" (UID: \"5baff989-9642-4963-816c-83904cab7811\") " pod="openstack/neutron-59d57fcc85-nkqp6" Nov 25 15:22:04 crc kubenswrapper[4890]: I1125 15:22:04.598222 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5baff989-9642-4963-816c-83904cab7811-ovndb-tls-certs\") pod \"neutron-59d57fcc85-nkqp6\" (UID: \"5baff989-9642-4963-816c-83904cab7811\") " pod="openstack/neutron-59d57fcc85-nkqp6" Nov 25 15:22:04 crc kubenswrapper[4890]: I1125 15:22:04.598242 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5baff989-9642-4963-816c-83904cab7811-public-tls-certs\") pod \"neutron-59d57fcc85-nkqp6\" (UID: \"5baff989-9642-4963-816c-83904cab7811\") " pod="openstack/neutron-59d57fcc85-nkqp6" Nov 25 15:22:04 crc kubenswrapper[4890]: I1125 15:22:04.598265 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5baff989-9642-4963-816c-83904cab7811-config\") pod \"neutron-59d57fcc85-nkqp6\" (UID: \"5baff989-9642-4963-816c-83904cab7811\") " pod="openstack/neutron-59d57fcc85-nkqp6" Nov 25 15:22:04 crc kubenswrapper[4890]: I1125 15:22:04.598319 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5baff989-9642-4963-816c-83904cab7811-combined-ca-bundle\") pod \"neutron-59d57fcc85-nkqp6\" (UID: \"5baff989-9642-4963-816c-83904cab7811\") " pod="openstack/neutron-59d57fcc85-nkqp6" Nov 25 15:22:04 crc kubenswrapper[4890]: I1125 15:22:04.598336 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcnjx\" (UniqueName: \"kubernetes.io/projected/5baff989-9642-4963-816c-83904cab7811-kube-api-access-mcnjx\") pod \"neutron-59d57fcc85-nkqp6\" (UID: \"5baff989-9642-4963-816c-83904cab7811\") " pod="openstack/neutron-59d57fcc85-nkqp6" Nov 25 15:22:04 crc kubenswrapper[4890]: I1125 15:22:04.605123 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5baff989-9642-4963-816c-83904cab7811-ovndb-tls-certs\") pod \"neutron-59d57fcc85-nkqp6\" (UID: \"5baff989-9642-4963-816c-83904cab7811\") " pod="openstack/neutron-59d57fcc85-nkqp6" Nov 25 15:22:04 crc kubenswrapper[4890]: I1125 15:22:04.605403 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5baff989-9642-4963-816c-83904cab7811-public-tls-certs\") pod \"neutron-59d57fcc85-nkqp6\" (UID: \"5baff989-9642-4963-816c-83904cab7811\") " pod="openstack/neutron-59d57fcc85-nkqp6" Nov 25 15:22:04 crc kubenswrapper[4890]: I1125 15:22:04.605776 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5baff989-9642-4963-816c-83904cab7811-combined-ca-bundle\") pod \"neutron-59d57fcc85-nkqp6\" (UID: \"5baff989-9642-4963-816c-83904cab7811\") " pod="openstack/neutron-59d57fcc85-nkqp6" Nov 25 15:22:04 crc kubenswrapper[4890]: I1125 15:22:04.606808 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5baff989-9642-4963-816c-83904cab7811-internal-tls-certs\") pod \"neutron-59d57fcc85-nkqp6\" (UID: \"5baff989-9642-4963-816c-83904cab7811\") " pod="openstack/neutron-59d57fcc85-nkqp6" Nov 25 15:22:04 crc kubenswrapper[4890]: I1125 15:22:04.607819 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5baff989-9642-4963-816c-83904cab7811-httpd-config\") pod \"neutron-59d57fcc85-nkqp6\" (UID: \"5baff989-9642-4963-816c-83904cab7811\") " pod="openstack/neutron-59d57fcc85-nkqp6" Nov 25 15:22:04 crc kubenswrapper[4890]: I1125 15:22:04.608155 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/5baff989-9642-4963-816c-83904cab7811-config\") pod \"neutron-59d57fcc85-nkqp6\" (UID: \"5baff989-9642-4963-816c-83904cab7811\") " pod="openstack/neutron-59d57fcc85-nkqp6" Nov 25 15:22:04 crc kubenswrapper[4890]: I1125 15:22:04.609444 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 25 15:22:04 crc kubenswrapper[4890]: I1125 15:22:04.610692 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 25 15:22:04 crc kubenswrapper[4890]: I1125 15:22:04.625641 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcnjx\" (UniqueName: \"kubernetes.io/projected/5baff989-9642-4963-816c-83904cab7811-kube-api-access-mcnjx\") pod \"neutron-59d57fcc85-nkqp6\" (UID: \"5baff989-9642-4963-816c-83904cab7811\") " pod="openstack/neutron-59d57fcc85-nkqp6" Nov 25 15:22:04 crc kubenswrapper[4890]: I1125 15:22:04.663210 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 25 15:22:04 crc kubenswrapper[4890]: I1125 15:22:04.677785 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 25 15:22:04 crc kubenswrapper[4890]: I1125 15:22:04.694781 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-59d57fcc85-nkqp6" Nov 25 15:22:05 crc kubenswrapper[4890]: I1125 15:22:05.186017 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 25 15:22:05 crc kubenswrapper[4890]: I1125 15:22:05.186306 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 25 15:22:05 crc kubenswrapper[4890]: I1125 15:22:05.297911 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-59d57fcc85-nkqp6"] Nov 25 15:22:06 crc kubenswrapper[4890]: I1125 15:22:06.194732 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-59d57fcc85-nkqp6" event={"ID":"5baff989-9642-4963-816c-83904cab7811","Type":"ContainerStarted","Data":"3e433ac3d21000d6c877500f77741afab50b88e0e18d4921365905f24f75efa3"} Nov 25 15:22:06 crc kubenswrapper[4890]: I1125 15:22:06.195261 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-59d57fcc85-nkqp6" event={"ID":"5baff989-9642-4963-816c-83904cab7811","Type":"ContainerStarted","Data":"238a0089c023378be9ecf3d11413cb58f1ecc4779eb6b4d9922275435be0a569"} Nov 25 15:22:06 crc kubenswrapper[4890]: I1125 15:22:06.195273 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-59d57fcc85-nkqp6" event={"ID":"5baff989-9642-4963-816c-83904cab7811","Type":"ContainerStarted","Data":"7c6d79df3f45d81607c4801a4f0df7292848a9535382cc5278d123c4c992d037"} Nov 25 15:22:06 crc kubenswrapper[4890]: I1125 15:22:06.214496 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-59d57fcc85-nkqp6" podStartSLOduration=2.214468299 podStartE2EDuration="2.214468299s" podCreationTimestamp="2025-11-25 15:22:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:22:06.211733871 +0000 UTC m=+1184.654196481" watchObservedRunningTime="2025-11-25 15:22:06.214468299 +0000 UTC m=+1184.656930909" Nov 25 15:22:07 crc kubenswrapper[4890]: I1125 15:22:07.209275 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9418e10d-9ec3-4b11-b3b9-5330acd24688","Type":"ContainerStarted","Data":"2c5ad099f24bb89df8e62266eadfab65a314af37d24b03fe815b564b5a53b930"} Nov 25 15:22:07 crc kubenswrapper[4890]: I1125 15:22:07.209322 4890 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 15:22:07 crc kubenswrapper[4890]: I1125 15:22:07.209807 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-59d57fcc85-nkqp6" Nov 25 15:22:07 crc kubenswrapper[4890]: I1125 15:22:07.209821 4890 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 15:22:07 crc kubenswrapper[4890]: I1125 15:22:07.209830 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 25 15:22:07 crc kubenswrapper[4890]: I1125 15:22:07.237049 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.663094172 podStartE2EDuration="8.237025038s" podCreationTimestamp="2025-11-25 15:21:59 +0000 UTC" firstStartedPulling="2025-11-25 15:22:00.023363481 +0000 UTC m=+1178.465826091" lastFinishedPulling="2025-11-25 15:22:06.597294357 +0000 UTC m=+1185.039756957" observedRunningTime="2025-11-25 15:22:07.229078598 +0000 UTC m=+1185.671541218" watchObservedRunningTime="2025-11-25 15:22:07.237025038 +0000 UTC m=+1185.679487658" Nov 25 15:22:07 crc kubenswrapper[4890]: I1125 15:22:07.408379 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 25 15:22:07 crc kubenswrapper[4890]: I1125 15:22:07.409289 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 25 15:22:11 crc kubenswrapper[4890]: I1125 15:22:11.598207 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c9776ccc5-kl5cw" Nov 25 15:22:11 crc kubenswrapper[4890]: I1125 15:22:11.689474 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-795f4db4bc-xwdbs"] Nov 25 15:22:11 crc kubenswrapper[4890]: I1125 15:22:11.689706 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-795f4db4bc-xwdbs" podUID="3dab95a0-f8d1-43c0-bd81-d3cafe943a16" containerName="dnsmasq-dns" containerID="cri-o://89e68f8d06e257d883b87fe64dd99aadb2b80168680ee45faf22df84dc8e8bb3" gracePeriod=10 Nov 25 15:22:11 crc kubenswrapper[4890]: I1125 15:22:11.915448 4890 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-795f4db4bc-xwdbs" podUID="3dab95a0-f8d1-43c0-bd81-d3cafe943a16" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.163:5353: connect: connection refused" Nov 25 15:22:14 crc kubenswrapper[4890]: I1125 15:22:14.290342 4890 generic.go:334] "Generic (PLEG): container finished" podID="3dab95a0-f8d1-43c0-bd81-d3cafe943a16" containerID="89e68f8d06e257d883b87fe64dd99aadb2b80168680ee45faf22df84dc8e8bb3" exitCode=0 Nov 25 15:22:14 crc kubenswrapper[4890]: I1125 15:22:14.290437 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-795f4db4bc-xwdbs" event={"ID":"3dab95a0-f8d1-43c0-bd81-d3cafe943a16","Type":"ContainerDied","Data":"89e68f8d06e257d883b87fe64dd99aadb2b80168680ee45faf22df84dc8e8bb3"} Nov 25 15:22:15 crc kubenswrapper[4890]: I1125 15:22:15.829633 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-795f4db4bc-xwdbs" Nov 25 15:22:15 crc kubenswrapper[4890]: I1125 15:22:15.944339 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3dab95a0-f8d1-43c0-bd81-d3cafe943a16-dns-swift-storage-0\") pod \"3dab95a0-f8d1-43c0-bd81-d3cafe943a16\" (UID: \"3dab95a0-f8d1-43c0-bd81-d3cafe943a16\") " Nov 25 15:22:15 crc kubenswrapper[4890]: I1125 15:22:15.944640 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3dab95a0-f8d1-43c0-bd81-d3cafe943a16-config\") pod \"3dab95a0-f8d1-43c0-bd81-d3cafe943a16\" (UID: \"3dab95a0-f8d1-43c0-bd81-d3cafe943a16\") " Nov 25 15:22:15 crc kubenswrapper[4890]: I1125 15:22:15.944669 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s59nt\" (UniqueName: \"kubernetes.io/projected/3dab95a0-f8d1-43c0-bd81-d3cafe943a16-kube-api-access-s59nt\") pod \"3dab95a0-f8d1-43c0-bd81-d3cafe943a16\" (UID: \"3dab95a0-f8d1-43c0-bd81-d3cafe943a16\") " Nov 25 15:22:15 crc kubenswrapper[4890]: I1125 15:22:15.944790 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3dab95a0-f8d1-43c0-bd81-d3cafe943a16-ovsdbserver-nb\") pod \"3dab95a0-f8d1-43c0-bd81-d3cafe943a16\" (UID: \"3dab95a0-f8d1-43c0-bd81-d3cafe943a16\") " Nov 25 15:22:15 crc kubenswrapper[4890]: I1125 15:22:15.944824 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3dab95a0-f8d1-43c0-bd81-d3cafe943a16-dns-svc\") pod \"3dab95a0-f8d1-43c0-bd81-d3cafe943a16\" (UID: \"3dab95a0-f8d1-43c0-bd81-d3cafe943a16\") " Nov 25 15:22:15 crc kubenswrapper[4890]: I1125 15:22:15.944856 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3dab95a0-f8d1-43c0-bd81-d3cafe943a16-ovsdbserver-sb\") pod \"3dab95a0-f8d1-43c0-bd81-d3cafe943a16\" (UID: \"3dab95a0-f8d1-43c0-bd81-d3cafe943a16\") " Nov 25 15:22:15 crc kubenswrapper[4890]: I1125 15:22:15.950370 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3dab95a0-f8d1-43c0-bd81-d3cafe943a16-kube-api-access-s59nt" (OuterVolumeSpecName: "kube-api-access-s59nt") pod "3dab95a0-f8d1-43c0-bd81-d3cafe943a16" (UID: "3dab95a0-f8d1-43c0-bd81-d3cafe943a16"). InnerVolumeSpecName "kube-api-access-s59nt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:22:15 crc kubenswrapper[4890]: I1125 15:22:15.997862 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3dab95a0-f8d1-43c0-bd81-d3cafe943a16-config" (OuterVolumeSpecName: "config") pod "3dab95a0-f8d1-43c0-bd81-d3cafe943a16" (UID: "3dab95a0-f8d1-43c0-bd81-d3cafe943a16"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:22:16 crc kubenswrapper[4890]: I1125 15:22:16.009010 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3dab95a0-f8d1-43c0-bd81-d3cafe943a16-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "3dab95a0-f8d1-43c0-bd81-d3cafe943a16" (UID: "3dab95a0-f8d1-43c0-bd81-d3cafe943a16"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:22:16 crc kubenswrapper[4890]: I1125 15:22:16.009052 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3dab95a0-f8d1-43c0-bd81-d3cafe943a16-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3dab95a0-f8d1-43c0-bd81-d3cafe943a16" (UID: "3dab95a0-f8d1-43c0-bd81-d3cafe943a16"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:22:16 crc kubenswrapper[4890]: I1125 15:22:16.015942 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3dab95a0-f8d1-43c0-bd81-d3cafe943a16-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3dab95a0-f8d1-43c0-bd81-d3cafe943a16" (UID: "3dab95a0-f8d1-43c0-bd81-d3cafe943a16"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:22:16 crc kubenswrapper[4890]: I1125 15:22:16.027021 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3dab95a0-f8d1-43c0-bd81-d3cafe943a16-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3dab95a0-f8d1-43c0-bd81-d3cafe943a16" (UID: "3dab95a0-f8d1-43c0-bd81-d3cafe943a16"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:22:16 crc kubenswrapper[4890]: I1125 15:22:16.046686 4890 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3dab95a0-f8d1-43c0-bd81-d3cafe943a16-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 15:22:16 crc kubenswrapper[4890]: I1125 15:22:16.046719 4890 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3dab95a0-f8d1-43c0-bd81-d3cafe943a16-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 15:22:16 crc kubenswrapper[4890]: I1125 15:22:16.046728 4890 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3dab95a0-f8d1-43c0-bd81-d3cafe943a16-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 15:22:16 crc kubenswrapper[4890]: I1125 15:22:16.046736 4890 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3dab95a0-f8d1-43c0-bd81-d3cafe943a16-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 25 15:22:16 crc kubenswrapper[4890]: I1125 15:22:16.046755 4890 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3dab95a0-f8d1-43c0-bd81-d3cafe943a16-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:22:16 crc kubenswrapper[4890]: I1125 15:22:16.046764 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s59nt\" (UniqueName: \"kubernetes.io/projected/3dab95a0-f8d1-43c0-bd81-d3cafe943a16-kube-api-access-s59nt\") on node \"crc\" DevicePath \"\"" Nov 25 15:22:16 crc kubenswrapper[4890]: I1125 15:22:16.309507 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-795f4db4bc-xwdbs" event={"ID":"3dab95a0-f8d1-43c0-bd81-d3cafe943a16","Type":"ContainerDied","Data":"f5cced960e733f360ae35de6797c9523edb2727f632862fc3204eb6efd8bb28d"} Nov 25 15:22:16 crc kubenswrapper[4890]: I1125 15:22:16.309575 4890 scope.go:117] "RemoveContainer" containerID="89e68f8d06e257d883b87fe64dd99aadb2b80168680ee45faf22df84dc8e8bb3" Nov 25 15:22:16 crc kubenswrapper[4890]: I1125 15:22:16.309730 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-795f4db4bc-xwdbs" Nov 25 15:22:16 crc kubenswrapper[4890]: I1125 15:22:16.314340 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-mlvqz" event={"ID":"910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0","Type":"ContainerStarted","Data":"b317dbbc6ec2fb8dfd39a367f59796a433cbc2e2182fcfcad92df15433c00b42"} Nov 25 15:22:16 crc kubenswrapper[4890]: I1125 15:22:16.331533 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-795f4db4bc-xwdbs"] Nov 25 15:22:16 crc kubenswrapper[4890]: I1125 15:22:16.341404 4890 scope.go:117] "RemoveContainer" containerID="25b66f459674dfda658c6f48abf96a0faaa14fa26d4b81871a7d4eba9d78f692" Nov 25 15:22:16 crc kubenswrapper[4890]: I1125 15:22:16.341787 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-795f4db4bc-xwdbs"] Nov 25 15:22:16 crc kubenswrapper[4890]: I1125 15:22:16.347234 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-mlvqz" podStartSLOduration=2.318724703 podStartE2EDuration="17.347216007s" podCreationTimestamp="2025-11-25 15:21:59 +0000 UTC" firstStartedPulling="2025-11-25 15:22:00.68209667 +0000 UTC m=+1179.124559280" lastFinishedPulling="2025-11-25 15:22:15.710587974 +0000 UTC m=+1194.153050584" observedRunningTime="2025-11-25 15:22:16.338671802 +0000 UTC m=+1194.781134412" watchObservedRunningTime="2025-11-25 15:22:16.347216007 +0000 UTC m=+1194.789678617" Nov 25 15:22:16 crc kubenswrapper[4890]: I1125 15:22:16.451909 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:22:16 crc kubenswrapper[4890]: I1125 15:22:16.452216 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9418e10d-9ec3-4b11-b3b9-5330acd24688" containerName="ceilometer-central-agent" containerID="cri-o://97d333fd0742a2e95e17ba84bccccc4c4554e2c99a9131c924546cd829172754" gracePeriod=30 Nov 25 15:22:16 crc kubenswrapper[4890]: I1125 15:22:16.452328 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9418e10d-9ec3-4b11-b3b9-5330acd24688" containerName="sg-core" containerID="cri-o://17e6fd8207fd512b5bef6d0a90026b5eee1015135c00a178f7535a2837111a52" gracePeriod=30 Nov 25 15:22:16 crc kubenswrapper[4890]: I1125 15:22:16.452324 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9418e10d-9ec3-4b11-b3b9-5330acd24688" containerName="proxy-httpd" containerID="cri-o://2c5ad099f24bb89df8e62266eadfab65a314af37d24b03fe815b564b5a53b930" gracePeriod=30 Nov 25 15:22:16 crc kubenswrapper[4890]: I1125 15:22:16.452428 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9418e10d-9ec3-4b11-b3b9-5330acd24688" containerName="ceilometer-notification-agent" containerID="cri-o://21a7bf5c99959d47a91b23c0cff11ab896c34faf7e6b72cb195b33a0416d9f21" gracePeriod=30 Nov 25 15:22:16 crc kubenswrapper[4890]: I1125 15:22:16.472303 4890 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="9418e10d-9ec3-4b11-b3b9-5330acd24688" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.177:3000/\": EOF" Nov 25 15:22:17 crc kubenswrapper[4890]: I1125 15:22:17.325110 4890 generic.go:334] "Generic (PLEG): container finished" podID="9418e10d-9ec3-4b11-b3b9-5330acd24688" containerID="2c5ad099f24bb89df8e62266eadfab65a314af37d24b03fe815b564b5a53b930" exitCode=0 Nov 25 15:22:17 crc kubenswrapper[4890]: I1125 15:22:17.325460 4890 generic.go:334] "Generic (PLEG): container finished" podID="9418e10d-9ec3-4b11-b3b9-5330acd24688" containerID="17e6fd8207fd512b5bef6d0a90026b5eee1015135c00a178f7535a2837111a52" exitCode=2 Nov 25 15:22:17 crc kubenswrapper[4890]: I1125 15:22:17.325474 4890 generic.go:334] "Generic (PLEG): container finished" podID="9418e10d-9ec3-4b11-b3b9-5330acd24688" containerID="97d333fd0742a2e95e17ba84bccccc4c4554e2c99a9131c924546cd829172754" exitCode=0 Nov 25 15:22:17 crc kubenswrapper[4890]: I1125 15:22:17.325259 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9418e10d-9ec3-4b11-b3b9-5330acd24688","Type":"ContainerDied","Data":"2c5ad099f24bb89df8e62266eadfab65a314af37d24b03fe815b564b5a53b930"} Nov 25 15:22:17 crc kubenswrapper[4890]: I1125 15:22:17.325565 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9418e10d-9ec3-4b11-b3b9-5330acd24688","Type":"ContainerDied","Data":"17e6fd8207fd512b5bef6d0a90026b5eee1015135c00a178f7535a2837111a52"} Nov 25 15:22:17 crc kubenswrapper[4890]: I1125 15:22:17.325607 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9418e10d-9ec3-4b11-b3b9-5330acd24688","Type":"ContainerDied","Data":"97d333fd0742a2e95e17ba84bccccc4c4554e2c99a9131c924546cd829172754"} Nov 25 15:22:18 crc kubenswrapper[4890]: I1125 15:22:18.184686 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3dab95a0-f8d1-43c0-bd81-d3cafe943a16" path="/var/lib/kubelet/pods/3dab95a0-f8d1-43c0-bd81-d3cafe943a16/volumes" Nov 25 15:22:20 crc kubenswrapper[4890]: I1125 15:22:20.354926 4890 generic.go:334] "Generic (PLEG): container finished" podID="9418e10d-9ec3-4b11-b3b9-5330acd24688" containerID="21a7bf5c99959d47a91b23c0cff11ab896c34faf7e6b72cb195b33a0416d9f21" exitCode=0 Nov 25 15:22:20 crc kubenswrapper[4890]: I1125 15:22:20.355005 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9418e10d-9ec3-4b11-b3b9-5330acd24688","Type":"ContainerDied","Data":"21a7bf5c99959d47a91b23c0cff11ab896c34faf7e6b72cb195b33a0416d9f21"} Nov 25 15:22:20 crc kubenswrapper[4890]: I1125 15:22:20.650442 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 15:22:20 crc kubenswrapper[4890]: I1125 15:22:20.734193 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9418e10d-9ec3-4b11-b3b9-5330acd24688-log-httpd\") pod \"9418e10d-9ec3-4b11-b3b9-5330acd24688\" (UID: \"9418e10d-9ec3-4b11-b3b9-5330acd24688\") " Nov 25 15:22:20 crc kubenswrapper[4890]: I1125 15:22:20.734251 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9418e10d-9ec3-4b11-b3b9-5330acd24688-config-data\") pod \"9418e10d-9ec3-4b11-b3b9-5330acd24688\" (UID: \"9418e10d-9ec3-4b11-b3b9-5330acd24688\") " Nov 25 15:22:20 crc kubenswrapper[4890]: I1125 15:22:20.734325 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9418e10d-9ec3-4b11-b3b9-5330acd24688-scripts\") pod \"9418e10d-9ec3-4b11-b3b9-5330acd24688\" (UID: \"9418e10d-9ec3-4b11-b3b9-5330acd24688\") " Nov 25 15:22:20 crc kubenswrapper[4890]: I1125 15:22:20.734506 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9418e10d-9ec3-4b11-b3b9-5330acd24688-run-httpd\") pod \"9418e10d-9ec3-4b11-b3b9-5330acd24688\" (UID: \"9418e10d-9ec3-4b11-b3b9-5330acd24688\") " Nov 25 15:22:20 crc kubenswrapper[4890]: I1125 15:22:20.734547 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9418e10d-9ec3-4b11-b3b9-5330acd24688-combined-ca-bundle\") pod \"9418e10d-9ec3-4b11-b3b9-5330acd24688\" (UID: \"9418e10d-9ec3-4b11-b3b9-5330acd24688\") " Nov 25 15:22:20 crc kubenswrapper[4890]: I1125 15:22:20.734642 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v2gs4\" (UniqueName: \"kubernetes.io/projected/9418e10d-9ec3-4b11-b3b9-5330acd24688-kube-api-access-v2gs4\") pod \"9418e10d-9ec3-4b11-b3b9-5330acd24688\" (UID: \"9418e10d-9ec3-4b11-b3b9-5330acd24688\") " Nov 25 15:22:20 crc kubenswrapper[4890]: I1125 15:22:20.734871 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9418e10d-9ec3-4b11-b3b9-5330acd24688-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "9418e10d-9ec3-4b11-b3b9-5330acd24688" (UID: "9418e10d-9ec3-4b11-b3b9-5330acd24688"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:22:20 crc kubenswrapper[4890]: I1125 15:22:20.734964 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9418e10d-9ec3-4b11-b3b9-5330acd24688-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "9418e10d-9ec3-4b11-b3b9-5330acd24688" (UID: "9418e10d-9ec3-4b11-b3b9-5330acd24688"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:22:20 crc kubenswrapper[4890]: I1125 15:22:20.735219 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9418e10d-9ec3-4b11-b3b9-5330acd24688-sg-core-conf-yaml\") pod \"9418e10d-9ec3-4b11-b3b9-5330acd24688\" (UID: \"9418e10d-9ec3-4b11-b3b9-5330acd24688\") " Nov 25 15:22:20 crc kubenswrapper[4890]: I1125 15:22:20.735922 4890 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9418e10d-9ec3-4b11-b3b9-5330acd24688-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 15:22:20 crc kubenswrapper[4890]: I1125 15:22:20.735953 4890 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9418e10d-9ec3-4b11-b3b9-5330acd24688-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 15:22:20 crc kubenswrapper[4890]: I1125 15:22:20.746383 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9418e10d-9ec3-4b11-b3b9-5330acd24688-kube-api-access-v2gs4" (OuterVolumeSpecName: "kube-api-access-v2gs4") pod "9418e10d-9ec3-4b11-b3b9-5330acd24688" (UID: "9418e10d-9ec3-4b11-b3b9-5330acd24688"). InnerVolumeSpecName "kube-api-access-v2gs4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:22:20 crc kubenswrapper[4890]: I1125 15:22:20.759373 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9418e10d-9ec3-4b11-b3b9-5330acd24688-scripts" (OuterVolumeSpecName: "scripts") pod "9418e10d-9ec3-4b11-b3b9-5330acd24688" (UID: "9418e10d-9ec3-4b11-b3b9-5330acd24688"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:22:20 crc kubenswrapper[4890]: I1125 15:22:20.767409 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9418e10d-9ec3-4b11-b3b9-5330acd24688-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "9418e10d-9ec3-4b11-b3b9-5330acd24688" (UID: "9418e10d-9ec3-4b11-b3b9-5330acd24688"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:22:20 crc kubenswrapper[4890]: I1125 15:22:20.828564 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9418e10d-9ec3-4b11-b3b9-5330acd24688-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9418e10d-9ec3-4b11-b3b9-5330acd24688" (UID: "9418e10d-9ec3-4b11-b3b9-5330acd24688"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:22:20 crc kubenswrapper[4890]: I1125 15:22:20.837325 4890 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9418e10d-9ec3-4b11-b3b9-5330acd24688-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:22:20 crc kubenswrapper[4890]: I1125 15:22:20.837362 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v2gs4\" (UniqueName: \"kubernetes.io/projected/9418e10d-9ec3-4b11-b3b9-5330acd24688-kube-api-access-v2gs4\") on node \"crc\" DevicePath \"\"" Nov 25 15:22:20 crc kubenswrapper[4890]: I1125 15:22:20.837378 4890 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9418e10d-9ec3-4b11-b3b9-5330acd24688-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 25 15:22:20 crc kubenswrapper[4890]: I1125 15:22:20.837394 4890 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9418e10d-9ec3-4b11-b3b9-5330acd24688-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 15:22:20 crc kubenswrapper[4890]: I1125 15:22:20.851291 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9418e10d-9ec3-4b11-b3b9-5330acd24688-config-data" (OuterVolumeSpecName: "config-data") pod "9418e10d-9ec3-4b11-b3b9-5330acd24688" (UID: "9418e10d-9ec3-4b11-b3b9-5330acd24688"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:22:20 crc kubenswrapper[4890]: I1125 15:22:20.939677 4890 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9418e10d-9ec3-4b11-b3b9-5330acd24688-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 15:22:21 crc kubenswrapper[4890]: I1125 15:22:21.366844 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9418e10d-9ec3-4b11-b3b9-5330acd24688","Type":"ContainerDied","Data":"8cc70c134ae977bfc0d007c279bf00a605bae4e3563035fc8cff49353c368a76"} Nov 25 15:22:21 crc kubenswrapper[4890]: I1125 15:22:21.367182 4890 scope.go:117] "RemoveContainer" containerID="2c5ad099f24bb89df8e62266eadfab65a314af37d24b03fe815b564b5a53b930" Nov 25 15:22:21 crc kubenswrapper[4890]: I1125 15:22:21.366954 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 15:22:21 crc kubenswrapper[4890]: I1125 15:22:21.402449 4890 scope.go:117] "RemoveContainer" containerID="17e6fd8207fd512b5bef6d0a90026b5eee1015135c00a178f7535a2837111a52" Nov 25 15:22:21 crc kubenswrapper[4890]: I1125 15:22:21.404196 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:22:21 crc kubenswrapper[4890]: I1125 15:22:21.421788 4890 scope.go:117] "RemoveContainer" containerID="21a7bf5c99959d47a91b23c0cff11ab896c34faf7e6b72cb195b33a0416d9f21" Nov 25 15:22:21 crc kubenswrapper[4890]: I1125 15:22:21.438238 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:22:21 crc kubenswrapper[4890]: I1125 15:22:21.459764 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:22:21 crc kubenswrapper[4890]: E1125 15:22:21.460212 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9418e10d-9ec3-4b11-b3b9-5330acd24688" containerName="ceilometer-notification-agent" Nov 25 15:22:21 crc kubenswrapper[4890]: I1125 15:22:21.460228 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="9418e10d-9ec3-4b11-b3b9-5330acd24688" containerName="ceilometer-notification-agent" Nov 25 15:22:21 crc kubenswrapper[4890]: E1125 15:22:21.460251 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dab95a0-f8d1-43c0-bd81-d3cafe943a16" containerName="init" Nov 25 15:22:21 crc kubenswrapper[4890]: I1125 15:22:21.460258 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dab95a0-f8d1-43c0-bd81-d3cafe943a16" containerName="init" Nov 25 15:22:21 crc kubenswrapper[4890]: E1125 15:22:21.460272 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9418e10d-9ec3-4b11-b3b9-5330acd24688" containerName="ceilometer-central-agent" Nov 25 15:22:21 crc kubenswrapper[4890]: I1125 15:22:21.460278 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="9418e10d-9ec3-4b11-b3b9-5330acd24688" containerName="ceilometer-central-agent" Nov 25 15:22:21 crc kubenswrapper[4890]: E1125 15:22:21.460297 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dab95a0-f8d1-43c0-bd81-d3cafe943a16" containerName="dnsmasq-dns" Nov 25 15:22:21 crc kubenswrapper[4890]: I1125 15:22:21.460303 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dab95a0-f8d1-43c0-bd81-d3cafe943a16" containerName="dnsmasq-dns" Nov 25 15:22:21 crc kubenswrapper[4890]: E1125 15:22:21.460318 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9418e10d-9ec3-4b11-b3b9-5330acd24688" containerName="proxy-httpd" Nov 25 15:22:21 crc kubenswrapper[4890]: I1125 15:22:21.460324 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="9418e10d-9ec3-4b11-b3b9-5330acd24688" containerName="proxy-httpd" Nov 25 15:22:21 crc kubenswrapper[4890]: E1125 15:22:21.460338 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9418e10d-9ec3-4b11-b3b9-5330acd24688" containerName="sg-core" Nov 25 15:22:21 crc kubenswrapper[4890]: I1125 15:22:21.460344 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="9418e10d-9ec3-4b11-b3b9-5330acd24688" containerName="sg-core" Nov 25 15:22:21 crc kubenswrapper[4890]: I1125 15:22:21.460520 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dab95a0-f8d1-43c0-bd81-d3cafe943a16" containerName="dnsmasq-dns" Nov 25 15:22:21 crc kubenswrapper[4890]: I1125 15:22:21.460531 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="9418e10d-9ec3-4b11-b3b9-5330acd24688" containerName="sg-core" Nov 25 15:22:21 crc kubenswrapper[4890]: I1125 15:22:21.460544 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="9418e10d-9ec3-4b11-b3b9-5330acd24688" containerName="proxy-httpd" Nov 25 15:22:21 crc kubenswrapper[4890]: I1125 15:22:21.460552 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="9418e10d-9ec3-4b11-b3b9-5330acd24688" containerName="ceilometer-notification-agent" Nov 25 15:22:21 crc kubenswrapper[4890]: I1125 15:22:21.460572 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="9418e10d-9ec3-4b11-b3b9-5330acd24688" containerName="ceilometer-central-agent" Nov 25 15:22:21 crc kubenswrapper[4890]: I1125 15:22:21.462212 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 15:22:21 crc kubenswrapper[4890]: I1125 15:22:21.466586 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 25 15:22:21 crc kubenswrapper[4890]: I1125 15:22:21.466872 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 25 15:22:21 crc kubenswrapper[4890]: I1125 15:22:21.474838 4890 scope.go:117] "RemoveContainer" containerID="97d333fd0742a2e95e17ba84bccccc4c4554e2c99a9131c924546cd829172754" Nov 25 15:22:21 crc kubenswrapper[4890]: I1125 15:22:21.475399 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:22:21 crc kubenswrapper[4890]: I1125 15:22:21.551633 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aaf63d02-9407-4093-8758-29ebe1caf5d8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aaf63d02-9407-4093-8758-29ebe1caf5d8\") " pod="openstack/ceilometer-0" Nov 25 15:22:21 crc kubenswrapper[4890]: I1125 15:22:21.551817 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aaf63d02-9407-4093-8758-29ebe1caf5d8-run-httpd\") pod \"ceilometer-0\" (UID: \"aaf63d02-9407-4093-8758-29ebe1caf5d8\") " pod="openstack/ceilometer-0" Nov 25 15:22:21 crc kubenswrapper[4890]: I1125 15:22:21.551849 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aaf63d02-9407-4093-8758-29ebe1caf5d8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aaf63d02-9407-4093-8758-29ebe1caf5d8\") " pod="openstack/ceilometer-0" Nov 25 15:22:21 crc kubenswrapper[4890]: I1125 15:22:21.551872 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aaf63d02-9407-4093-8758-29ebe1caf5d8-config-data\") pod \"ceilometer-0\" (UID: \"aaf63d02-9407-4093-8758-29ebe1caf5d8\") " pod="openstack/ceilometer-0" Nov 25 15:22:21 crc kubenswrapper[4890]: I1125 15:22:21.551921 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aaf63d02-9407-4093-8758-29ebe1caf5d8-scripts\") pod \"ceilometer-0\" (UID: \"aaf63d02-9407-4093-8758-29ebe1caf5d8\") " pod="openstack/ceilometer-0" Nov 25 15:22:21 crc kubenswrapper[4890]: I1125 15:22:21.551940 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bqzt\" (UniqueName: \"kubernetes.io/projected/aaf63d02-9407-4093-8758-29ebe1caf5d8-kube-api-access-4bqzt\") pod \"ceilometer-0\" (UID: \"aaf63d02-9407-4093-8758-29ebe1caf5d8\") " pod="openstack/ceilometer-0" Nov 25 15:22:21 crc kubenswrapper[4890]: I1125 15:22:21.551977 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aaf63d02-9407-4093-8758-29ebe1caf5d8-log-httpd\") pod \"ceilometer-0\" (UID: \"aaf63d02-9407-4093-8758-29ebe1caf5d8\") " pod="openstack/ceilometer-0" Nov 25 15:22:21 crc kubenswrapper[4890]: I1125 15:22:21.653566 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aaf63d02-9407-4093-8758-29ebe1caf5d8-run-httpd\") pod \"ceilometer-0\" (UID: \"aaf63d02-9407-4093-8758-29ebe1caf5d8\") " pod="openstack/ceilometer-0" Nov 25 15:22:21 crc kubenswrapper[4890]: I1125 15:22:21.653630 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aaf63d02-9407-4093-8758-29ebe1caf5d8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aaf63d02-9407-4093-8758-29ebe1caf5d8\") " pod="openstack/ceilometer-0" Nov 25 15:22:21 crc kubenswrapper[4890]: I1125 15:22:21.653665 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aaf63d02-9407-4093-8758-29ebe1caf5d8-config-data\") pod \"ceilometer-0\" (UID: \"aaf63d02-9407-4093-8758-29ebe1caf5d8\") " pod="openstack/ceilometer-0" Nov 25 15:22:21 crc kubenswrapper[4890]: I1125 15:22:21.653706 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aaf63d02-9407-4093-8758-29ebe1caf5d8-scripts\") pod \"ceilometer-0\" (UID: \"aaf63d02-9407-4093-8758-29ebe1caf5d8\") " pod="openstack/ceilometer-0" Nov 25 15:22:21 crc kubenswrapper[4890]: I1125 15:22:21.653727 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bqzt\" (UniqueName: \"kubernetes.io/projected/aaf63d02-9407-4093-8758-29ebe1caf5d8-kube-api-access-4bqzt\") pod \"ceilometer-0\" (UID: \"aaf63d02-9407-4093-8758-29ebe1caf5d8\") " pod="openstack/ceilometer-0" Nov 25 15:22:21 crc kubenswrapper[4890]: I1125 15:22:21.653752 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aaf63d02-9407-4093-8758-29ebe1caf5d8-log-httpd\") pod \"ceilometer-0\" (UID: \"aaf63d02-9407-4093-8758-29ebe1caf5d8\") " pod="openstack/ceilometer-0" Nov 25 15:22:21 crc kubenswrapper[4890]: I1125 15:22:21.653804 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aaf63d02-9407-4093-8758-29ebe1caf5d8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aaf63d02-9407-4093-8758-29ebe1caf5d8\") " pod="openstack/ceilometer-0" Nov 25 15:22:21 crc kubenswrapper[4890]: I1125 15:22:21.654806 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aaf63d02-9407-4093-8758-29ebe1caf5d8-run-httpd\") pod \"ceilometer-0\" (UID: \"aaf63d02-9407-4093-8758-29ebe1caf5d8\") " pod="openstack/ceilometer-0" Nov 25 15:22:21 crc kubenswrapper[4890]: I1125 15:22:21.655377 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aaf63d02-9407-4093-8758-29ebe1caf5d8-log-httpd\") pod \"ceilometer-0\" (UID: \"aaf63d02-9407-4093-8758-29ebe1caf5d8\") " pod="openstack/ceilometer-0" Nov 25 15:22:21 crc kubenswrapper[4890]: I1125 15:22:21.660521 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aaf63d02-9407-4093-8758-29ebe1caf5d8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aaf63d02-9407-4093-8758-29ebe1caf5d8\") " pod="openstack/ceilometer-0" Nov 25 15:22:21 crc kubenswrapper[4890]: I1125 15:22:21.660719 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aaf63d02-9407-4093-8758-29ebe1caf5d8-scripts\") pod \"ceilometer-0\" (UID: \"aaf63d02-9407-4093-8758-29ebe1caf5d8\") " pod="openstack/ceilometer-0" Nov 25 15:22:21 crc kubenswrapper[4890]: I1125 15:22:21.660998 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aaf63d02-9407-4093-8758-29ebe1caf5d8-config-data\") pod \"ceilometer-0\" (UID: \"aaf63d02-9407-4093-8758-29ebe1caf5d8\") " pod="openstack/ceilometer-0" Nov 25 15:22:21 crc kubenswrapper[4890]: I1125 15:22:21.671431 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aaf63d02-9407-4093-8758-29ebe1caf5d8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aaf63d02-9407-4093-8758-29ebe1caf5d8\") " pod="openstack/ceilometer-0" Nov 25 15:22:21 crc kubenswrapper[4890]: I1125 15:22:21.672061 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bqzt\" (UniqueName: \"kubernetes.io/projected/aaf63d02-9407-4093-8758-29ebe1caf5d8-kube-api-access-4bqzt\") pod \"ceilometer-0\" (UID: \"aaf63d02-9407-4093-8758-29ebe1caf5d8\") " pod="openstack/ceilometer-0" Nov 25 15:22:21 crc kubenswrapper[4890]: I1125 15:22:21.797720 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 15:22:22 crc kubenswrapper[4890]: I1125 15:22:22.195248 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9418e10d-9ec3-4b11-b3b9-5330acd24688" path="/var/lib/kubelet/pods/9418e10d-9ec3-4b11-b3b9-5330acd24688/volumes" Nov 25 15:22:22 crc kubenswrapper[4890]: I1125 15:22:22.263453 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:22:22 crc kubenswrapper[4890]: I1125 15:22:22.273687 4890 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 15:22:22 crc kubenswrapper[4890]: I1125 15:22:22.378666 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aaf63d02-9407-4093-8758-29ebe1caf5d8","Type":"ContainerStarted","Data":"fbb44c7d8991e6f6df617d90838ab3a5195fae17aa1d0f3e70a117b48830f4fe"} Nov 25 15:22:24 crc kubenswrapper[4890]: I1125 15:22:24.395961 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aaf63d02-9407-4093-8758-29ebe1caf5d8","Type":"ContainerStarted","Data":"5c6b5cc35c60dc33a6bf9c3e4d5e57afcdcd13faab809cbdd8ef561c28ca5f39"} Nov 25 15:22:26 crc kubenswrapper[4890]: I1125 15:22:26.417944 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aaf63d02-9407-4093-8758-29ebe1caf5d8","Type":"ContainerStarted","Data":"a24642291597b8e41a24c60bad3543d78ae08c81bb7d0837708068c58402aef4"} Nov 25 15:22:28 crc kubenswrapper[4890]: I1125 15:22:28.436682 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aaf63d02-9407-4093-8758-29ebe1caf5d8","Type":"ContainerStarted","Data":"bc907fd6e2326efbb65b750dcae20d4cdcc532dbe56d012e59a8d83ad8d3384a"} Nov 25 15:22:30 crc kubenswrapper[4890]: I1125 15:22:30.454824 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aaf63d02-9407-4093-8758-29ebe1caf5d8","Type":"ContainerStarted","Data":"6ad9fbe38ac0708550c685d8125b2a6141698328f641b0bbcf9d878023e72290"} Nov 25 15:22:30 crc kubenswrapper[4890]: I1125 15:22:30.456245 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 25 15:22:30 crc kubenswrapper[4890]: I1125 15:22:30.475175 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.135312173 podStartE2EDuration="9.475145249s" podCreationTimestamp="2025-11-25 15:22:21 +0000 UTC" firstStartedPulling="2025-11-25 15:22:22.273453421 +0000 UTC m=+1200.715916021" lastFinishedPulling="2025-11-25 15:22:29.613286487 +0000 UTC m=+1208.055749097" observedRunningTime="2025-11-25 15:22:30.473516698 +0000 UTC m=+1208.915979318" watchObservedRunningTime="2025-11-25 15:22:30.475145249 +0000 UTC m=+1208.917607859" Nov 25 15:22:30 crc kubenswrapper[4890]: I1125 15:22:30.727781 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:22:31 crc kubenswrapper[4890]: I1125 15:22:31.827223 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-87c97d56d-x6xqf" Nov 25 15:22:32 crc kubenswrapper[4890]: I1125 15:22:32.471645 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aaf63d02-9407-4093-8758-29ebe1caf5d8" containerName="ceilometer-central-agent" containerID="cri-o://5c6b5cc35c60dc33a6bf9c3e4d5e57afcdcd13faab809cbdd8ef561c28ca5f39" gracePeriod=30 Nov 25 15:22:32 crc kubenswrapper[4890]: I1125 15:22:32.471695 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aaf63d02-9407-4093-8758-29ebe1caf5d8" containerName="ceilometer-notification-agent" containerID="cri-o://a24642291597b8e41a24c60bad3543d78ae08c81bb7d0837708068c58402aef4" gracePeriod=30 Nov 25 15:22:32 crc kubenswrapper[4890]: I1125 15:22:32.471717 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aaf63d02-9407-4093-8758-29ebe1caf5d8" containerName="proxy-httpd" containerID="cri-o://6ad9fbe38ac0708550c685d8125b2a6141698328f641b0bbcf9d878023e72290" gracePeriod=30 Nov 25 15:22:32 crc kubenswrapper[4890]: I1125 15:22:32.471717 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aaf63d02-9407-4093-8758-29ebe1caf5d8" containerName="sg-core" containerID="cri-o://bc907fd6e2326efbb65b750dcae20d4cdcc532dbe56d012e59a8d83ad8d3384a" gracePeriod=30 Nov 25 15:22:33 crc kubenswrapper[4890]: I1125 15:22:33.482872 4890 generic.go:334] "Generic (PLEG): container finished" podID="aaf63d02-9407-4093-8758-29ebe1caf5d8" containerID="6ad9fbe38ac0708550c685d8125b2a6141698328f641b0bbcf9d878023e72290" exitCode=0 Nov 25 15:22:33 crc kubenswrapper[4890]: I1125 15:22:33.483136 4890 generic.go:334] "Generic (PLEG): container finished" podID="aaf63d02-9407-4093-8758-29ebe1caf5d8" containerID="bc907fd6e2326efbb65b750dcae20d4cdcc532dbe56d012e59a8d83ad8d3384a" exitCode=2 Nov 25 15:22:33 crc kubenswrapper[4890]: I1125 15:22:33.483144 4890 generic.go:334] "Generic (PLEG): container finished" podID="aaf63d02-9407-4093-8758-29ebe1caf5d8" containerID="a24642291597b8e41a24c60bad3543d78ae08c81bb7d0837708068c58402aef4" exitCode=0 Nov 25 15:22:33 crc kubenswrapper[4890]: I1125 15:22:33.482942 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aaf63d02-9407-4093-8758-29ebe1caf5d8","Type":"ContainerDied","Data":"6ad9fbe38ac0708550c685d8125b2a6141698328f641b0bbcf9d878023e72290"} Nov 25 15:22:33 crc kubenswrapper[4890]: I1125 15:22:33.483182 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aaf63d02-9407-4093-8758-29ebe1caf5d8","Type":"ContainerDied","Data":"bc907fd6e2326efbb65b750dcae20d4cdcc532dbe56d012e59a8d83ad8d3384a"} Nov 25 15:22:33 crc kubenswrapper[4890]: I1125 15:22:33.483193 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aaf63d02-9407-4093-8758-29ebe1caf5d8","Type":"ContainerDied","Data":"a24642291597b8e41a24c60bad3543d78ae08c81bb7d0837708068c58402aef4"} Nov 25 15:22:34 crc kubenswrapper[4890]: I1125 15:22:34.710074 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-59d57fcc85-nkqp6" Nov 25 15:22:34 crc kubenswrapper[4890]: I1125 15:22:34.785645 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-87c97d56d-x6xqf"] Nov 25 15:22:34 crc kubenswrapper[4890]: I1125 15:22:34.785953 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-87c97d56d-x6xqf" podUID="dbb94a13-5e59-489e-8f1a-3a75a1cbcece" containerName="neutron-api" containerID="cri-o://28d2ca6e727256b908ff796bafab8becc4b4af8ccfa2df60a6604a6701f0e265" gracePeriod=30 Nov 25 15:22:34 crc kubenswrapper[4890]: I1125 15:22:34.786061 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-87c97d56d-x6xqf" podUID="dbb94a13-5e59-489e-8f1a-3a75a1cbcece" containerName="neutron-httpd" containerID="cri-o://c744f49d9533e645e19af2518ab1f8782de52653f5ee3e3e14434b00e721a8cd" gracePeriod=30 Nov 25 15:22:35 crc kubenswrapper[4890]: I1125 15:22:35.504622 4890 generic.go:334] "Generic (PLEG): container finished" podID="dbb94a13-5e59-489e-8f1a-3a75a1cbcece" containerID="c744f49d9533e645e19af2518ab1f8782de52653f5ee3e3e14434b00e721a8cd" exitCode=0 Nov 25 15:22:35 crc kubenswrapper[4890]: I1125 15:22:35.504649 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-87c97d56d-x6xqf" event={"ID":"dbb94a13-5e59-489e-8f1a-3a75a1cbcece","Type":"ContainerDied","Data":"c744f49d9533e645e19af2518ab1f8782de52653f5ee3e3e14434b00e721a8cd"} Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.381431 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-87c97d56d-x6xqf" Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.437620 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/dbb94a13-5e59-489e-8f1a-3a75a1cbcece-httpd-config\") pod \"dbb94a13-5e59-489e-8f1a-3a75a1cbcece\" (UID: \"dbb94a13-5e59-489e-8f1a-3a75a1cbcece\") " Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.437762 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbb94a13-5e59-489e-8f1a-3a75a1cbcece-ovndb-tls-certs\") pod \"dbb94a13-5e59-489e-8f1a-3a75a1cbcece\" (UID: \"dbb94a13-5e59-489e-8f1a-3a75a1cbcece\") " Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.437844 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbb94a13-5e59-489e-8f1a-3a75a1cbcece-combined-ca-bundle\") pod \"dbb94a13-5e59-489e-8f1a-3a75a1cbcece\" (UID: \"dbb94a13-5e59-489e-8f1a-3a75a1cbcece\") " Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.437950 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kc6kb\" (UniqueName: \"kubernetes.io/projected/dbb94a13-5e59-489e-8f1a-3a75a1cbcece-kube-api-access-kc6kb\") pod \"dbb94a13-5e59-489e-8f1a-3a75a1cbcece\" (UID: \"dbb94a13-5e59-489e-8f1a-3a75a1cbcece\") " Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.437984 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/dbb94a13-5e59-489e-8f1a-3a75a1cbcece-config\") pod \"dbb94a13-5e59-489e-8f1a-3a75a1cbcece\" (UID: \"dbb94a13-5e59-489e-8f1a-3a75a1cbcece\") " Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.463055 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dbb94a13-5e59-489e-8f1a-3a75a1cbcece-kube-api-access-kc6kb" (OuterVolumeSpecName: "kube-api-access-kc6kb") pod "dbb94a13-5e59-489e-8f1a-3a75a1cbcece" (UID: "dbb94a13-5e59-489e-8f1a-3a75a1cbcece"). InnerVolumeSpecName "kube-api-access-kc6kb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.471390 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbb94a13-5e59-489e-8f1a-3a75a1cbcece-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "dbb94a13-5e59-489e-8f1a-3a75a1cbcece" (UID: "dbb94a13-5e59-489e-8f1a-3a75a1cbcece"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.524275 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbb94a13-5e59-489e-8f1a-3a75a1cbcece-config" (OuterVolumeSpecName: "config") pod "dbb94a13-5e59-489e-8f1a-3a75a1cbcece" (UID: "dbb94a13-5e59-489e-8f1a-3a75a1cbcece"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.540097 4890 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/dbb94a13-5e59-489e-8f1a-3a75a1cbcece-httpd-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.540134 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kc6kb\" (UniqueName: \"kubernetes.io/projected/dbb94a13-5e59-489e-8f1a-3a75a1cbcece-kube-api-access-kc6kb\") on node \"crc\" DevicePath \"\"" Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.540149 4890 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/dbb94a13-5e59-489e-8f1a-3a75a1cbcece-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.540788 4890 generic.go:334] "Generic (PLEG): container finished" podID="aaf63d02-9407-4093-8758-29ebe1caf5d8" containerID="5c6b5cc35c60dc33a6bf9c3e4d5e57afcdcd13faab809cbdd8ef561c28ca5f39" exitCode=0 Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.540884 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aaf63d02-9407-4093-8758-29ebe1caf5d8","Type":"ContainerDied","Data":"5c6b5cc35c60dc33a6bf9c3e4d5e57afcdcd13faab809cbdd8ef561c28ca5f39"} Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.541827 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbb94a13-5e59-489e-8f1a-3a75a1cbcece-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dbb94a13-5e59-489e-8f1a-3a75a1cbcece" (UID: "dbb94a13-5e59-489e-8f1a-3a75a1cbcece"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.543286 4890 generic.go:334] "Generic (PLEG): container finished" podID="dbb94a13-5e59-489e-8f1a-3a75a1cbcece" containerID="28d2ca6e727256b908ff796bafab8becc4b4af8ccfa2df60a6604a6701f0e265" exitCode=0 Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.543317 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-87c97d56d-x6xqf" event={"ID":"dbb94a13-5e59-489e-8f1a-3a75a1cbcece","Type":"ContainerDied","Data":"28d2ca6e727256b908ff796bafab8becc4b4af8ccfa2df60a6604a6701f0e265"} Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.543339 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-87c97d56d-x6xqf" event={"ID":"dbb94a13-5e59-489e-8f1a-3a75a1cbcece","Type":"ContainerDied","Data":"bbf330ee5fe4a96d267faaf6e14c5b54f1fa56e521e29cf61fa956850eb13c6e"} Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.543357 4890 scope.go:117] "RemoveContainer" containerID="c744f49d9533e645e19af2518ab1f8782de52653f5ee3e3e14434b00e721a8cd" Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.543471 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-87c97d56d-x6xqf" Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.571452 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbb94a13-5e59-489e-8f1a-3a75a1cbcece-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "dbb94a13-5e59-489e-8f1a-3a75a1cbcece" (UID: "dbb94a13-5e59-489e-8f1a-3a75a1cbcece"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.572769 4890 scope.go:117] "RemoveContainer" containerID="28d2ca6e727256b908ff796bafab8becc4b4af8ccfa2df60a6604a6701f0e265" Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.572875 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.618328 4890 scope.go:117] "RemoveContainer" containerID="c744f49d9533e645e19af2518ab1f8782de52653f5ee3e3e14434b00e721a8cd" Nov 25 15:22:36 crc kubenswrapper[4890]: E1125 15:22:36.621299 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c744f49d9533e645e19af2518ab1f8782de52653f5ee3e3e14434b00e721a8cd\": container with ID starting with c744f49d9533e645e19af2518ab1f8782de52653f5ee3e3e14434b00e721a8cd not found: ID does not exist" containerID="c744f49d9533e645e19af2518ab1f8782de52653f5ee3e3e14434b00e721a8cd" Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.621348 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c744f49d9533e645e19af2518ab1f8782de52653f5ee3e3e14434b00e721a8cd"} err="failed to get container status \"c744f49d9533e645e19af2518ab1f8782de52653f5ee3e3e14434b00e721a8cd\": rpc error: code = NotFound desc = could not find container \"c744f49d9533e645e19af2518ab1f8782de52653f5ee3e3e14434b00e721a8cd\": container with ID starting with c744f49d9533e645e19af2518ab1f8782de52653f5ee3e3e14434b00e721a8cd not found: ID does not exist" Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.621379 4890 scope.go:117] "RemoveContainer" containerID="28d2ca6e727256b908ff796bafab8becc4b4af8ccfa2df60a6604a6701f0e265" Nov 25 15:22:36 crc kubenswrapper[4890]: E1125 15:22:36.625325 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28d2ca6e727256b908ff796bafab8becc4b4af8ccfa2df60a6604a6701f0e265\": container with ID starting with 28d2ca6e727256b908ff796bafab8becc4b4af8ccfa2df60a6604a6701f0e265 not found: ID does not exist" containerID="28d2ca6e727256b908ff796bafab8becc4b4af8ccfa2df60a6604a6701f0e265" Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.625367 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28d2ca6e727256b908ff796bafab8becc4b4af8ccfa2df60a6604a6701f0e265"} err="failed to get container status \"28d2ca6e727256b908ff796bafab8becc4b4af8ccfa2df60a6604a6701f0e265\": rpc error: code = NotFound desc = could not find container \"28d2ca6e727256b908ff796bafab8becc4b4af8ccfa2df60a6604a6701f0e265\": container with ID starting with 28d2ca6e727256b908ff796bafab8becc4b4af8ccfa2df60a6604a6701f0e265 not found: ID does not exist" Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.640908 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aaf63d02-9407-4093-8758-29ebe1caf5d8-scripts\") pod \"aaf63d02-9407-4093-8758-29ebe1caf5d8\" (UID: \"aaf63d02-9407-4093-8758-29ebe1caf5d8\") " Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.641000 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aaf63d02-9407-4093-8758-29ebe1caf5d8-combined-ca-bundle\") pod \"aaf63d02-9407-4093-8758-29ebe1caf5d8\" (UID: \"aaf63d02-9407-4093-8758-29ebe1caf5d8\") " Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.641071 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aaf63d02-9407-4093-8758-29ebe1caf5d8-run-httpd\") pod \"aaf63d02-9407-4093-8758-29ebe1caf5d8\" (UID: \"aaf63d02-9407-4093-8758-29ebe1caf5d8\") " Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.641105 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aaf63d02-9407-4093-8758-29ebe1caf5d8-config-data\") pod \"aaf63d02-9407-4093-8758-29ebe1caf5d8\" (UID: \"aaf63d02-9407-4093-8758-29ebe1caf5d8\") " Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.641290 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aaf63d02-9407-4093-8758-29ebe1caf5d8-sg-core-conf-yaml\") pod \"aaf63d02-9407-4093-8758-29ebe1caf5d8\" (UID: \"aaf63d02-9407-4093-8758-29ebe1caf5d8\") " Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.641327 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4bqzt\" (UniqueName: \"kubernetes.io/projected/aaf63d02-9407-4093-8758-29ebe1caf5d8-kube-api-access-4bqzt\") pod \"aaf63d02-9407-4093-8758-29ebe1caf5d8\" (UID: \"aaf63d02-9407-4093-8758-29ebe1caf5d8\") " Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.641348 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aaf63d02-9407-4093-8758-29ebe1caf5d8-log-httpd\") pod \"aaf63d02-9407-4093-8758-29ebe1caf5d8\" (UID: \"aaf63d02-9407-4093-8758-29ebe1caf5d8\") " Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.641688 4890 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbb94a13-5e59-489e-8f1a-3a75a1cbcece-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.641705 4890 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbb94a13-5e59-489e-8f1a-3a75a1cbcece-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.642031 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aaf63d02-9407-4093-8758-29ebe1caf5d8-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "aaf63d02-9407-4093-8758-29ebe1caf5d8" (UID: "aaf63d02-9407-4093-8758-29ebe1caf5d8"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.642434 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aaf63d02-9407-4093-8758-29ebe1caf5d8-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "aaf63d02-9407-4093-8758-29ebe1caf5d8" (UID: "aaf63d02-9407-4093-8758-29ebe1caf5d8"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.645444 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aaf63d02-9407-4093-8758-29ebe1caf5d8-scripts" (OuterVolumeSpecName: "scripts") pod "aaf63d02-9407-4093-8758-29ebe1caf5d8" (UID: "aaf63d02-9407-4093-8758-29ebe1caf5d8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.646456 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aaf63d02-9407-4093-8758-29ebe1caf5d8-kube-api-access-4bqzt" (OuterVolumeSpecName: "kube-api-access-4bqzt") pod "aaf63d02-9407-4093-8758-29ebe1caf5d8" (UID: "aaf63d02-9407-4093-8758-29ebe1caf5d8"). InnerVolumeSpecName "kube-api-access-4bqzt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.729503 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aaf63d02-9407-4093-8758-29ebe1caf5d8-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "aaf63d02-9407-4093-8758-29ebe1caf5d8" (UID: "aaf63d02-9407-4093-8758-29ebe1caf5d8"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.743554 4890 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aaf63d02-9407-4093-8758-29ebe1caf5d8-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.743588 4890 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aaf63d02-9407-4093-8758-29ebe1caf5d8-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.743601 4890 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aaf63d02-9407-4093-8758-29ebe1caf5d8-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.743612 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4bqzt\" (UniqueName: \"kubernetes.io/projected/aaf63d02-9407-4093-8758-29ebe1caf5d8-kube-api-access-4bqzt\") on node \"crc\" DevicePath \"\"" Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.743622 4890 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aaf63d02-9407-4093-8758-29ebe1caf5d8-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.823325 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aaf63d02-9407-4093-8758-29ebe1caf5d8-config-data" (OuterVolumeSpecName: "config-data") pod "aaf63d02-9407-4093-8758-29ebe1caf5d8" (UID: "aaf63d02-9407-4093-8758-29ebe1caf5d8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.845238 4890 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aaf63d02-9407-4093-8758-29ebe1caf5d8-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.848287 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aaf63d02-9407-4093-8758-29ebe1caf5d8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aaf63d02-9407-4093-8758-29ebe1caf5d8" (UID: "aaf63d02-9407-4093-8758-29ebe1caf5d8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.886296 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-87c97d56d-x6xqf"] Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.897194 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-87c97d56d-x6xqf"] Nov 25 15:22:36 crc kubenswrapper[4890]: I1125 15:22:36.947376 4890 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aaf63d02-9407-4093-8758-29ebe1caf5d8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:22:37 crc kubenswrapper[4890]: I1125 15:22:37.555074 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aaf63d02-9407-4093-8758-29ebe1caf5d8","Type":"ContainerDied","Data":"fbb44c7d8991e6f6df617d90838ab3a5195fae17aa1d0f3e70a117b48830f4fe"} Nov 25 15:22:37 crc kubenswrapper[4890]: I1125 15:22:37.555102 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 15:22:37 crc kubenswrapper[4890]: I1125 15:22:37.555126 4890 scope.go:117] "RemoveContainer" containerID="6ad9fbe38ac0708550c685d8125b2a6141698328f641b0bbcf9d878023e72290" Nov 25 15:22:37 crc kubenswrapper[4890]: I1125 15:22:37.582250 4890 scope.go:117] "RemoveContainer" containerID="bc907fd6e2326efbb65b750dcae20d4cdcc532dbe56d012e59a8d83ad8d3384a" Nov 25 15:22:37 crc kubenswrapper[4890]: I1125 15:22:37.630434 4890 scope.go:117] "RemoveContainer" containerID="a24642291597b8e41a24c60bad3543d78ae08c81bb7d0837708068c58402aef4" Nov 25 15:22:37 crc kubenswrapper[4890]: I1125 15:22:37.632213 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:22:37 crc kubenswrapper[4890]: I1125 15:22:37.647863 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:22:37 crc kubenswrapper[4890]: I1125 15:22:37.657529 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:22:37 crc kubenswrapper[4890]: E1125 15:22:37.658015 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbb94a13-5e59-489e-8f1a-3a75a1cbcece" containerName="neutron-httpd" Nov 25 15:22:37 crc kubenswrapper[4890]: I1125 15:22:37.658038 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbb94a13-5e59-489e-8f1a-3a75a1cbcece" containerName="neutron-httpd" Nov 25 15:22:37 crc kubenswrapper[4890]: E1125 15:22:37.658058 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaf63d02-9407-4093-8758-29ebe1caf5d8" containerName="ceilometer-central-agent" Nov 25 15:22:37 crc kubenswrapper[4890]: I1125 15:22:37.658066 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaf63d02-9407-4093-8758-29ebe1caf5d8" containerName="ceilometer-central-agent" Nov 25 15:22:37 crc kubenswrapper[4890]: E1125 15:22:37.658087 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaf63d02-9407-4093-8758-29ebe1caf5d8" containerName="proxy-httpd" Nov 25 15:22:37 crc kubenswrapper[4890]: I1125 15:22:37.658102 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaf63d02-9407-4093-8758-29ebe1caf5d8" containerName="proxy-httpd" Nov 25 15:22:37 crc kubenswrapper[4890]: E1125 15:22:37.658123 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaf63d02-9407-4093-8758-29ebe1caf5d8" containerName="sg-core" Nov 25 15:22:37 crc kubenswrapper[4890]: I1125 15:22:37.658130 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaf63d02-9407-4093-8758-29ebe1caf5d8" containerName="sg-core" Nov 25 15:22:37 crc kubenswrapper[4890]: E1125 15:22:37.658143 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbb94a13-5e59-489e-8f1a-3a75a1cbcece" containerName="neutron-api" Nov 25 15:22:37 crc kubenswrapper[4890]: I1125 15:22:37.658149 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbb94a13-5e59-489e-8f1a-3a75a1cbcece" containerName="neutron-api" Nov 25 15:22:37 crc kubenswrapper[4890]: E1125 15:22:37.658375 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaf63d02-9407-4093-8758-29ebe1caf5d8" containerName="ceilometer-notification-agent" Nov 25 15:22:37 crc kubenswrapper[4890]: I1125 15:22:37.658389 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaf63d02-9407-4093-8758-29ebe1caf5d8" containerName="ceilometer-notification-agent" Nov 25 15:22:37 crc kubenswrapper[4890]: I1125 15:22:37.658594 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="aaf63d02-9407-4093-8758-29ebe1caf5d8" containerName="ceilometer-central-agent" Nov 25 15:22:37 crc kubenswrapper[4890]: I1125 15:22:37.658616 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbb94a13-5e59-489e-8f1a-3a75a1cbcece" containerName="neutron-httpd" Nov 25 15:22:37 crc kubenswrapper[4890]: I1125 15:22:37.658631 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="aaf63d02-9407-4093-8758-29ebe1caf5d8" containerName="proxy-httpd" Nov 25 15:22:37 crc kubenswrapper[4890]: I1125 15:22:37.658641 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="aaf63d02-9407-4093-8758-29ebe1caf5d8" containerName="ceilometer-notification-agent" Nov 25 15:22:37 crc kubenswrapper[4890]: I1125 15:22:37.658659 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbb94a13-5e59-489e-8f1a-3a75a1cbcece" containerName="neutron-api" Nov 25 15:22:37 crc kubenswrapper[4890]: I1125 15:22:37.658679 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="aaf63d02-9407-4093-8758-29ebe1caf5d8" containerName="sg-core" Nov 25 15:22:37 crc kubenswrapper[4890]: I1125 15:22:37.662081 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 15:22:37 crc kubenswrapper[4890]: I1125 15:22:37.665368 4890 scope.go:117] "RemoveContainer" containerID="5c6b5cc35c60dc33a6bf9c3e4d5e57afcdcd13faab809cbdd8ef561c28ca5f39" Nov 25 15:22:37 crc kubenswrapper[4890]: I1125 15:22:37.665681 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 25 15:22:37 crc kubenswrapper[4890]: I1125 15:22:37.665749 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 25 15:22:37 crc kubenswrapper[4890]: I1125 15:22:37.694486 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:22:37 crc kubenswrapper[4890]: I1125 15:22:37.761609 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f40cb730-942e-48c2-8bd5-4b1b9c98e608-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f40cb730-942e-48c2-8bd5-4b1b9c98e608\") " pod="openstack/ceilometer-0" Nov 25 15:22:37 crc kubenswrapper[4890]: I1125 15:22:37.761685 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f40cb730-942e-48c2-8bd5-4b1b9c98e608-config-data\") pod \"ceilometer-0\" (UID: \"f40cb730-942e-48c2-8bd5-4b1b9c98e608\") " pod="openstack/ceilometer-0" Nov 25 15:22:37 crc kubenswrapper[4890]: I1125 15:22:37.761725 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f40cb730-942e-48c2-8bd5-4b1b9c98e608-log-httpd\") pod \"ceilometer-0\" (UID: \"f40cb730-942e-48c2-8bd5-4b1b9c98e608\") " pod="openstack/ceilometer-0" Nov 25 15:22:37 crc kubenswrapper[4890]: I1125 15:22:37.761771 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f40cb730-942e-48c2-8bd5-4b1b9c98e608-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f40cb730-942e-48c2-8bd5-4b1b9c98e608\") " pod="openstack/ceilometer-0" Nov 25 15:22:37 crc kubenswrapper[4890]: I1125 15:22:37.761792 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f40cb730-942e-48c2-8bd5-4b1b9c98e608-run-httpd\") pod \"ceilometer-0\" (UID: \"f40cb730-942e-48c2-8bd5-4b1b9c98e608\") " pod="openstack/ceilometer-0" Nov 25 15:22:37 crc kubenswrapper[4890]: I1125 15:22:37.761830 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f40cb730-942e-48c2-8bd5-4b1b9c98e608-scripts\") pod \"ceilometer-0\" (UID: \"f40cb730-942e-48c2-8bd5-4b1b9c98e608\") " pod="openstack/ceilometer-0" Nov 25 15:22:37 crc kubenswrapper[4890]: I1125 15:22:37.762137 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7t6bv\" (UniqueName: \"kubernetes.io/projected/f40cb730-942e-48c2-8bd5-4b1b9c98e608-kube-api-access-7t6bv\") pod \"ceilometer-0\" (UID: \"f40cb730-942e-48c2-8bd5-4b1b9c98e608\") " pod="openstack/ceilometer-0" Nov 25 15:22:37 crc kubenswrapper[4890]: I1125 15:22:37.864665 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f40cb730-942e-48c2-8bd5-4b1b9c98e608-config-data\") pod \"ceilometer-0\" (UID: \"f40cb730-942e-48c2-8bd5-4b1b9c98e608\") " pod="openstack/ceilometer-0" Nov 25 15:22:37 crc kubenswrapper[4890]: I1125 15:22:37.864799 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f40cb730-942e-48c2-8bd5-4b1b9c98e608-log-httpd\") pod \"ceilometer-0\" (UID: \"f40cb730-942e-48c2-8bd5-4b1b9c98e608\") " pod="openstack/ceilometer-0" Nov 25 15:22:37 crc kubenswrapper[4890]: I1125 15:22:37.864884 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f40cb730-942e-48c2-8bd5-4b1b9c98e608-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f40cb730-942e-48c2-8bd5-4b1b9c98e608\") " pod="openstack/ceilometer-0" Nov 25 15:22:37 crc kubenswrapper[4890]: I1125 15:22:37.864925 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f40cb730-942e-48c2-8bd5-4b1b9c98e608-run-httpd\") pod \"ceilometer-0\" (UID: \"f40cb730-942e-48c2-8bd5-4b1b9c98e608\") " pod="openstack/ceilometer-0" Nov 25 15:22:37 crc kubenswrapper[4890]: I1125 15:22:37.864967 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f40cb730-942e-48c2-8bd5-4b1b9c98e608-scripts\") pod \"ceilometer-0\" (UID: \"f40cb730-942e-48c2-8bd5-4b1b9c98e608\") " pod="openstack/ceilometer-0" Nov 25 15:22:37 crc kubenswrapper[4890]: I1125 15:22:37.865036 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7t6bv\" (UniqueName: \"kubernetes.io/projected/f40cb730-942e-48c2-8bd5-4b1b9c98e608-kube-api-access-7t6bv\") pod \"ceilometer-0\" (UID: \"f40cb730-942e-48c2-8bd5-4b1b9c98e608\") " pod="openstack/ceilometer-0" Nov 25 15:22:37 crc kubenswrapper[4890]: I1125 15:22:37.865153 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f40cb730-942e-48c2-8bd5-4b1b9c98e608-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f40cb730-942e-48c2-8bd5-4b1b9c98e608\") " pod="openstack/ceilometer-0" Nov 25 15:22:37 crc kubenswrapper[4890]: I1125 15:22:37.865485 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f40cb730-942e-48c2-8bd5-4b1b9c98e608-log-httpd\") pod \"ceilometer-0\" (UID: \"f40cb730-942e-48c2-8bd5-4b1b9c98e608\") " pod="openstack/ceilometer-0" Nov 25 15:22:37 crc kubenswrapper[4890]: I1125 15:22:37.865618 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f40cb730-942e-48c2-8bd5-4b1b9c98e608-run-httpd\") pod \"ceilometer-0\" (UID: \"f40cb730-942e-48c2-8bd5-4b1b9c98e608\") " pod="openstack/ceilometer-0" Nov 25 15:22:37 crc kubenswrapper[4890]: I1125 15:22:37.871741 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f40cb730-942e-48c2-8bd5-4b1b9c98e608-scripts\") pod \"ceilometer-0\" (UID: \"f40cb730-942e-48c2-8bd5-4b1b9c98e608\") " pod="openstack/ceilometer-0" Nov 25 15:22:37 crc kubenswrapper[4890]: I1125 15:22:37.871834 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f40cb730-942e-48c2-8bd5-4b1b9c98e608-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f40cb730-942e-48c2-8bd5-4b1b9c98e608\") " pod="openstack/ceilometer-0" Nov 25 15:22:37 crc kubenswrapper[4890]: I1125 15:22:37.872895 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f40cb730-942e-48c2-8bd5-4b1b9c98e608-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f40cb730-942e-48c2-8bd5-4b1b9c98e608\") " pod="openstack/ceilometer-0" Nov 25 15:22:37 crc kubenswrapper[4890]: I1125 15:22:37.873005 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f40cb730-942e-48c2-8bd5-4b1b9c98e608-config-data\") pod \"ceilometer-0\" (UID: \"f40cb730-942e-48c2-8bd5-4b1b9c98e608\") " pod="openstack/ceilometer-0" Nov 25 15:22:37 crc kubenswrapper[4890]: I1125 15:22:37.882993 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7t6bv\" (UniqueName: \"kubernetes.io/projected/f40cb730-942e-48c2-8bd5-4b1b9c98e608-kube-api-access-7t6bv\") pod \"ceilometer-0\" (UID: \"f40cb730-942e-48c2-8bd5-4b1b9c98e608\") " pod="openstack/ceilometer-0" Nov 25 15:22:37 crc kubenswrapper[4890]: I1125 15:22:37.988883 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 15:22:38 crc kubenswrapper[4890]: I1125 15:22:38.182950 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aaf63d02-9407-4093-8758-29ebe1caf5d8" path="/var/lib/kubelet/pods/aaf63d02-9407-4093-8758-29ebe1caf5d8/volumes" Nov 25 15:22:38 crc kubenswrapper[4890]: I1125 15:22:38.183816 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dbb94a13-5e59-489e-8f1a-3a75a1cbcece" path="/var/lib/kubelet/pods/dbb94a13-5e59-489e-8f1a-3a75a1cbcece/volumes" Nov 25 15:22:38 crc kubenswrapper[4890]: I1125 15:22:38.495540 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:22:38 crc kubenswrapper[4890]: I1125 15:22:38.567671 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f40cb730-942e-48c2-8bd5-4b1b9c98e608","Type":"ContainerStarted","Data":"a98905f20f1a69fa34a4efc6b71969134542453b9c04771f8160092cd956b20e"} Nov 25 15:22:39 crc kubenswrapper[4890]: I1125 15:22:39.590924 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f40cb730-942e-48c2-8bd5-4b1b9c98e608","Type":"ContainerStarted","Data":"7185cab16ba5a88ec74be5b5c46e19d70303e92c0117bcb50a92fd44153fcf57"} Nov 25 15:22:40 crc kubenswrapper[4890]: I1125 15:22:40.602281 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f40cb730-942e-48c2-8bd5-4b1b9c98e608","Type":"ContainerStarted","Data":"17ff51b5eb9a63fcee40bf814a40b33bf0ad95865cbb62f436c7e6dd452aaeee"} Nov 25 15:22:41 crc kubenswrapper[4890]: I1125 15:22:41.621149 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f40cb730-942e-48c2-8bd5-4b1b9c98e608","Type":"ContainerStarted","Data":"14d6c045577a49293ddbae7176231d13c17a436c0e7a7f9d56bc3829d7075e9e"} Nov 25 15:22:42 crc kubenswrapper[4890]: I1125 15:22:42.634209 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f40cb730-942e-48c2-8bd5-4b1b9c98e608","Type":"ContainerStarted","Data":"0abb23e0d0efe4d34e82a2e2f1746922424fbcc72a80f6cc99c5500a750ad497"} Nov 25 15:22:42 crc kubenswrapper[4890]: I1125 15:22:42.636630 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 25 15:22:42 crc kubenswrapper[4890]: I1125 15:22:42.658470 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.4507110069999998 podStartE2EDuration="5.658449789s" podCreationTimestamp="2025-11-25 15:22:37 +0000 UTC" firstStartedPulling="2025-11-25 15:22:38.495016518 +0000 UTC m=+1216.937479128" lastFinishedPulling="2025-11-25 15:22:41.7027553 +0000 UTC m=+1220.145217910" observedRunningTime="2025-11-25 15:22:42.657283269 +0000 UTC m=+1221.099745899" watchObservedRunningTime="2025-11-25 15:22:42.658449789 +0000 UTC m=+1221.100912399" Nov 25 15:22:42 crc kubenswrapper[4890]: I1125 15:22:42.738458 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:22:44 crc kubenswrapper[4890]: I1125 15:22:44.652155 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f40cb730-942e-48c2-8bd5-4b1b9c98e608" containerName="ceilometer-central-agent" containerID="cri-o://7185cab16ba5a88ec74be5b5c46e19d70303e92c0117bcb50a92fd44153fcf57" gracePeriod=30 Nov 25 15:22:44 crc kubenswrapper[4890]: I1125 15:22:44.652267 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f40cb730-942e-48c2-8bd5-4b1b9c98e608" containerName="proxy-httpd" containerID="cri-o://0abb23e0d0efe4d34e82a2e2f1746922424fbcc72a80f6cc99c5500a750ad497" gracePeriod=30 Nov 25 15:22:44 crc kubenswrapper[4890]: I1125 15:22:44.652270 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f40cb730-942e-48c2-8bd5-4b1b9c98e608" containerName="ceilometer-notification-agent" containerID="cri-o://17ff51b5eb9a63fcee40bf814a40b33bf0ad95865cbb62f436c7e6dd452aaeee" gracePeriod=30 Nov 25 15:22:44 crc kubenswrapper[4890]: I1125 15:22:44.652215 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f40cb730-942e-48c2-8bd5-4b1b9c98e608" containerName="sg-core" containerID="cri-o://14d6c045577a49293ddbae7176231d13c17a436c0e7a7f9d56bc3829d7075e9e" gracePeriod=30 Nov 25 15:22:45 crc kubenswrapper[4890]: I1125 15:22:45.664275 4890 generic.go:334] "Generic (PLEG): container finished" podID="f40cb730-942e-48c2-8bd5-4b1b9c98e608" containerID="0abb23e0d0efe4d34e82a2e2f1746922424fbcc72a80f6cc99c5500a750ad497" exitCode=0 Nov 25 15:22:45 crc kubenswrapper[4890]: I1125 15:22:45.665508 4890 generic.go:334] "Generic (PLEG): container finished" podID="f40cb730-942e-48c2-8bd5-4b1b9c98e608" containerID="14d6c045577a49293ddbae7176231d13c17a436c0e7a7f9d56bc3829d7075e9e" exitCode=2 Nov 25 15:22:45 crc kubenswrapper[4890]: I1125 15:22:45.665580 4890 generic.go:334] "Generic (PLEG): container finished" podID="f40cb730-942e-48c2-8bd5-4b1b9c98e608" containerID="17ff51b5eb9a63fcee40bf814a40b33bf0ad95865cbb62f436c7e6dd452aaeee" exitCode=0 Nov 25 15:22:45 crc kubenswrapper[4890]: I1125 15:22:45.664369 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f40cb730-942e-48c2-8bd5-4b1b9c98e608","Type":"ContainerDied","Data":"0abb23e0d0efe4d34e82a2e2f1746922424fbcc72a80f6cc99c5500a750ad497"} Nov 25 15:22:45 crc kubenswrapper[4890]: I1125 15:22:45.665771 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f40cb730-942e-48c2-8bd5-4b1b9c98e608","Type":"ContainerDied","Data":"14d6c045577a49293ddbae7176231d13c17a436c0e7a7f9d56bc3829d7075e9e"} Nov 25 15:22:45 crc kubenswrapper[4890]: I1125 15:22:45.665870 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f40cb730-942e-48c2-8bd5-4b1b9c98e608","Type":"ContainerDied","Data":"17ff51b5eb9a63fcee40bf814a40b33bf0ad95865cbb62f436c7e6dd452aaeee"} Nov 25 15:22:47 crc kubenswrapper[4890]: I1125 15:22:47.685791 4890 generic.go:334] "Generic (PLEG): container finished" podID="f40cb730-942e-48c2-8bd5-4b1b9c98e608" containerID="7185cab16ba5a88ec74be5b5c46e19d70303e92c0117bcb50a92fd44153fcf57" exitCode=0 Nov 25 15:22:47 crc kubenswrapper[4890]: I1125 15:22:47.686098 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f40cb730-942e-48c2-8bd5-4b1b9c98e608","Type":"ContainerDied","Data":"7185cab16ba5a88ec74be5b5c46e19d70303e92c0117bcb50a92fd44153fcf57"} Nov 25 15:22:47 crc kubenswrapper[4890]: I1125 15:22:47.910045 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 15:22:47 crc kubenswrapper[4890]: I1125 15:22:47.983439 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f40cb730-942e-48c2-8bd5-4b1b9c98e608-log-httpd\") pod \"f40cb730-942e-48c2-8bd5-4b1b9c98e608\" (UID: \"f40cb730-942e-48c2-8bd5-4b1b9c98e608\") " Nov 25 15:22:47 crc kubenswrapper[4890]: I1125 15:22:47.983578 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f40cb730-942e-48c2-8bd5-4b1b9c98e608-sg-core-conf-yaml\") pod \"f40cb730-942e-48c2-8bd5-4b1b9c98e608\" (UID: \"f40cb730-942e-48c2-8bd5-4b1b9c98e608\") " Nov 25 15:22:47 crc kubenswrapper[4890]: I1125 15:22:47.983643 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7t6bv\" (UniqueName: \"kubernetes.io/projected/f40cb730-942e-48c2-8bd5-4b1b9c98e608-kube-api-access-7t6bv\") pod \"f40cb730-942e-48c2-8bd5-4b1b9c98e608\" (UID: \"f40cb730-942e-48c2-8bd5-4b1b9c98e608\") " Nov 25 15:22:47 crc kubenswrapper[4890]: I1125 15:22:47.986065 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f40cb730-942e-48c2-8bd5-4b1b9c98e608-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f40cb730-942e-48c2-8bd5-4b1b9c98e608" (UID: "f40cb730-942e-48c2-8bd5-4b1b9c98e608"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:22:47 crc kubenswrapper[4890]: I1125 15:22:47.997646 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f40cb730-942e-48c2-8bd5-4b1b9c98e608-kube-api-access-7t6bv" (OuterVolumeSpecName: "kube-api-access-7t6bv") pod "f40cb730-942e-48c2-8bd5-4b1b9c98e608" (UID: "f40cb730-942e-48c2-8bd5-4b1b9c98e608"). InnerVolumeSpecName "kube-api-access-7t6bv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:22:48 crc kubenswrapper[4890]: I1125 15:22:48.021288 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f40cb730-942e-48c2-8bd5-4b1b9c98e608-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f40cb730-942e-48c2-8bd5-4b1b9c98e608" (UID: "f40cb730-942e-48c2-8bd5-4b1b9c98e608"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:22:48 crc kubenswrapper[4890]: I1125 15:22:48.086395 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f40cb730-942e-48c2-8bd5-4b1b9c98e608-scripts\") pod \"f40cb730-942e-48c2-8bd5-4b1b9c98e608\" (UID: \"f40cb730-942e-48c2-8bd5-4b1b9c98e608\") " Nov 25 15:22:48 crc kubenswrapper[4890]: I1125 15:22:48.086447 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f40cb730-942e-48c2-8bd5-4b1b9c98e608-combined-ca-bundle\") pod \"f40cb730-942e-48c2-8bd5-4b1b9c98e608\" (UID: \"f40cb730-942e-48c2-8bd5-4b1b9c98e608\") " Nov 25 15:22:48 crc kubenswrapper[4890]: I1125 15:22:48.086615 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f40cb730-942e-48c2-8bd5-4b1b9c98e608-config-data\") pod \"f40cb730-942e-48c2-8bd5-4b1b9c98e608\" (UID: \"f40cb730-942e-48c2-8bd5-4b1b9c98e608\") " Nov 25 15:22:48 crc kubenswrapper[4890]: I1125 15:22:48.087013 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f40cb730-942e-48c2-8bd5-4b1b9c98e608-run-httpd\") pod \"f40cb730-942e-48c2-8bd5-4b1b9c98e608\" (UID: \"f40cb730-942e-48c2-8bd5-4b1b9c98e608\") " Nov 25 15:22:48 crc kubenswrapper[4890]: I1125 15:22:48.087655 4890 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f40cb730-942e-48c2-8bd5-4b1b9c98e608-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 15:22:48 crc kubenswrapper[4890]: I1125 15:22:48.087687 4890 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f40cb730-942e-48c2-8bd5-4b1b9c98e608-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 25 15:22:48 crc kubenswrapper[4890]: I1125 15:22:48.087704 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7t6bv\" (UniqueName: \"kubernetes.io/projected/f40cb730-942e-48c2-8bd5-4b1b9c98e608-kube-api-access-7t6bv\") on node \"crc\" DevicePath \"\"" Nov 25 15:22:48 crc kubenswrapper[4890]: I1125 15:22:48.088378 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f40cb730-942e-48c2-8bd5-4b1b9c98e608-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f40cb730-942e-48c2-8bd5-4b1b9c98e608" (UID: "f40cb730-942e-48c2-8bd5-4b1b9c98e608"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:22:48 crc kubenswrapper[4890]: I1125 15:22:48.090309 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f40cb730-942e-48c2-8bd5-4b1b9c98e608-scripts" (OuterVolumeSpecName: "scripts") pod "f40cb730-942e-48c2-8bd5-4b1b9c98e608" (UID: "f40cb730-942e-48c2-8bd5-4b1b9c98e608"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:22:48 crc kubenswrapper[4890]: I1125 15:22:48.159682 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f40cb730-942e-48c2-8bd5-4b1b9c98e608-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f40cb730-942e-48c2-8bd5-4b1b9c98e608" (UID: "f40cb730-942e-48c2-8bd5-4b1b9c98e608"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:22:48 crc kubenswrapper[4890]: I1125 15:22:48.183690 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f40cb730-942e-48c2-8bd5-4b1b9c98e608-config-data" (OuterVolumeSpecName: "config-data") pod "f40cb730-942e-48c2-8bd5-4b1b9c98e608" (UID: "f40cb730-942e-48c2-8bd5-4b1b9c98e608"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:22:48 crc kubenswrapper[4890]: I1125 15:22:48.188736 4890 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f40cb730-942e-48c2-8bd5-4b1b9c98e608-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 15:22:48 crc kubenswrapper[4890]: I1125 15:22:48.188790 4890 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f40cb730-942e-48c2-8bd5-4b1b9c98e608-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 15:22:48 crc kubenswrapper[4890]: I1125 15:22:48.188806 4890 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f40cb730-942e-48c2-8bd5-4b1b9c98e608-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:22:48 crc kubenswrapper[4890]: I1125 15:22:48.188823 4890 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f40cb730-942e-48c2-8bd5-4b1b9c98e608-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 15:22:48 crc kubenswrapper[4890]: I1125 15:22:48.703145 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f40cb730-942e-48c2-8bd5-4b1b9c98e608","Type":"ContainerDied","Data":"a98905f20f1a69fa34a4efc6b71969134542453b9c04771f8160092cd956b20e"} Nov 25 15:22:48 crc kubenswrapper[4890]: I1125 15:22:48.703264 4890 scope.go:117] "RemoveContainer" containerID="0abb23e0d0efe4d34e82a2e2f1746922424fbcc72a80f6cc99c5500a750ad497" Nov 25 15:22:48 crc kubenswrapper[4890]: I1125 15:22:48.703282 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 15:22:48 crc kubenswrapper[4890]: I1125 15:22:48.734046 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:22:48 crc kubenswrapper[4890]: I1125 15:22:48.736982 4890 scope.go:117] "RemoveContainer" containerID="14d6c045577a49293ddbae7176231d13c17a436c0e7a7f9d56bc3829d7075e9e" Nov 25 15:22:48 crc kubenswrapper[4890]: I1125 15:22:48.764050 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:22:48 crc kubenswrapper[4890]: I1125 15:22:48.785440 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:22:48 crc kubenswrapper[4890]: E1125 15:22:48.787042 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f40cb730-942e-48c2-8bd5-4b1b9c98e608" containerName="proxy-httpd" Nov 25 15:22:48 crc kubenswrapper[4890]: I1125 15:22:48.787088 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="f40cb730-942e-48c2-8bd5-4b1b9c98e608" containerName="proxy-httpd" Nov 25 15:22:48 crc kubenswrapper[4890]: E1125 15:22:48.787189 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f40cb730-942e-48c2-8bd5-4b1b9c98e608" containerName="ceilometer-notification-agent" Nov 25 15:22:48 crc kubenswrapper[4890]: I1125 15:22:48.787201 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="f40cb730-942e-48c2-8bd5-4b1b9c98e608" containerName="ceilometer-notification-agent" Nov 25 15:22:48 crc kubenswrapper[4890]: E1125 15:22:48.787238 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f40cb730-942e-48c2-8bd5-4b1b9c98e608" containerName="sg-core" Nov 25 15:22:48 crc kubenswrapper[4890]: I1125 15:22:48.787251 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="f40cb730-942e-48c2-8bd5-4b1b9c98e608" containerName="sg-core" Nov 25 15:22:48 crc kubenswrapper[4890]: E1125 15:22:48.787294 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f40cb730-942e-48c2-8bd5-4b1b9c98e608" containerName="ceilometer-central-agent" Nov 25 15:22:48 crc kubenswrapper[4890]: I1125 15:22:48.787306 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="f40cb730-942e-48c2-8bd5-4b1b9c98e608" containerName="ceilometer-central-agent" Nov 25 15:22:48 crc kubenswrapper[4890]: I1125 15:22:48.788110 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="f40cb730-942e-48c2-8bd5-4b1b9c98e608" containerName="proxy-httpd" Nov 25 15:22:48 crc kubenswrapper[4890]: I1125 15:22:48.788139 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="f40cb730-942e-48c2-8bd5-4b1b9c98e608" containerName="ceilometer-central-agent" Nov 25 15:22:48 crc kubenswrapper[4890]: I1125 15:22:48.788180 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="f40cb730-942e-48c2-8bd5-4b1b9c98e608" containerName="ceilometer-notification-agent" Nov 25 15:22:48 crc kubenswrapper[4890]: I1125 15:22:48.788206 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="f40cb730-942e-48c2-8bd5-4b1b9c98e608" containerName="sg-core" Nov 25 15:22:48 crc kubenswrapper[4890]: I1125 15:22:48.794309 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 15:22:48 crc kubenswrapper[4890]: I1125 15:22:48.797076 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 25 15:22:48 crc kubenswrapper[4890]: I1125 15:22:48.797131 4890 scope.go:117] "RemoveContainer" containerID="17ff51b5eb9a63fcee40bf814a40b33bf0ad95865cbb62f436c7e6dd452aaeee" Nov 25 15:22:48 crc kubenswrapper[4890]: I1125 15:22:48.799483 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 25 15:22:48 crc kubenswrapper[4890]: I1125 15:22:48.803521 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:22:48 crc kubenswrapper[4890]: I1125 15:22:48.828456 4890 scope.go:117] "RemoveContainer" containerID="7185cab16ba5a88ec74be5b5c46e19d70303e92c0117bcb50a92fd44153fcf57" Nov 25 15:22:48 crc kubenswrapper[4890]: I1125 15:22:48.903586 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/936ff525-5705-492b-8262-dc886cd0ccfa-log-httpd\") pod \"ceilometer-0\" (UID: \"936ff525-5705-492b-8262-dc886cd0ccfa\") " pod="openstack/ceilometer-0" Nov 25 15:22:48 crc kubenswrapper[4890]: I1125 15:22:48.903658 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/936ff525-5705-492b-8262-dc886cd0ccfa-config-data\") pod \"ceilometer-0\" (UID: \"936ff525-5705-492b-8262-dc886cd0ccfa\") " pod="openstack/ceilometer-0" Nov 25 15:22:48 crc kubenswrapper[4890]: I1125 15:22:48.903718 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/936ff525-5705-492b-8262-dc886cd0ccfa-scripts\") pod \"ceilometer-0\" (UID: \"936ff525-5705-492b-8262-dc886cd0ccfa\") " pod="openstack/ceilometer-0" Nov 25 15:22:48 crc kubenswrapper[4890]: I1125 15:22:48.903756 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/936ff525-5705-492b-8262-dc886cd0ccfa-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"936ff525-5705-492b-8262-dc886cd0ccfa\") " pod="openstack/ceilometer-0" Nov 25 15:22:48 crc kubenswrapper[4890]: I1125 15:22:48.903780 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/936ff525-5705-492b-8262-dc886cd0ccfa-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"936ff525-5705-492b-8262-dc886cd0ccfa\") " pod="openstack/ceilometer-0" Nov 25 15:22:48 crc kubenswrapper[4890]: I1125 15:22:48.903807 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/936ff525-5705-492b-8262-dc886cd0ccfa-run-httpd\") pod \"ceilometer-0\" (UID: \"936ff525-5705-492b-8262-dc886cd0ccfa\") " pod="openstack/ceilometer-0" Nov 25 15:22:48 crc kubenswrapper[4890]: I1125 15:22:48.903940 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lptvc\" (UniqueName: \"kubernetes.io/projected/936ff525-5705-492b-8262-dc886cd0ccfa-kube-api-access-lptvc\") pod \"ceilometer-0\" (UID: \"936ff525-5705-492b-8262-dc886cd0ccfa\") " pod="openstack/ceilometer-0" Nov 25 15:22:49 crc kubenswrapper[4890]: I1125 15:22:49.005879 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/936ff525-5705-492b-8262-dc886cd0ccfa-config-data\") pod \"ceilometer-0\" (UID: \"936ff525-5705-492b-8262-dc886cd0ccfa\") " pod="openstack/ceilometer-0" Nov 25 15:22:49 crc kubenswrapper[4890]: I1125 15:22:49.005975 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/936ff525-5705-492b-8262-dc886cd0ccfa-scripts\") pod \"ceilometer-0\" (UID: \"936ff525-5705-492b-8262-dc886cd0ccfa\") " pod="openstack/ceilometer-0" Nov 25 15:22:49 crc kubenswrapper[4890]: I1125 15:22:49.006033 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/936ff525-5705-492b-8262-dc886cd0ccfa-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"936ff525-5705-492b-8262-dc886cd0ccfa\") " pod="openstack/ceilometer-0" Nov 25 15:22:49 crc kubenswrapper[4890]: I1125 15:22:49.006053 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/936ff525-5705-492b-8262-dc886cd0ccfa-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"936ff525-5705-492b-8262-dc886cd0ccfa\") " pod="openstack/ceilometer-0" Nov 25 15:22:49 crc kubenswrapper[4890]: I1125 15:22:49.006073 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/936ff525-5705-492b-8262-dc886cd0ccfa-run-httpd\") pod \"ceilometer-0\" (UID: \"936ff525-5705-492b-8262-dc886cd0ccfa\") " pod="openstack/ceilometer-0" Nov 25 15:22:49 crc kubenswrapper[4890]: I1125 15:22:49.006122 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lptvc\" (UniqueName: \"kubernetes.io/projected/936ff525-5705-492b-8262-dc886cd0ccfa-kube-api-access-lptvc\") pod \"ceilometer-0\" (UID: \"936ff525-5705-492b-8262-dc886cd0ccfa\") " pod="openstack/ceilometer-0" Nov 25 15:22:49 crc kubenswrapper[4890]: I1125 15:22:49.006188 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/936ff525-5705-492b-8262-dc886cd0ccfa-log-httpd\") pod \"ceilometer-0\" (UID: \"936ff525-5705-492b-8262-dc886cd0ccfa\") " pod="openstack/ceilometer-0" Nov 25 15:22:49 crc kubenswrapper[4890]: I1125 15:22:49.006603 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/936ff525-5705-492b-8262-dc886cd0ccfa-log-httpd\") pod \"ceilometer-0\" (UID: \"936ff525-5705-492b-8262-dc886cd0ccfa\") " pod="openstack/ceilometer-0" Nov 25 15:22:49 crc kubenswrapper[4890]: I1125 15:22:49.006714 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/936ff525-5705-492b-8262-dc886cd0ccfa-run-httpd\") pod \"ceilometer-0\" (UID: \"936ff525-5705-492b-8262-dc886cd0ccfa\") " pod="openstack/ceilometer-0" Nov 25 15:22:49 crc kubenswrapper[4890]: I1125 15:22:49.009921 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/936ff525-5705-492b-8262-dc886cd0ccfa-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"936ff525-5705-492b-8262-dc886cd0ccfa\") " pod="openstack/ceilometer-0" Nov 25 15:22:49 crc kubenswrapper[4890]: I1125 15:22:49.010722 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/936ff525-5705-492b-8262-dc886cd0ccfa-config-data\") pod \"ceilometer-0\" (UID: \"936ff525-5705-492b-8262-dc886cd0ccfa\") " pod="openstack/ceilometer-0" Nov 25 15:22:49 crc kubenswrapper[4890]: I1125 15:22:49.011095 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/936ff525-5705-492b-8262-dc886cd0ccfa-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"936ff525-5705-492b-8262-dc886cd0ccfa\") " pod="openstack/ceilometer-0" Nov 25 15:22:49 crc kubenswrapper[4890]: I1125 15:22:49.020782 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/936ff525-5705-492b-8262-dc886cd0ccfa-scripts\") pod \"ceilometer-0\" (UID: \"936ff525-5705-492b-8262-dc886cd0ccfa\") " pod="openstack/ceilometer-0" Nov 25 15:22:49 crc kubenswrapper[4890]: I1125 15:22:49.023235 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lptvc\" (UniqueName: \"kubernetes.io/projected/936ff525-5705-492b-8262-dc886cd0ccfa-kube-api-access-lptvc\") pod \"ceilometer-0\" (UID: \"936ff525-5705-492b-8262-dc886cd0ccfa\") " pod="openstack/ceilometer-0" Nov 25 15:22:49 crc kubenswrapper[4890]: I1125 15:22:49.119764 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 15:22:49 crc kubenswrapper[4890]: I1125 15:22:49.555597 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:22:49 crc kubenswrapper[4890]: I1125 15:22:49.714324 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"936ff525-5705-492b-8262-dc886cd0ccfa","Type":"ContainerStarted","Data":"9c4bd9e2d7855c2478ee5b38022cafc88c168a8cb03c2308b367d31d0d5bec96"} Nov 25 15:22:50 crc kubenswrapper[4890]: I1125 15:22:50.183744 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f40cb730-942e-48c2-8bd5-4b1b9c98e608" path="/var/lib/kubelet/pods/f40cb730-942e-48c2-8bd5-4b1b9c98e608/volumes" Nov 25 15:22:51 crc kubenswrapper[4890]: I1125 15:22:51.740719 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"936ff525-5705-492b-8262-dc886cd0ccfa","Type":"ContainerStarted","Data":"e1d0efa602b582a36d30b22f424477e447da02cbf0301407dfc60beceffddfe4"} Nov 25 15:22:52 crc kubenswrapper[4890]: I1125 15:22:52.771060 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"936ff525-5705-492b-8262-dc886cd0ccfa","Type":"ContainerStarted","Data":"dcace7be575ca39c462893475e6bbccce00e1f0de451fabe14cb9ef826a5dbd7"} Nov 25 15:22:53 crc kubenswrapper[4890]: I1125 15:22:53.779882 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"936ff525-5705-492b-8262-dc886cd0ccfa","Type":"ContainerStarted","Data":"6c5b96bfb343f6da2c3f5521049b204f2b60420ceb97afee565b4bb5a609af83"} Nov 25 15:22:54 crc kubenswrapper[4890]: I1125 15:22:54.796005 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"936ff525-5705-492b-8262-dc886cd0ccfa","Type":"ContainerStarted","Data":"663454459ec9d503abbc83dafe5617cd37228e8a850da5b7b16f3ca8dece0567"} Nov 25 15:22:54 crc kubenswrapper[4890]: I1125 15:22:54.796369 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 25 15:22:54 crc kubenswrapper[4890]: I1125 15:22:54.830407 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.10204976 podStartE2EDuration="6.830389083s" podCreationTimestamp="2025-11-25 15:22:48 +0000 UTC" firstStartedPulling="2025-11-25 15:22:49.564091257 +0000 UTC m=+1228.006553867" lastFinishedPulling="2025-11-25 15:22:54.29243058 +0000 UTC m=+1232.734893190" observedRunningTime="2025-11-25 15:22:54.821834998 +0000 UTC m=+1233.264297608" watchObservedRunningTime="2025-11-25 15:22:54.830389083 +0000 UTC m=+1233.272851693" Nov 25 15:23:02 crc kubenswrapper[4890]: I1125 15:23:02.875825 4890 generic.go:334] "Generic (PLEG): container finished" podID="910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0" containerID="b317dbbc6ec2fb8dfd39a367f59796a433cbc2e2182fcfcad92df15433c00b42" exitCode=0 Nov 25 15:23:02 crc kubenswrapper[4890]: I1125 15:23:02.875918 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-mlvqz" event={"ID":"910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0","Type":"ContainerDied","Data":"b317dbbc6ec2fb8dfd39a367f59796a433cbc2e2182fcfcad92df15433c00b42"} Nov 25 15:23:04 crc kubenswrapper[4890]: I1125 15:23:04.241626 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-mlvqz" Nov 25 15:23:04 crc kubenswrapper[4890]: I1125 15:23:04.381464 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0-config-data\") pod \"910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0\" (UID: \"910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0\") " Nov 25 15:23:04 crc kubenswrapper[4890]: I1125 15:23:04.381832 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0-combined-ca-bundle\") pod \"910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0\" (UID: \"910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0\") " Nov 25 15:23:04 crc kubenswrapper[4890]: I1125 15:23:04.382043 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0-scripts\") pod \"910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0\" (UID: \"910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0\") " Nov 25 15:23:04 crc kubenswrapper[4890]: I1125 15:23:04.382225 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mbqv8\" (UniqueName: \"kubernetes.io/projected/910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0-kube-api-access-mbqv8\") pod \"910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0\" (UID: \"910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0\") " Nov 25 15:23:04 crc kubenswrapper[4890]: I1125 15:23:04.386944 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0-scripts" (OuterVolumeSpecName: "scripts") pod "910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0" (UID: "910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:23:04 crc kubenswrapper[4890]: I1125 15:23:04.388405 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0-kube-api-access-mbqv8" (OuterVolumeSpecName: "kube-api-access-mbqv8") pod "910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0" (UID: "910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0"). InnerVolumeSpecName "kube-api-access-mbqv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:23:04 crc kubenswrapper[4890]: I1125 15:23:04.409400 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0-config-data" (OuterVolumeSpecName: "config-data") pod "910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0" (UID: "910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:23:04 crc kubenswrapper[4890]: I1125 15:23:04.409969 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0" (UID: "910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:23:04 crc kubenswrapper[4890]: I1125 15:23:04.484458 4890 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 15:23:04 crc kubenswrapper[4890]: I1125 15:23:04.484499 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mbqv8\" (UniqueName: \"kubernetes.io/projected/910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0-kube-api-access-mbqv8\") on node \"crc\" DevicePath \"\"" Nov 25 15:23:04 crc kubenswrapper[4890]: I1125 15:23:04.484513 4890 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 15:23:04 crc kubenswrapper[4890]: I1125 15:23:04.484523 4890 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:23:04 crc kubenswrapper[4890]: I1125 15:23:04.896096 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-mlvqz" event={"ID":"910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0","Type":"ContainerDied","Data":"15611717d8cf7ee7164c9af8f15896d018e145df2180ab4abda3052da41e4310"} Nov 25 15:23:04 crc kubenswrapper[4890]: I1125 15:23:04.896135 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="15611717d8cf7ee7164c9af8f15896d018e145df2180ab4abda3052da41e4310" Nov 25 15:23:04 crc kubenswrapper[4890]: I1125 15:23:04.896190 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-mlvqz" Nov 25 15:23:05 crc kubenswrapper[4890]: I1125 15:23:05.168637 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 25 15:23:05 crc kubenswrapper[4890]: E1125 15:23:05.170308 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0" containerName="nova-cell0-conductor-db-sync" Nov 25 15:23:05 crc kubenswrapper[4890]: I1125 15:23:05.171121 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0" containerName="nova-cell0-conductor-db-sync" Nov 25 15:23:05 crc kubenswrapper[4890]: I1125 15:23:05.171447 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0" containerName="nova-cell0-conductor-db-sync" Nov 25 15:23:05 crc kubenswrapper[4890]: I1125 15:23:05.172104 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 25 15:23:05 crc kubenswrapper[4890]: I1125 15:23:05.174427 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-qvvbs" Nov 25 15:23:05 crc kubenswrapper[4890]: I1125 15:23:05.174617 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 25 15:23:05 crc kubenswrapper[4890]: I1125 15:23:05.197670 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7929bde-463b-4dff-b685-db4bf8368909-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"f7929bde-463b-4dff-b685-db4bf8368909\") " pod="openstack/nova-cell0-conductor-0" Nov 25 15:23:05 crc kubenswrapper[4890]: I1125 15:23:05.197732 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmt6c\" (UniqueName: \"kubernetes.io/projected/f7929bde-463b-4dff-b685-db4bf8368909-kube-api-access-pmt6c\") pod \"nova-cell0-conductor-0\" (UID: \"f7929bde-463b-4dff-b685-db4bf8368909\") " pod="openstack/nova-cell0-conductor-0" Nov 25 15:23:05 crc kubenswrapper[4890]: I1125 15:23:05.197779 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7929bde-463b-4dff-b685-db4bf8368909-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"f7929bde-463b-4dff-b685-db4bf8368909\") " pod="openstack/nova-cell0-conductor-0" Nov 25 15:23:05 crc kubenswrapper[4890]: I1125 15:23:05.197858 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 25 15:23:05 crc kubenswrapper[4890]: I1125 15:23:05.299124 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7929bde-463b-4dff-b685-db4bf8368909-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"f7929bde-463b-4dff-b685-db4bf8368909\") " pod="openstack/nova-cell0-conductor-0" Nov 25 15:23:05 crc kubenswrapper[4890]: I1125 15:23:05.299203 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmt6c\" (UniqueName: \"kubernetes.io/projected/f7929bde-463b-4dff-b685-db4bf8368909-kube-api-access-pmt6c\") pod \"nova-cell0-conductor-0\" (UID: \"f7929bde-463b-4dff-b685-db4bf8368909\") " pod="openstack/nova-cell0-conductor-0" Nov 25 15:23:05 crc kubenswrapper[4890]: I1125 15:23:05.299255 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7929bde-463b-4dff-b685-db4bf8368909-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"f7929bde-463b-4dff-b685-db4bf8368909\") " pod="openstack/nova-cell0-conductor-0" Nov 25 15:23:05 crc kubenswrapper[4890]: I1125 15:23:05.304811 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7929bde-463b-4dff-b685-db4bf8368909-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"f7929bde-463b-4dff-b685-db4bf8368909\") " pod="openstack/nova-cell0-conductor-0" Nov 25 15:23:05 crc kubenswrapper[4890]: I1125 15:23:05.309962 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7929bde-463b-4dff-b685-db4bf8368909-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"f7929bde-463b-4dff-b685-db4bf8368909\") " pod="openstack/nova-cell0-conductor-0" Nov 25 15:23:05 crc kubenswrapper[4890]: I1125 15:23:05.319271 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmt6c\" (UniqueName: \"kubernetes.io/projected/f7929bde-463b-4dff-b685-db4bf8368909-kube-api-access-pmt6c\") pod \"nova-cell0-conductor-0\" (UID: \"f7929bde-463b-4dff-b685-db4bf8368909\") " pod="openstack/nova-cell0-conductor-0" Nov 25 15:23:05 crc kubenswrapper[4890]: I1125 15:23:05.492657 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 25 15:23:06 crc kubenswrapper[4890]: I1125 15:23:06.001475 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 25 15:23:06 crc kubenswrapper[4890]: W1125 15:23:06.008896 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf7929bde_463b_4dff_b685_db4bf8368909.slice/crio-f38317574f9651942fb837cbf5208df912d42d928e8b1a288f7478e2d2eae077 WatchSource:0}: Error finding container f38317574f9651942fb837cbf5208df912d42d928e8b1a288f7478e2d2eae077: Status 404 returned error can't find the container with id f38317574f9651942fb837cbf5208df912d42d928e8b1a288f7478e2d2eae077 Nov 25 15:23:06 crc kubenswrapper[4890]: I1125 15:23:06.916334 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"f7929bde-463b-4dff-b685-db4bf8368909","Type":"ContainerStarted","Data":"99afa6dbc91a0fe549f4ec14bfec500df11fa72b489a1c8dcb77c15059e7b19d"} Nov 25 15:23:06 crc kubenswrapper[4890]: I1125 15:23:06.916691 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"f7929bde-463b-4dff-b685-db4bf8368909","Type":"ContainerStarted","Data":"f38317574f9651942fb837cbf5208df912d42d928e8b1a288f7478e2d2eae077"} Nov 25 15:23:06 crc kubenswrapper[4890]: I1125 15:23:06.916984 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Nov 25 15:23:06 crc kubenswrapper[4890]: I1125 15:23:06.950745 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=1.950724562 podStartE2EDuration="1.950724562s" podCreationTimestamp="2025-11-25 15:23:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:23:06.93713037 +0000 UTC m=+1245.379592980" watchObservedRunningTime="2025-11-25 15:23:06.950724562 +0000 UTC m=+1245.393187172" Nov 25 15:23:15 crc kubenswrapper[4890]: I1125 15:23:15.529413 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.041356 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-t4rb6"] Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.042875 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-t4rb6" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.046105 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.046945 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.052462 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-t4rb6"] Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.214672 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/872e8100-002a-4ac1-b699-b35111573d7d-config-data\") pod \"nova-cell0-cell-mapping-t4rb6\" (UID: \"872e8100-002a-4ac1-b699-b35111573d7d\") " pod="openstack/nova-cell0-cell-mapping-t4rb6" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.214782 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hq2s\" (UniqueName: \"kubernetes.io/projected/872e8100-002a-4ac1-b699-b35111573d7d-kube-api-access-9hq2s\") pod \"nova-cell0-cell-mapping-t4rb6\" (UID: \"872e8100-002a-4ac1-b699-b35111573d7d\") " pod="openstack/nova-cell0-cell-mapping-t4rb6" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.214901 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/872e8100-002a-4ac1-b699-b35111573d7d-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-t4rb6\" (UID: \"872e8100-002a-4ac1-b699-b35111573d7d\") " pod="openstack/nova-cell0-cell-mapping-t4rb6" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.214938 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/872e8100-002a-4ac1-b699-b35111573d7d-scripts\") pod \"nova-cell0-cell-mapping-t4rb6\" (UID: \"872e8100-002a-4ac1-b699-b35111573d7d\") " pod="openstack/nova-cell0-cell-mapping-t4rb6" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.236210 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.251074 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.255601 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.257334 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.321059 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/872e8100-002a-4ac1-b699-b35111573d7d-scripts\") pod \"nova-cell0-cell-mapping-t4rb6\" (UID: \"872e8100-002a-4ac1-b699-b35111573d7d\") " pod="openstack/nova-cell0-cell-mapping-t4rb6" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.321437 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/872e8100-002a-4ac1-b699-b35111573d7d-config-data\") pod \"nova-cell0-cell-mapping-t4rb6\" (UID: \"872e8100-002a-4ac1-b699-b35111573d7d\") " pod="openstack/nova-cell0-cell-mapping-t4rb6" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.321540 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hq2s\" (UniqueName: \"kubernetes.io/projected/872e8100-002a-4ac1-b699-b35111573d7d-kube-api-access-9hq2s\") pod \"nova-cell0-cell-mapping-t4rb6\" (UID: \"872e8100-002a-4ac1-b699-b35111573d7d\") " pod="openstack/nova-cell0-cell-mapping-t4rb6" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.321666 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/872e8100-002a-4ac1-b699-b35111573d7d-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-t4rb6\" (UID: \"872e8100-002a-4ac1-b699-b35111573d7d\") " pod="openstack/nova-cell0-cell-mapping-t4rb6" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.364930 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/872e8100-002a-4ac1-b699-b35111573d7d-scripts\") pod \"nova-cell0-cell-mapping-t4rb6\" (UID: \"872e8100-002a-4ac1-b699-b35111573d7d\") " pod="openstack/nova-cell0-cell-mapping-t4rb6" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.369128 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/872e8100-002a-4ac1-b699-b35111573d7d-config-data\") pod \"nova-cell0-cell-mapping-t4rb6\" (UID: \"872e8100-002a-4ac1-b699-b35111573d7d\") " pod="openstack/nova-cell0-cell-mapping-t4rb6" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.369795 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/872e8100-002a-4ac1-b699-b35111573d7d-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-t4rb6\" (UID: \"872e8100-002a-4ac1-b699-b35111573d7d\") " pod="openstack/nova-cell0-cell-mapping-t4rb6" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.379835 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hq2s\" (UniqueName: \"kubernetes.io/projected/872e8100-002a-4ac1-b699-b35111573d7d-kube-api-access-9hq2s\") pod \"nova-cell0-cell-mapping-t4rb6\" (UID: \"872e8100-002a-4ac1-b699-b35111573d7d\") " pod="openstack/nova-cell0-cell-mapping-t4rb6" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.402569 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.404032 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.404972 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-t4rb6" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.417033 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.424655 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d2cf4a9-6672-4006-8d0d-6b380f442fc6-logs\") pod \"nova-metadata-0\" (UID: \"3d2cf4a9-6672-4006-8d0d-6b380f442fc6\") " pod="openstack/nova-metadata-0" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.424720 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bg8ns\" (UniqueName: \"kubernetes.io/projected/3d2cf4a9-6672-4006-8d0d-6b380f442fc6-kube-api-access-bg8ns\") pod \"nova-metadata-0\" (UID: \"3d2cf4a9-6672-4006-8d0d-6b380f442fc6\") " pod="openstack/nova-metadata-0" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.424824 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d2cf4a9-6672-4006-8d0d-6b380f442fc6-config-data\") pod \"nova-metadata-0\" (UID: \"3d2cf4a9-6672-4006-8d0d-6b380f442fc6\") " pod="openstack/nova-metadata-0" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.424901 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d2cf4a9-6672-4006-8d0d-6b380f442fc6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3d2cf4a9-6672-4006-8d0d-6b380f442fc6\") " pod="openstack/nova-metadata-0" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.434399 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-kjvns"] Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.436704 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-kjvns" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.449305 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.477323 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-kjvns"] Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.487762 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.489619 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.501627 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.520706 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.527650 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d2cf4a9-6672-4006-8d0d-6b380f442fc6-logs\") pod \"nova-metadata-0\" (UID: \"3d2cf4a9-6672-4006-8d0d-6b380f442fc6\") " pod="openstack/nova-metadata-0" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.527709 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7874ec1-b9cb-4fc3-a5af-002ceb718857-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"f7874ec1-b9cb-4fc3-a5af-002ceb718857\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.527750 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bg8ns\" (UniqueName: \"kubernetes.io/projected/3d2cf4a9-6672-4006-8d0d-6b380f442fc6-kube-api-access-bg8ns\") pod \"nova-metadata-0\" (UID: \"3d2cf4a9-6672-4006-8d0d-6b380f442fc6\") " pod="openstack/nova-metadata-0" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.527791 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lp6cx\" (UniqueName: \"kubernetes.io/projected/f7874ec1-b9cb-4fc3-a5af-002ceb718857-kube-api-access-lp6cx\") pod \"nova-cell1-novncproxy-0\" (UID: \"f7874ec1-b9cb-4fc3-a5af-002ceb718857\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.527826 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7874ec1-b9cb-4fc3-a5af-002ceb718857-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"f7874ec1-b9cb-4fc3-a5af-002ceb718857\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.527912 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d2cf4a9-6672-4006-8d0d-6b380f442fc6-config-data\") pod \"nova-metadata-0\" (UID: \"3d2cf4a9-6672-4006-8d0d-6b380f442fc6\") " pod="openstack/nova-metadata-0" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.528003 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d2cf4a9-6672-4006-8d0d-6b380f442fc6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3d2cf4a9-6672-4006-8d0d-6b380f442fc6\") " pod="openstack/nova-metadata-0" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.531768 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d2cf4a9-6672-4006-8d0d-6b380f442fc6-logs\") pod \"nova-metadata-0\" (UID: \"3d2cf4a9-6672-4006-8d0d-6b380f442fc6\") " pod="openstack/nova-metadata-0" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.536575 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d2cf4a9-6672-4006-8d0d-6b380f442fc6-config-data\") pod \"nova-metadata-0\" (UID: \"3d2cf4a9-6672-4006-8d0d-6b380f442fc6\") " pod="openstack/nova-metadata-0" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.536895 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d2cf4a9-6672-4006-8d0d-6b380f442fc6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3d2cf4a9-6672-4006-8d0d-6b380f442fc6\") " pod="openstack/nova-metadata-0" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.552747 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.554234 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.556477 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.579942 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bg8ns\" (UniqueName: \"kubernetes.io/projected/3d2cf4a9-6672-4006-8d0d-6b380f442fc6-kube-api-access-bg8ns\") pod \"nova-metadata-0\" (UID: \"3d2cf4a9-6672-4006-8d0d-6b380f442fc6\") " pod="openstack/nova-metadata-0" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.580891 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.593623 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.629816 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82rvh\" (UniqueName: \"kubernetes.io/projected/49e416d1-e37c-4293-b007-5ec30c040757-kube-api-access-82rvh\") pod \"nova-api-0\" (UID: \"49e416d1-e37c-4293-b007-5ec30c040757\") " pod="openstack/nova-api-0" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.630042 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m28qt\" (UniqueName: \"kubernetes.io/projected/9060524b-48eb-42ba-b1e7-5a852d89d94f-kube-api-access-m28qt\") pod \"dnsmasq-dns-757b4f8459-kjvns\" (UID: \"9060524b-48eb-42ba-b1e7-5a852d89d94f\") " pod="openstack/dnsmasq-dns-757b4f8459-kjvns" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.630069 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49e416d1-e37c-4293-b007-5ec30c040757-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"49e416d1-e37c-4293-b007-5ec30c040757\") " pod="openstack/nova-api-0" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.630141 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9060524b-48eb-42ba-b1e7-5a852d89d94f-config\") pod \"dnsmasq-dns-757b4f8459-kjvns\" (UID: \"9060524b-48eb-42ba-b1e7-5a852d89d94f\") " pod="openstack/dnsmasq-dns-757b4f8459-kjvns" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.630190 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9060524b-48eb-42ba-b1e7-5a852d89d94f-dns-svc\") pod \"dnsmasq-dns-757b4f8459-kjvns\" (UID: \"9060524b-48eb-42ba-b1e7-5a852d89d94f\") " pod="openstack/dnsmasq-dns-757b4f8459-kjvns" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.630222 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49e416d1-e37c-4293-b007-5ec30c040757-config-data\") pod \"nova-api-0\" (UID: \"49e416d1-e37c-4293-b007-5ec30c040757\") " pod="openstack/nova-api-0" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.630278 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/49e416d1-e37c-4293-b007-5ec30c040757-logs\") pod \"nova-api-0\" (UID: \"49e416d1-e37c-4293-b007-5ec30c040757\") " pod="openstack/nova-api-0" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.630317 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9060524b-48eb-42ba-b1e7-5a852d89d94f-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-kjvns\" (UID: \"9060524b-48eb-42ba-b1e7-5a852d89d94f\") " pod="openstack/dnsmasq-dns-757b4f8459-kjvns" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.630355 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7874ec1-b9cb-4fc3-a5af-002ceb718857-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"f7874ec1-b9cb-4fc3-a5af-002ceb718857\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.630405 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lp6cx\" (UniqueName: \"kubernetes.io/projected/f7874ec1-b9cb-4fc3-a5af-002ceb718857-kube-api-access-lp6cx\") pod \"nova-cell1-novncproxy-0\" (UID: \"f7874ec1-b9cb-4fc3-a5af-002ceb718857\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.630441 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7874ec1-b9cb-4fc3-a5af-002ceb718857-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"f7874ec1-b9cb-4fc3-a5af-002ceb718857\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.630463 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9060524b-48eb-42ba-b1e7-5a852d89d94f-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-kjvns\" (UID: \"9060524b-48eb-42ba-b1e7-5a852d89d94f\") " pod="openstack/dnsmasq-dns-757b4f8459-kjvns" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.630537 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9060524b-48eb-42ba-b1e7-5a852d89d94f-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-kjvns\" (UID: \"9060524b-48eb-42ba-b1e7-5a852d89d94f\") " pod="openstack/dnsmasq-dns-757b4f8459-kjvns" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.641137 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7874ec1-b9cb-4fc3-a5af-002ceb718857-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"f7874ec1-b9cb-4fc3-a5af-002ceb718857\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.643111 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7874ec1-b9cb-4fc3-a5af-002ceb718857-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"f7874ec1-b9cb-4fc3-a5af-002ceb718857\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.647145 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lp6cx\" (UniqueName: \"kubernetes.io/projected/f7874ec1-b9cb-4fc3-a5af-002ceb718857-kube-api-access-lp6cx\") pod \"nova-cell1-novncproxy-0\" (UID: \"f7874ec1-b9cb-4fc3-a5af-002ceb718857\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.732680 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqfrz\" (UniqueName: \"kubernetes.io/projected/92350db6-3f3b-44c6-915c-8223296fe946-kube-api-access-tqfrz\") pod \"nova-scheduler-0\" (UID: \"92350db6-3f3b-44c6-915c-8223296fe946\") " pod="openstack/nova-scheduler-0" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.732726 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9060524b-48eb-42ba-b1e7-5a852d89d94f-config\") pod \"dnsmasq-dns-757b4f8459-kjvns\" (UID: \"9060524b-48eb-42ba-b1e7-5a852d89d94f\") " pod="openstack/dnsmasq-dns-757b4f8459-kjvns" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.732750 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9060524b-48eb-42ba-b1e7-5a852d89d94f-dns-svc\") pod \"dnsmasq-dns-757b4f8459-kjvns\" (UID: \"9060524b-48eb-42ba-b1e7-5a852d89d94f\") " pod="openstack/dnsmasq-dns-757b4f8459-kjvns" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.732773 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49e416d1-e37c-4293-b007-5ec30c040757-config-data\") pod \"nova-api-0\" (UID: \"49e416d1-e37c-4293-b007-5ec30c040757\") " pod="openstack/nova-api-0" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.732826 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/49e416d1-e37c-4293-b007-5ec30c040757-logs\") pod \"nova-api-0\" (UID: \"49e416d1-e37c-4293-b007-5ec30c040757\") " pod="openstack/nova-api-0" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.732851 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92350db6-3f3b-44c6-915c-8223296fe946-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"92350db6-3f3b-44c6-915c-8223296fe946\") " pod="openstack/nova-scheduler-0" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.732881 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9060524b-48eb-42ba-b1e7-5a852d89d94f-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-kjvns\" (UID: \"9060524b-48eb-42ba-b1e7-5a852d89d94f\") " pod="openstack/dnsmasq-dns-757b4f8459-kjvns" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.732937 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9060524b-48eb-42ba-b1e7-5a852d89d94f-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-kjvns\" (UID: \"9060524b-48eb-42ba-b1e7-5a852d89d94f\") " pod="openstack/dnsmasq-dns-757b4f8459-kjvns" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.732979 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9060524b-48eb-42ba-b1e7-5a852d89d94f-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-kjvns\" (UID: \"9060524b-48eb-42ba-b1e7-5a852d89d94f\") " pod="openstack/dnsmasq-dns-757b4f8459-kjvns" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.733026 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82rvh\" (UniqueName: \"kubernetes.io/projected/49e416d1-e37c-4293-b007-5ec30c040757-kube-api-access-82rvh\") pod \"nova-api-0\" (UID: \"49e416d1-e37c-4293-b007-5ec30c040757\") " pod="openstack/nova-api-0" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.733046 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m28qt\" (UniqueName: \"kubernetes.io/projected/9060524b-48eb-42ba-b1e7-5a852d89d94f-kube-api-access-m28qt\") pod \"dnsmasq-dns-757b4f8459-kjvns\" (UID: \"9060524b-48eb-42ba-b1e7-5a852d89d94f\") " pod="openstack/dnsmasq-dns-757b4f8459-kjvns" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.733065 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49e416d1-e37c-4293-b007-5ec30c040757-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"49e416d1-e37c-4293-b007-5ec30c040757\") " pod="openstack/nova-api-0" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.733087 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92350db6-3f3b-44c6-915c-8223296fe946-config-data\") pod \"nova-scheduler-0\" (UID: \"92350db6-3f3b-44c6-915c-8223296fe946\") " pod="openstack/nova-scheduler-0" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.734413 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9060524b-48eb-42ba-b1e7-5a852d89d94f-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-kjvns\" (UID: \"9060524b-48eb-42ba-b1e7-5a852d89d94f\") " pod="openstack/dnsmasq-dns-757b4f8459-kjvns" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.734752 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9060524b-48eb-42ba-b1e7-5a852d89d94f-config\") pod \"dnsmasq-dns-757b4f8459-kjvns\" (UID: \"9060524b-48eb-42ba-b1e7-5a852d89d94f\") " pod="openstack/dnsmasq-dns-757b4f8459-kjvns" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.734904 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9060524b-48eb-42ba-b1e7-5a852d89d94f-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-kjvns\" (UID: \"9060524b-48eb-42ba-b1e7-5a852d89d94f\") " pod="openstack/dnsmasq-dns-757b4f8459-kjvns" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.736198 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9060524b-48eb-42ba-b1e7-5a852d89d94f-dns-svc\") pod \"dnsmasq-dns-757b4f8459-kjvns\" (UID: \"9060524b-48eb-42ba-b1e7-5a852d89d94f\") " pod="openstack/dnsmasq-dns-757b4f8459-kjvns" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.738086 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/49e416d1-e37c-4293-b007-5ec30c040757-logs\") pod \"nova-api-0\" (UID: \"49e416d1-e37c-4293-b007-5ec30c040757\") " pod="openstack/nova-api-0" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.738237 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9060524b-48eb-42ba-b1e7-5a852d89d94f-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-kjvns\" (UID: \"9060524b-48eb-42ba-b1e7-5a852d89d94f\") " pod="openstack/dnsmasq-dns-757b4f8459-kjvns" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.741125 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49e416d1-e37c-4293-b007-5ec30c040757-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"49e416d1-e37c-4293-b007-5ec30c040757\") " pod="openstack/nova-api-0" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.749259 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82rvh\" (UniqueName: \"kubernetes.io/projected/49e416d1-e37c-4293-b007-5ec30c040757-kube-api-access-82rvh\") pod \"nova-api-0\" (UID: \"49e416d1-e37c-4293-b007-5ec30c040757\") " pod="openstack/nova-api-0" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.751978 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49e416d1-e37c-4293-b007-5ec30c040757-config-data\") pod \"nova-api-0\" (UID: \"49e416d1-e37c-4293-b007-5ec30c040757\") " pod="openstack/nova-api-0" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.756231 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m28qt\" (UniqueName: \"kubernetes.io/projected/9060524b-48eb-42ba-b1e7-5a852d89d94f-kube-api-access-m28qt\") pod \"dnsmasq-dns-757b4f8459-kjvns\" (UID: \"9060524b-48eb-42ba-b1e7-5a852d89d94f\") " pod="openstack/dnsmasq-dns-757b4f8459-kjvns" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.834606 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqfrz\" (UniqueName: \"kubernetes.io/projected/92350db6-3f3b-44c6-915c-8223296fe946-kube-api-access-tqfrz\") pod \"nova-scheduler-0\" (UID: \"92350db6-3f3b-44c6-915c-8223296fe946\") " pod="openstack/nova-scheduler-0" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.834710 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92350db6-3f3b-44c6-915c-8223296fe946-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"92350db6-3f3b-44c6-915c-8223296fe946\") " pod="openstack/nova-scheduler-0" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.834839 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92350db6-3f3b-44c6-915c-8223296fe946-config-data\") pod \"nova-scheduler-0\" (UID: \"92350db6-3f3b-44c6-915c-8223296fe946\") " pod="openstack/nova-scheduler-0" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.838326 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92350db6-3f3b-44c6-915c-8223296fe946-config-data\") pod \"nova-scheduler-0\" (UID: \"92350db6-3f3b-44c6-915c-8223296fe946\") " pod="openstack/nova-scheduler-0" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.841655 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92350db6-3f3b-44c6-915c-8223296fe946-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"92350db6-3f3b-44c6-915c-8223296fe946\") " pod="openstack/nova-scheduler-0" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.856768 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqfrz\" (UniqueName: \"kubernetes.io/projected/92350db6-3f3b-44c6-915c-8223296fe946-kube-api-access-tqfrz\") pod \"nova-scheduler-0\" (UID: \"92350db6-3f3b-44c6-915c-8223296fe946\") " pod="openstack/nova-scheduler-0" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.937533 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.955426 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-kjvns" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.977972 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 15:23:16 crc kubenswrapper[4890]: I1125 15:23:16.987296 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 15:23:17 crc kubenswrapper[4890]: I1125 15:23:17.088248 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-828xw"] Nov 25 15:23:17 crc kubenswrapper[4890]: I1125 15:23:17.089947 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-828xw" Nov 25 15:23:17 crc kubenswrapper[4890]: I1125 15:23:17.094769 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Nov 25 15:23:17 crc kubenswrapper[4890]: I1125 15:23:17.095127 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 25 15:23:17 crc kubenswrapper[4890]: I1125 15:23:17.111828 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-828xw"] Nov 25 15:23:17 crc kubenswrapper[4890]: I1125 15:23:17.127438 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-t4rb6"] Nov 25 15:23:17 crc kubenswrapper[4890]: I1125 15:23:17.218837 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 15:23:17 crc kubenswrapper[4890]: W1125 15:23:17.233567 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3d2cf4a9_6672_4006_8d0d_6b380f442fc6.slice/crio-f3dc9ed2521c1a50aff723fb95d650c03a3f5153706b401f231c395f0da61006 WatchSource:0}: Error finding container f3dc9ed2521c1a50aff723fb95d650c03a3f5153706b401f231c395f0da61006: Status 404 returned error can't find the container with id f3dc9ed2521c1a50aff723fb95d650c03a3f5153706b401f231c395f0da61006 Nov 25 15:23:17 crc kubenswrapper[4890]: I1125 15:23:17.246717 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70c6170b-9975-487c-ac2d-a2a309833658-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-828xw\" (UID: \"70c6170b-9975-487c-ac2d-a2a309833658\") " pod="openstack/nova-cell1-conductor-db-sync-828xw" Nov 25 15:23:17 crc kubenswrapper[4890]: I1125 15:23:17.246767 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qcbxm\" (UniqueName: \"kubernetes.io/projected/70c6170b-9975-487c-ac2d-a2a309833658-kube-api-access-qcbxm\") pod \"nova-cell1-conductor-db-sync-828xw\" (UID: \"70c6170b-9975-487c-ac2d-a2a309833658\") " pod="openstack/nova-cell1-conductor-db-sync-828xw" Nov 25 15:23:17 crc kubenswrapper[4890]: I1125 15:23:17.246828 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70c6170b-9975-487c-ac2d-a2a309833658-scripts\") pod \"nova-cell1-conductor-db-sync-828xw\" (UID: \"70c6170b-9975-487c-ac2d-a2a309833658\") " pod="openstack/nova-cell1-conductor-db-sync-828xw" Nov 25 15:23:17 crc kubenswrapper[4890]: I1125 15:23:17.246902 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70c6170b-9975-487c-ac2d-a2a309833658-config-data\") pod \"nova-cell1-conductor-db-sync-828xw\" (UID: \"70c6170b-9975-487c-ac2d-a2a309833658\") " pod="openstack/nova-cell1-conductor-db-sync-828xw" Nov 25 15:23:17 crc kubenswrapper[4890]: I1125 15:23:17.349184 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70c6170b-9975-487c-ac2d-a2a309833658-config-data\") pod \"nova-cell1-conductor-db-sync-828xw\" (UID: \"70c6170b-9975-487c-ac2d-a2a309833658\") " pod="openstack/nova-cell1-conductor-db-sync-828xw" Nov 25 15:23:17 crc kubenswrapper[4890]: I1125 15:23:17.349399 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70c6170b-9975-487c-ac2d-a2a309833658-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-828xw\" (UID: \"70c6170b-9975-487c-ac2d-a2a309833658\") " pod="openstack/nova-cell1-conductor-db-sync-828xw" Nov 25 15:23:17 crc kubenswrapper[4890]: I1125 15:23:17.349427 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qcbxm\" (UniqueName: \"kubernetes.io/projected/70c6170b-9975-487c-ac2d-a2a309833658-kube-api-access-qcbxm\") pod \"nova-cell1-conductor-db-sync-828xw\" (UID: \"70c6170b-9975-487c-ac2d-a2a309833658\") " pod="openstack/nova-cell1-conductor-db-sync-828xw" Nov 25 15:23:17 crc kubenswrapper[4890]: I1125 15:23:17.349477 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70c6170b-9975-487c-ac2d-a2a309833658-scripts\") pod \"nova-cell1-conductor-db-sync-828xw\" (UID: \"70c6170b-9975-487c-ac2d-a2a309833658\") " pod="openstack/nova-cell1-conductor-db-sync-828xw" Nov 25 15:23:17 crc kubenswrapper[4890]: I1125 15:23:17.357684 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70c6170b-9975-487c-ac2d-a2a309833658-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-828xw\" (UID: \"70c6170b-9975-487c-ac2d-a2a309833658\") " pod="openstack/nova-cell1-conductor-db-sync-828xw" Nov 25 15:23:17 crc kubenswrapper[4890]: I1125 15:23:17.357691 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70c6170b-9975-487c-ac2d-a2a309833658-config-data\") pod \"nova-cell1-conductor-db-sync-828xw\" (UID: \"70c6170b-9975-487c-ac2d-a2a309833658\") " pod="openstack/nova-cell1-conductor-db-sync-828xw" Nov 25 15:23:17 crc kubenswrapper[4890]: I1125 15:23:17.364080 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70c6170b-9975-487c-ac2d-a2a309833658-scripts\") pod \"nova-cell1-conductor-db-sync-828xw\" (UID: \"70c6170b-9975-487c-ac2d-a2a309833658\") " pod="openstack/nova-cell1-conductor-db-sync-828xw" Nov 25 15:23:17 crc kubenswrapper[4890]: I1125 15:23:17.375984 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qcbxm\" (UniqueName: \"kubernetes.io/projected/70c6170b-9975-487c-ac2d-a2a309833658-kube-api-access-qcbxm\") pod \"nova-cell1-conductor-db-sync-828xw\" (UID: \"70c6170b-9975-487c-ac2d-a2a309833658\") " pod="openstack/nova-cell1-conductor-db-sync-828xw" Nov 25 15:23:17 crc kubenswrapper[4890]: I1125 15:23:17.415061 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-828xw" Nov 25 15:23:17 crc kubenswrapper[4890]: I1125 15:23:17.562293 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 15:23:17 crc kubenswrapper[4890]: I1125 15:23:17.640747 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 15:23:17 crc kubenswrapper[4890]: I1125 15:23:17.710860 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 15:23:17 crc kubenswrapper[4890]: W1125 15:23:17.719293 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod49e416d1_e37c_4293_b007_5ec30c040757.slice/crio-64306c3cfa1b006d1e9b88a1b7b9b99bc6fdc44247d3a141caa8131c2f65489b WatchSource:0}: Error finding container 64306c3cfa1b006d1e9b88a1b7b9b99bc6fdc44247d3a141caa8131c2f65489b: Status 404 returned error can't find the container with id 64306c3cfa1b006d1e9b88a1b7b9b99bc6fdc44247d3a141caa8131c2f65489b Nov 25 15:23:17 crc kubenswrapper[4890]: I1125 15:23:17.722299 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-kjvns"] Nov 25 15:23:17 crc kubenswrapper[4890]: W1125 15:23:17.971119 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod70c6170b_9975_487c_ac2d_a2a309833658.slice/crio-2edd2f1868d79aada5c6f4f563af558ef2fb3b1efebc4ef8844df88a7a9cef20 WatchSource:0}: Error finding container 2edd2f1868d79aada5c6f4f563af558ef2fb3b1efebc4ef8844df88a7a9cef20: Status 404 returned error can't find the container with id 2edd2f1868d79aada5c6f4f563af558ef2fb3b1efebc4ef8844df88a7a9cef20 Nov 25 15:23:17 crc kubenswrapper[4890]: I1125 15:23:17.974053 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-828xw"] Nov 25 15:23:18 crc kubenswrapper[4890]: I1125 15:23:18.033033 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-828xw" event={"ID":"70c6170b-9975-487c-ac2d-a2a309833658","Type":"ContainerStarted","Data":"2edd2f1868d79aada5c6f4f563af558ef2fb3b1efebc4ef8844df88a7a9cef20"} Nov 25 15:23:18 crc kubenswrapper[4890]: I1125 15:23:18.034618 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3d2cf4a9-6672-4006-8d0d-6b380f442fc6","Type":"ContainerStarted","Data":"f3dc9ed2521c1a50aff723fb95d650c03a3f5153706b401f231c395f0da61006"} Nov 25 15:23:18 crc kubenswrapper[4890]: I1125 15:23:18.037065 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-t4rb6" event={"ID":"872e8100-002a-4ac1-b699-b35111573d7d","Type":"ContainerStarted","Data":"f38ab6ac425c917e9cfcdf44dedf4ba6b91c5339078f0d15f992406d87b07283"} Nov 25 15:23:18 crc kubenswrapper[4890]: I1125 15:23:18.037098 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-t4rb6" event={"ID":"872e8100-002a-4ac1-b699-b35111573d7d","Type":"ContainerStarted","Data":"1ddfcb0480bd4cfa71d8707d0f61ffcd795836bde5146b735f36c1b141387fe0"} Nov 25 15:23:18 crc kubenswrapper[4890]: I1125 15:23:18.040437 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"92350db6-3f3b-44c6-915c-8223296fe946","Type":"ContainerStarted","Data":"4976e86c186e922cf0c220529e46409ef2648632691213e5018ab6db88632282"} Nov 25 15:23:18 crc kubenswrapper[4890]: I1125 15:23:18.042319 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"49e416d1-e37c-4293-b007-5ec30c040757","Type":"ContainerStarted","Data":"64306c3cfa1b006d1e9b88a1b7b9b99bc6fdc44247d3a141caa8131c2f65489b"} Nov 25 15:23:18 crc kubenswrapper[4890]: I1125 15:23:18.047962 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-kjvns" event={"ID":"9060524b-48eb-42ba-b1e7-5a852d89d94f","Type":"ContainerStarted","Data":"dd6d428a2188a9073530062c1a18cd685809e99a554254915e6b1d0f3fbdac6e"} Nov 25 15:23:18 crc kubenswrapper[4890]: I1125 15:23:18.048027 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-kjvns" event={"ID":"9060524b-48eb-42ba-b1e7-5a852d89d94f","Type":"ContainerStarted","Data":"d5f2da4c503149a55d72aab22d036a67883e3df32d7c9dc4e4a2c1a7b60e8f81"} Nov 25 15:23:18 crc kubenswrapper[4890]: I1125 15:23:18.049697 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"f7874ec1-b9cb-4fc3-a5af-002ceb718857","Type":"ContainerStarted","Data":"c1ce1936e66e601c7026e7f2e98633c46322a184ebb215d6739126f049f88146"} Nov 25 15:23:18 crc kubenswrapper[4890]: I1125 15:23:18.062696 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-t4rb6" podStartSLOduration=2.062673306 podStartE2EDuration="2.062673306s" podCreationTimestamp="2025-11-25 15:23:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:23:18.053319691 +0000 UTC m=+1256.495782321" watchObservedRunningTime="2025-11-25 15:23:18.062673306 +0000 UTC m=+1256.505135916" Nov 25 15:23:19 crc kubenswrapper[4890]: I1125 15:23:19.094344 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-828xw" event={"ID":"70c6170b-9975-487c-ac2d-a2a309833658","Type":"ContainerStarted","Data":"b76d7672c9af03f6cfa1ad825433495b765e29311771d9a0c6a518019d549333"} Nov 25 15:23:19 crc kubenswrapper[4890]: I1125 15:23:19.099347 4890 generic.go:334] "Generic (PLEG): container finished" podID="9060524b-48eb-42ba-b1e7-5a852d89d94f" containerID="dd6d428a2188a9073530062c1a18cd685809e99a554254915e6b1d0f3fbdac6e" exitCode=0 Nov 25 15:23:19 crc kubenswrapper[4890]: I1125 15:23:19.100617 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-kjvns" event={"ID":"9060524b-48eb-42ba-b1e7-5a852d89d94f","Type":"ContainerDied","Data":"dd6d428a2188a9073530062c1a18cd685809e99a554254915e6b1d0f3fbdac6e"} Nov 25 15:23:19 crc kubenswrapper[4890]: I1125 15:23:19.100646 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-kjvns" event={"ID":"9060524b-48eb-42ba-b1e7-5a852d89d94f","Type":"ContainerStarted","Data":"9a1c99eefeb5490ba9f1a7c457143c55e8a76e9cf439ecb289dfeec1c9681662"} Nov 25 15:23:19 crc kubenswrapper[4890]: I1125 15:23:19.100659 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-757b4f8459-kjvns" Nov 25 15:23:19 crc kubenswrapper[4890]: I1125 15:23:19.114151 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-828xw" podStartSLOduration=2.11413137 podStartE2EDuration="2.11413137s" podCreationTimestamp="2025-11-25 15:23:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:23:19.111774101 +0000 UTC m=+1257.554236711" watchObservedRunningTime="2025-11-25 15:23:19.11413137 +0000 UTC m=+1257.556593970" Nov 25 15:23:19 crc kubenswrapper[4890]: I1125 15:23:19.124941 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 25 15:23:19 crc kubenswrapper[4890]: I1125 15:23:19.134882 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-757b4f8459-kjvns" podStartSLOduration=3.134862741 podStartE2EDuration="3.134862741s" podCreationTimestamp="2025-11-25 15:23:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:23:19.127143237 +0000 UTC m=+1257.569605847" watchObservedRunningTime="2025-11-25 15:23:19.134862741 +0000 UTC m=+1257.577325351" Nov 25 15:23:19 crc kubenswrapper[4890]: I1125 15:23:19.683766 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 15:23:19 crc kubenswrapper[4890]: I1125 15:23:19.699211 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 15:23:22 crc kubenswrapper[4890]: I1125 15:23:22.149281 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"49e416d1-e37c-4293-b007-5ec30c040757","Type":"ContainerStarted","Data":"884812514b8afe9a3b5eb43645839401f17cbda7213ace64048c1facc0d00a74"} Nov 25 15:23:22 crc kubenswrapper[4890]: I1125 15:23:22.150111 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"49e416d1-e37c-4293-b007-5ec30c040757","Type":"ContainerStarted","Data":"8f13efbe74ff4d27e1662ce4246ec230c23743e053662b64e118dce0f23a53a6"} Nov 25 15:23:22 crc kubenswrapper[4890]: I1125 15:23:22.154781 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"f7874ec1-b9cb-4fc3-a5af-002ceb718857","Type":"ContainerStarted","Data":"86bba2dc4137e2dc190867b6156e8f0a083e36324a8432d17e4da208e31170c7"} Nov 25 15:23:22 crc kubenswrapper[4890]: I1125 15:23:22.154857 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="f7874ec1-b9cb-4fc3-a5af-002ceb718857" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://86bba2dc4137e2dc190867b6156e8f0a083e36324a8432d17e4da208e31170c7" gracePeriod=30 Nov 25 15:23:22 crc kubenswrapper[4890]: I1125 15:23:22.157568 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="3d2cf4a9-6672-4006-8d0d-6b380f442fc6" containerName="nova-metadata-log" containerID="cri-o://cf0197010721c650d8f7fc6c3d03d72fd098df4e1fb0da58970153b76410a9fb" gracePeriod=30 Nov 25 15:23:22 crc kubenswrapper[4890]: I1125 15:23:22.157637 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3d2cf4a9-6672-4006-8d0d-6b380f442fc6","Type":"ContainerStarted","Data":"f58634dc38cd8fda480de3324ee3c11444290ffeca8461028579b08c12265727"} Nov 25 15:23:22 crc kubenswrapper[4890]: I1125 15:23:22.157659 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3d2cf4a9-6672-4006-8d0d-6b380f442fc6","Type":"ContainerStarted","Data":"cf0197010721c650d8f7fc6c3d03d72fd098df4e1fb0da58970153b76410a9fb"} Nov 25 15:23:22 crc kubenswrapper[4890]: I1125 15:23:22.157686 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="3d2cf4a9-6672-4006-8d0d-6b380f442fc6" containerName="nova-metadata-metadata" containerID="cri-o://f58634dc38cd8fda480de3324ee3c11444290ffeca8461028579b08c12265727" gracePeriod=30 Nov 25 15:23:22 crc kubenswrapper[4890]: I1125 15:23:22.160957 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"92350db6-3f3b-44c6-915c-8223296fe946","Type":"ContainerStarted","Data":"15321cc73c2abb27fce208f45be1bbe219056863f1de914f8bb7838f82f05ffb"} Nov 25 15:23:22 crc kubenswrapper[4890]: I1125 15:23:22.173501 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.92741461 podStartE2EDuration="6.173484855s" podCreationTimestamp="2025-11-25 15:23:16 +0000 UTC" firstStartedPulling="2025-11-25 15:23:17.72885926 +0000 UTC m=+1256.171321870" lastFinishedPulling="2025-11-25 15:23:20.974929505 +0000 UTC m=+1259.417392115" observedRunningTime="2025-11-25 15:23:22.171348531 +0000 UTC m=+1260.613811171" watchObservedRunningTime="2025-11-25 15:23:22.173484855 +0000 UTC m=+1260.615947475" Nov 25 15:23:22 crc kubenswrapper[4890]: I1125 15:23:22.202698 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.879342682 podStartE2EDuration="6.202679448s" podCreationTimestamp="2025-11-25 15:23:16 +0000 UTC" firstStartedPulling="2025-11-25 15:23:17.6556066 +0000 UTC m=+1256.098069210" lastFinishedPulling="2025-11-25 15:23:20.978943366 +0000 UTC m=+1259.421405976" observedRunningTime="2025-11-25 15:23:22.196705538 +0000 UTC m=+1260.639168148" watchObservedRunningTime="2025-11-25 15:23:22.202679448 +0000 UTC m=+1260.645142058" Nov 25 15:23:22 crc kubenswrapper[4890]: I1125 15:23:22.218592 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.868728865 podStartE2EDuration="6.218571778s" podCreationTimestamp="2025-11-25 15:23:16 +0000 UTC" firstStartedPulling="2025-11-25 15:23:17.623907613 +0000 UTC m=+1256.066370223" lastFinishedPulling="2025-11-25 15:23:20.973750526 +0000 UTC m=+1259.416213136" observedRunningTime="2025-11-25 15:23:22.217499631 +0000 UTC m=+1260.659962241" watchObservedRunningTime="2025-11-25 15:23:22.218571778 +0000 UTC m=+1260.661034388" Nov 25 15:23:22 crc kubenswrapper[4890]: I1125 15:23:22.241087 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.505702866 podStartE2EDuration="6.241070473s" podCreationTimestamp="2025-11-25 15:23:16 +0000 UTC" firstStartedPulling="2025-11-25 15:23:17.237726132 +0000 UTC m=+1255.680188742" lastFinishedPulling="2025-11-25 15:23:20.973093739 +0000 UTC m=+1259.415556349" observedRunningTime="2025-11-25 15:23:22.234570879 +0000 UTC m=+1260.677033489" watchObservedRunningTime="2025-11-25 15:23:22.241070473 +0000 UTC m=+1260.683533083" Nov 25 15:23:23 crc kubenswrapper[4890]: I1125 15:23:23.171818 4890 generic.go:334] "Generic (PLEG): container finished" podID="3d2cf4a9-6672-4006-8d0d-6b380f442fc6" containerID="cf0197010721c650d8f7fc6c3d03d72fd098df4e1fb0da58970153b76410a9fb" exitCode=143 Nov 25 15:23:23 crc kubenswrapper[4890]: I1125 15:23:23.171905 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3d2cf4a9-6672-4006-8d0d-6b380f442fc6","Type":"ContainerDied","Data":"cf0197010721c650d8f7fc6c3d03d72fd098df4e1fb0da58970153b76410a9fb"} Nov 25 15:23:23 crc kubenswrapper[4890]: I1125 15:23:23.277577 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 15:23:23 crc kubenswrapper[4890]: I1125 15:23:23.277790 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="24000ee2-b285-4f1a-afe1-75e643c1e172" containerName="kube-state-metrics" containerID="cri-o://49320dd334704d9959f25dc7d2d46346328fbd1b58da0ba378c9421bba59f6ee" gracePeriod=30 Nov 25 15:23:24 crc kubenswrapper[4890]: I1125 15:23:24.202221 4890 generic.go:334] "Generic (PLEG): container finished" podID="24000ee2-b285-4f1a-afe1-75e643c1e172" containerID="49320dd334704d9959f25dc7d2d46346328fbd1b58da0ba378c9421bba59f6ee" exitCode=2 Nov 25 15:23:24 crc kubenswrapper[4890]: I1125 15:23:24.203094 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"24000ee2-b285-4f1a-afe1-75e643c1e172","Type":"ContainerDied","Data":"49320dd334704d9959f25dc7d2d46346328fbd1b58da0ba378c9421bba59f6ee"} Nov 25 15:23:24 crc kubenswrapper[4890]: I1125 15:23:24.213112 4890 generic.go:334] "Generic (PLEG): container finished" podID="3d2cf4a9-6672-4006-8d0d-6b380f442fc6" containerID="f58634dc38cd8fda480de3324ee3c11444290ffeca8461028579b08c12265727" exitCode=0 Nov 25 15:23:24 crc kubenswrapper[4890]: I1125 15:23:24.213186 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3d2cf4a9-6672-4006-8d0d-6b380f442fc6","Type":"ContainerDied","Data":"f58634dc38cd8fda480de3324ee3c11444290ffeca8461028579b08c12265727"} Nov 25 15:23:24 crc kubenswrapper[4890]: I1125 15:23:24.394047 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 15:23:24 crc kubenswrapper[4890]: I1125 15:23:24.402709 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 25 15:23:24 crc kubenswrapper[4890]: I1125 15:23:24.526847 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d2cf4a9-6672-4006-8d0d-6b380f442fc6-logs\") pod \"3d2cf4a9-6672-4006-8d0d-6b380f442fc6\" (UID: \"3d2cf4a9-6672-4006-8d0d-6b380f442fc6\") " Nov 25 15:23:24 crc kubenswrapper[4890]: I1125 15:23:24.526957 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bg8ns\" (UniqueName: \"kubernetes.io/projected/3d2cf4a9-6672-4006-8d0d-6b380f442fc6-kube-api-access-bg8ns\") pod \"3d2cf4a9-6672-4006-8d0d-6b380f442fc6\" (UID: \"3d2cf4a9-6672-4006-8d0d-6b380f442fc6\") " Nov 25 15:23:24 crc kubenswrapper[4890]: I1125 15:23:24.527019 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d2cf4a9-6672-4006-8d0d-6b380f442fc6-combined-ca-bundle\") pod \"3d2cf4a9-6672-4006-8d0d-6b380f442fc6\" (UID: \"3d2cf4a9-6672-4006-8d0d-6b380f442fc6\") " Nov 25 15:23:24 crc kubenswrapper[4890]: I1125 15:23:24.527089 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d2cf4a9-6672-4006-8d0d-6b380f442fc6-config-data\") pod \"3d2cf4a9-6672-4006-8d0d-6b380f442fc6\" (UID: \"3d2cf4a9-6672-4006-8d0d-6b380f442fc6\") " Nov 25 15:23:24 crc kubenswrapper[4890]: I1125 15:23:24.527262 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mtcq\" (UniqueName: \"kubernetes.io/projected/24000ee2-b285-4f1a-afe1-75e643c1e172-kube-api-access-6mtcq\") pod \"24000ee2-b285-4f1a-afe1-75e643c1e172\" (UID: \"24000ee2-b285-4f1a-afe1-75e643c1e172\") " Nov 25 15:23:24 crc kubenswrapper[4890]: I1125 15:23:24.527431 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d2cf4a9-6672-4006-8d0d-6b380f442fc6-logs" (OuterVolumeSpecName: "logs") pod "3d2cf4a9-6672-4006-8d0d-6b380f442fc6" (UID: "3d2cf4a9-6672-4006-8d0d-6b380f442fc6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:23:24 crc kubenswrapper[4890]: I1125 15:23:24.527960 4890 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d2cf4a9-6672-4006-8d0d-6b380f442fc6-logs\") on node \"crc\" DevicePath \"\"" Nov 25 15:23:24 crc kubenswrapper[4890]: I1125 15:23:24.540690 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d2cf4a9-6672-4006-8d0d-6b380f442fc6-kube-api-access-bg8ns" (OuterVolumeSpecName: "kube-api-access-bg8ns") pod "3d2cf4a9-6672-4006-8d0d-6b380f442fc6" (UID: "3d2cf4a9-6672-4006-8d0d-6b380f442fc6"). InnerVolumeSpecName "kube-api-access-bg8ns". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:23:24 crc kubenswrapper[4890]: I1125 15:23:24.540771 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24000ee2-b285-4f1a-afe1-75e643c1e172-kube-api-access-6mtcq" (OuterVolumeSpecName: "kube-api-access-6mtcq") pod "24000ee2-b285-4f1a-afe1-75e643c1e172" (UID: "24000ee2-b285-4f1a-afe1-75e643c1e172"). InnerVolumeSpecName "kube-api-access-6mtcq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:23:24 crc kubenswrapper[4890]: I1125 15:23:24.562073 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d2cf4a9-6672-4006-8d0d-6b380f442fc6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3d2cf4a9-6672-4006-8d0d-6b380f442fc6" (UID: "3d2cf4a9-6672-4006-8d0d-6b380f442fc6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:23:24 crc kubenswrapper[4890]: I1125 15:23:24.587588 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d2cf4a9-6672-4006-8d0d-6b380f442fc6-config-data" (OuterVolumeSpecName: "config-data") pod "3d2cf4a9-6672-4006-8d0d-6b380f442fc6" (UID: "3d2cf4a9-6672-4006-8d0d-6b380f442fc6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:23:24 crc kubenswrapper[4890]: I1125 15:23:24.629675 4890 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d2cf4a9-6672-4006-8d0d-6b380f442fc6-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 15:23:24 crc kubenswrapper[4890]: I1125 15:23:24.629706 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mtcq\" (UniqueName: \"kubernetes.io/projected/24000ee2-b285-4f1a-afe1-75e643c1e172-kube-api-access-6mtcq\") on node \"crc\" DevicePath \"\"" Nov 25 15:23:24 crc kubenswrapper[4890]: I1125 15:23:24.629717 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bg8ns\" (UniqueName: \"kubernetes.io/projected/3d2cf4a9-6672-4006-8d0d-6b380f442fc6-kube-api-access-bg8ns\") on node \"crc\" DevicePath \"\"" Nov 25 15:23:24 crc kubenswrapper[4890]: I1125 15:23:24.629727 4890 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d2cf4a9-6672-4006-8d0d-6b380f442fc6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.114624 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.115439 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="936ff525-5705-492b-8262-dc886cd0ccfa" containerName="ceilometer-notification-agent" containerID="cri-o://dcace7be575ca39c462893475e6bbccce00e1f0de451fabe14cb9ef826a5dbd7" gracePeriod=30 Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.115445 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="936ff525-5705-492b-8262-dc886cd0ccfa" containerName="proxy-httpd" containerID="cri-o://663454459ec9d503abbc83dafe5617cd37228e8a850da5b7b16f3ca8dece0567" gracePeriod=30 Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.115439 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="936ff525-5705-492b-8262-dc886cd0ccfa" containerName="sg-core" containerID="cri-o://6c5b96bfb343f6da2c3f5521049b204f2b60420ceb97afee565b4bb5a609af83" gracePeriod=30 Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.115688 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="936ff525-5705-492b-8262-dc886cd0ccfa" containerName="ceilometer-central-agent" containerID="cri-o://e1d0efa602b582a36d30b22f424477e447da02cbf0301407dfc60beceffddfe4" gracePeriod=30 Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.225216 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"24000ee2-b285-4f1a-afe1-75e643c1e172","Type":"ContainerDied","Data":"9a252e68275b5109f09d5c7b05b3bff493c11dbbe5426349770b48dfde18848e"} Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.225254 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.225265 4890 scope.go:117] "RemoveContainer" containerID="49320dd334704d9959f25dc7d2d46346328fbd1b58da0ba378c9421bba59f6ee" Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.238284 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3d2cf4a9-6672-4006-8d0d-6b380f442fc6","Type":"ContainerDied","Data":"f3dc9ed2521c1a50aff723fb95d650c03a3f5153706b401f231c395f0da61006"} Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.238361 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.332119 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.339308 4890 scope.go:117] "RemoveContainer" containerID="f58634dc38cd8fda480de3324ee3c11444290ffeca8461028579b08c12265727" Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.350248 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.370654 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.377332 4890 scope.go:117] "RemoveContainer" containerID="cf0197010721c650d8f7fc6c3d03d72fd098df4e1fb0da58970153b76410a9fb" Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.383305 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.400965 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 15:23:25 crc kubenswrapper[4890]: E1125 15:23:25.405632 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24000ee2-b285-4f1a-afe1-75e643c1e172" containerName="kube-state-metrics" Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.405676 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="24000ee2-b285-4f1a-afe1-75e643c1e172" containerName="kube-state-metrics" Nov 25 15:23:25 crc kubenswrapper[4890]: E1125 15:23:25.405694 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d2cf4a9-6672-4006-8d0d-6b380f442fc6" containerName="nova-metadata-metadata" Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.405702 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d2cf4a9-6672-4006-8d0d-6b380f442fc6" containerName="nova-metadata-metadata" Nov 25 15:23:25 crc kubenswrapper[4890]: E1125 15:23:25.405722 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d2cf4a9-6672-4006-8d0d-6b380f442fc6" containerName="nova-metadata-log" Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.405730 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d2cf4a9-6672-4006-8d0d-6b380f442fc6" containerName="nova-metadata-log" Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.405945 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d2cf4a9-6672-4006-8d0d-6b380f442fc6" containerName="nova-metadata-metadata" Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.405971 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d2cf4a9-6672-4006-8d0d-6b380f442fc6" containerName="nova-metadata-log" Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.405988 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="24000ee2-b285-4f1a-afe1-75e643c1e172" containerName="kube-state-metrics" Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.407659 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.411553 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.411613 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.425223 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.443385 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.449982 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.453439 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.453622 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.463644 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.547830 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/a47203e9-a88a-42f7-86e1-164c6ad532b5-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"a47203e9-a88a-42f7-86e1-164c6ad532b5\") " pod="openstack/kube-state-metrics-0" Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.547881 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwvfq\" (UniqueName: \"kubernetes.io/projected/a47203e9-a88a-42f7-86e1-164c6ad532b5-kube-api-access-mwvfq\") pod \"kube-state-metrics-0\" (UID: \"a47203e9-a88a-42f7-86e1-164c6ad532b5\") " pod="openstack/kube-state-metrics-0" Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.547933 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfc2z\" (UniqueName: \"kubernetes.io/projected/9355e0f0-8d98-49a0-9dd4-720841673727-kube-api-access-nfc2z\") pod \"nova-metadata-0\" (UID: \"9355e0f0-8d98-49a0-9dd4-720841673727\") " pod="openstack/nova-metadata-0" Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.548037 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9355e0f0-8d98-49a0-9dd4-720841673727-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9355e0f0-8d98-49a0-9dd4-720841673727\") " pod="openstack/nova-metadata-0" Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.548123 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9355e0f0-8d98-49a0-9dd4-720841673727-logs\") pod \"nova-metadata-0\" (UID: \"9355e0f0-8d98-49a0-9dd4-720841673727\") " pod="openstack/nova-metadata-0" Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.548174 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a47203e9-a88a-42f7-86e1-164c6ad532b5-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"a47203e9-a88a-42f7-86e1-164c6ad532b5\") " pod="openstack/kube-state-metrics-0" Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.548336 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9355e0f0-8d98-49a0-9dd4-720841673727-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"9355e0f0-8d98-49a0-9dd4-720841673727\") " pod="openstack/nova-metadata-0" Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.548426 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9355e0f0-8d98-49a0-9dd4-720841673727-config-data\") pod \"nova-metadata-0\" (UID: \"9355e0f0-8d98-49a0-9dd4-720841673727\") " pod="openstack/nova-metadata-0" Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.548495 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/a47203e9-a88a-42f7-86e1-164c6ad532b5-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"a47203e9-a88a-42f7-86e1-164c6ad532b5\") " pod="openstack/kube-state-metrics-0" Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.650654 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9355e0f0-8d98-49a0-9dd4-720841673727-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"9355e0f0-8d98-49a0-9dd4-720841673727\") " pod="openstack/nova-metadata-0" Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.650761 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9355e0f0-8d98-49a0-9dd4-720841673727-config-data\") pod \"nova-metadata-0\" (UID: \"9355e0f0-8d98-49a0-9dd4-720841673727\") " pod="openstack/nova-metadata-0" Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.650815 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/a47203e9-a88a-42f7-86e1-164c6ad532b5-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"a47203e9-a88a-42f7-86e1-164c6ad532b5\") " pod="openstack/kube-state-metrics-0" Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.650852 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/a47203e9-a88a-42f7-86e1-164c6ad532b5-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"a47203e9-a88a-42f7-86e1-164c6ad532b5\") " pod="openstack/kube-state-metrics-0" Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.650882 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwvfq\" (UniqueName: \"kubernetes.io/projected/a47203e9-a88a-42f7-86e1-164c6ad532b5-kube-api-access-mwvfq\") pod \"kube-state-metrics-0\" (UID: \"a47203e9-a88a-42f7-86e1-164c6ad532b5\") " pod="openstack/kube-state-metrics-0" Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.650923 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfc2z\" (UniqueName: \"kubernetes.io/projected/9355e0f0-8d98-49a0-9dd4-720841673727-kube-api-access-nfc2z\") pod \"nova-metadata-0\" (UID: \"9355e0f0-8d98-49a0-9dd4-720841673727\") " pod="openstack/nova-metadata-0" Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.650965 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9355e0f0-8d98-49a0-9dd4-720841673727-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9355e0f0-8d98-49a0-9dd4-720841673727\") " pod="openstack/nova-metadata-0" Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.650996 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9355e0f0-8d98-49a0-9dd4-720841673727-logs\") pod \"nova-metadata-0\" (UID: \"9355e0f0-8d98-49a0-9dd4-720841673727\") " pod="openstack/nova-metadata-0" Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.651015 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a47203e9-a88a-42f7-86e1-164c6ad532b5-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"a47203e9-a88a-42f7-86e1-164c6ad532b5\") " pod="openstack/kube-state-metrics-0" Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.654738 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9355e0f0-8d98-49a0-9dd4-720841673727-logs\") pod \"nova-metadata-0\" (UID: \"9355e0f0-8d98-49a0-9dd4-720841673727\") " pod="openstack/nova-metadata-0" Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.655922 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9355e0f0-8d98-49a0-9dd4-720841673727-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"9355e0f0-8d98-49a0-9dd4-720841673727\") " pod="openstack/nova-metadata-0" Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.656961 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/a47203e9-a88a-42f7-86e1-164c6ad532b5-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"a47203e9-a88a-42f7-86e1-164c6ad532b5\") " pod="openstack/kube-state-metrics-0" Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.656976 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9355e0f0-8d98-49a0-9dd4-720841673727-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9355e0f0-8d98-49a0-9dd4-720841673727\") " pod="openstack/nova-metadata-0" Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.658965 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a47203e9-a88a-42f7-86e1-164c6ad532b5-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"a47203e9-a88a-42f7-86e1-164c6ad532b5\") " pod="openstack/kube-state-metrics-0" Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.659291 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9355e0f0-8d98-49a0-9dd4-720841673727-config-data\") pod \"nova-metadata-0\" (UID: \"9355e0f0-8d98-49a0-9dd4-720841673727\") " pod="openstack/nova-metadata-0" Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.659760 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/a47203e9-a88a-42f7-86e1-164c6ad532b5-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"a47203e9-a88a-42f7-86e1-164c6ad532b5\") " pod="openstack/kube-state-metrics-0" Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.672517 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwvfq\" (UniqueName: \"kubernetes.io/projected/a47203e9-a88a-42f7-86e1-164c6ad532b5-kube-api-access-mwvfq\") pod \"kube-state-metrics-0\" (UID: \"a47203e9-a88a-42f7-86e1-164c6ad532b5\") " pod="openstack/kube-state-metrics-0" Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.684226 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfc2z\" (UniqueName: \"kubernetes.io/projected/9355e0f0-8d98-49a0-9dd4-720841673727-kube-api-access-nfc2z\") pod \"nova-metadata-0\" (UID: \"9355e0f0-8d98-49a0-9dd4-720841673727\") " pod="openstack/nova-metadata-0" Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.738615 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 25 15:23:25 crc kubenswrapper[4890]: I1125 15:23:25.777522 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 15:23:26 crc kubenswrapper[4890]: I1125 15:23:26.216266 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24000ee2-b285-4f1a-afe1-75e643c1e172" path="/var/lib/kubelet/pods/24000ee2-b285-4f1a-afe1-75e643c1e172/volumes" Nov 25 15:23:26 crc kubenswrapper[4890]: I1125 15:23:26.218121 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d2cf4a9-6672-4006-8d0d-6b380f442fc6" path="/var/lib/kubelet/pods/3d2cf4a9-6672-4006-8d0d-6b380f442fc6/volumes" Nov 25 15:23:26 crc kubenswrapper[4890]: I1125 15:23:26.277034 4890 generic.go:334] "Generic (PLEG): container finished" podID="936ff525-5705-492b-8262-dc886cd0ccfa" containerID="663454459ec9d503abbc83dafe5617cd37228e8a850da5b7b16f3ca8dece0567" exitCode=0 Nov 25 15:23:26 crc kubenswrapper[4890]: I1125 15:23:26.277067 4890 generic.go:334] "Generic (PLEG): container finished" podID="936ff525-5705-492b-8262-dc886cd0ccfa" containerID="6c5b96bfb343f6da2c3f5521049b204f2b60420ceb97afee565b4bb5a609af83" exitCode=2 Nov 25 15:23:26 crc kubenswrapper[4890]: I1125 15:23:26.277078 4890 generic.go:334] "Generic (PLEG): container finished" podID="936ff525-5705-492b-8262-dc886cd0ccfa" containerID="e1d0efa602b582a36d30b22f424477e447da02cbf0301407dfc60beceffddfe4" exitCode=0 Nov 25 15:23:26 crc kubenswrapper[4890]: I1125 15:23:26.277127 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"936ff525-5705-492b-8262-dc886cd0ccfa","Type":"ContainerDied","Data":"663454459ec9d503abbc83dafe5617cd37228e8a850da5b7b16f3ca8dece0567"} Nov 25 15:23:26 crc kubenswrapper[4890]: I1125 15:23:26.277174 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"936ff525-5705-492b-8262-dc886cd0ccfa","Type":"ContainerDied","Data":"6c5b96bfb343f6da2c3f5521049b204f2b60420ceb97afee565b4bb5a609af83"} Nov 25 15:23:26 crc kubenswrapper[4890]: I1125 15:23:26.277189 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"936ff525-5705-492b-8262-dc886cd0ccfa","Type":"ContainerDied","Data":"e1d0efa602b582a36d30b22f424477e447da02cbf0301407dfc60beceffddfe4"} Nov 25 15:23:26 crc kubenswrapper[4890]: I1125 15:23:26.288077 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 15:23:26 crc kubenswrapper[4890]: I1125 15:23:26.301759 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 15:23:26 crc kubenswrapper[4890]: I1125 15:23:26.938473 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 25 15:23:26 crc kubenswrapper[4890]: I1125 15:23:26.957092 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-757b4f8459-kjvns" Nov 25 15:23:26 crc kubenswrapper[4890]: I1125 15:23:26.978470 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 25 15:23:26 crc kubenswrapper[4890]: I1125 15:23:26.978508 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 25 15:23:26 crc kubenswrapper[4890]: I1125 15:23:26.988328 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 25 15:23:26 crc kubenswrapper[4890]: I1125 15:23:26.988369 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 25 15:23:27 crc kubenswrapper[4890]: I1125 15:23:27.041897 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 25 15:23:27 crc kubenswrapper[4890]: I1125 15:23:27.052239 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-kl5cw"] Nov 25 15:23:27 crc kubenswrapper[4890]: I1125 15:23:27.052755 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c9776ccc5-kl5cw" podUID="690fc634-5a1a-4f30-bbba-264e5f3d0b5d" containerName="dnsmasq-dns" containerID="cri-o://07c84e1d442d2b461dc07c93f2e5691075954cc4f96c0f66bc25c474ca99f41b" gracePeriod=10 Nov 25 15:23:27 crc kubenswrapper[4890]: I1125 15:23:27.294668 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9355e0f0-8d98-49a0-9dd4-720841673727","Type":"ContainerStarted","Data":"e250fa327ea48096afcfbf3d3a485c7eff758a9583e480b94a2b41767fdf1607"} Nov 25 15:23:27 crc kubenswrapper[4890]: I1125 15:23:27.295188 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9355e0f0-8d98-49a0-9dd4-720841673727","Type":"ContainerStarted","Data":"5c87c8a4cf5e8d339eae1d1cff2df87aca8d18a2acc8d4d86741296390b30367"} Nov 25 15:23:27 crc kubenswrapper[4890]: I1125 15:23:27.295207 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9355e0f0-8d98-49a0-9dd4-720841673727","Type":"ContainerStarted","Data":"580ef2b49d1cab70b26dd8079a4bbe7ac7e8a9f2ef238e0a0c2d246806306e09"} Nov 25 15:23:27 crc kubenswrapper[4890]: I1125 15:23:27.296400 4890 generic.go:334] "Generic (PLEG): container finished" podID="690fc634-5a1a-4f30-bbba-264e5f3d0b5d" containerID="07c84e1d442d2b461dc07c93f2e5691075954cc4f96c0f66bc25c474ca99f41b" exitCode=0 Nov 25 15:23:27 crc kubenswrapper[4890]: I1125 15:23:27.296497 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-kl5cw" event={"ID":"690fc634-5a1a-4f30-bbba-264e5f3d0b5d","Type":"ContainerDied","Data":"07c84e1d442d2b461dc07c93f2e5691075954cc4f96c0f66bc25c474ca99f41b"} Nov 25 15:23:27 crc kubenswrapper[4890]: I1125 15:23:27.297964 4890 generic.go:334] "Generic (PLEG): container finished" podID="872e8100-002a-4ac1-b699-b35111573d7d" containerID="f38ab6ac425c917e9cfcdf44dedf4ba6b91c5339078f0d15f992406d87b07283" exitCode=0 Nov 25 15:23:27 crc kubenswrapper[4890]: I1125 15:23:27.298035 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-t4rb6" event={"ID":"872e8100-002a-4ac1-b699-b35111573d7d","Type":"ContainerDied","Data":"f38ab6ac425c917e9cfcdf44dedf4ba6b91c5339078f0d15f992406d87b07283"} Nov 25 15:23:27 crc kubenswrapper[4890]: I1125 15:23:27.300095 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"a47203e9-a88a-42f7-86e1-164c6ad532b5","Type":"ContainerStarted","Data":"f6c757f7b337424920e92dc7103283ff1d00267dabeb93ed84f34ff593609e9a"} Nov 25 15:23:27 crc kubenswrapper[4890]: I1125 15:23:27.300173 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"a47203e9-a88a-42f7-86e1-164c6ad532b5","Type":"ContainerStarted","Data":"9d98de83e860a74f8754adaf53b959b5ceeb3f7536da3969394b8f584cd22636"} Nov 25 15:23:27 crc kubenswrapper[4890]: I1125 15:23:27.339871 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.33984533 podStartE2EDuration="2.33984533s" podCreationTimestamp="2025-11-25 15:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:23:27.328635078 +0000 UTC m=+1265.771097708" watchObservedRunningTime="2025-11-25 15:23:27.33984533 +0000 UTC m=+1265.782307940" Nov 25 15:23:27 crc kubenswrapper[4890]: I1125 15:23:27.351936 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 25 15:23:27 crc kubenswrapper[4890]: I1125 15:23:27.372603 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.93736262 podStartE2EDuration="2.372580512s" podCreationTimestamp="2025-11-25 15:23:25 +0000 UTC" firstStartedPulling="2025-11-25 15:23:26.302085211 +0000 UTC m=+1264.744547821" lastFinishedPulling="2025-11-25 15:23:26.737303103 +0000 UTC m=+1265.179765713" observedRunningTime="2025-11-25 15:23:27.360797436 +0000 UTC m=+1265.803260076" watchObservedRunningTime="2025-11-25 15:23:27.372580512 +0000 UTC m=+1265.815043122" Nov 25 15:23:27 crc kubenswrapper[4890]: I1125 15:23:27.589649 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-kl5cw" Nov 25 15:23:27 crc kubenswrapper[4890]: I1125 15:23:27.712338 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/690fc634-5a1a-4f30-bbba-264e5f3d0b5d-ovsdbserver-sb\") pod \"690fc634-5a1a-4f30-bbba-264e5f3d0b5d\" (UID: \"690fc634-5a1a-4f30-bbba-264e5f3d0b5d\") " Nov 25 15:23:27 crc kubenswrapper[4890]: I1125 15:23:27.712483 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/690fc634-5a1a-4f30-bbba-264e5f3d0b5d-dns-svc\") pod \"690fc634-5a1a-4f30-bbba-264e5f3d0b5d\" (UID: \"690fc634-5a1a-4f30-bbba-264e5f3d0b5d\") " Nov 25 15:23:27 crc kubenswrapper[4890]: I1125 15:23:27.712601 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/690fc634-5a1a-4f30-bbba-264e5f3d0b5d-config\") pod \"690fc634-5a1a-4f30-bbba-264e5f3d0b5d\" (UID: \"690fc634-5a1a-4f30-bbba-264e5f3d0b5d\") " Nov 25 15:23:27 crc kubenswrapper[4890]: I1125 15:23:27.712647 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/690fc634-5a1a-4f30-bbba-264e5f3d0b5d-dns-swift-storage-0\") pod \"690fc634-5a1a-4f30-bbba-264e5f3d0b5d\" (UID: \"690fc634-5a1a-4f30-bbba-264e5f3d0b5d\") " Nov 25 15:23:27 crc kubenswrapper[4890]: I1125 15:23:27.712678 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wr4s5\" (UniqueName: \"kubernetes.io/projected/690fc634-5a1a-4f30-bbba-264e5f3d0b5d-kube-api-access-wr4s5\") pod \"690fc634-5a1a-4f30-bbba-264e5f3d0b5d\" (UID: \"690fc634-5a1a-4f30-bbba-264e5f3d0b5d\") " Nov 25 15:23:27 crc kubenswrapper[4890]: I1125 15:23:27.712721 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/690fc634-5a1a-4f30-bbba-264e5f3d0b5d-ovsdbserver-nb\") pod \"690fc634-5a1a-4f30-bbba-264e5f3d0b5d\" (UID: \"690fc634-5a1a-4f30-bbba-264e5f3d0b5d\") " Nov 25 15:23:27 crc kubenswrapper[4890]: I1125 15:23:27.720645 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/690fc634-5a1a-4f30-bbba-264e5f3d0b5d-kube-api-access-wr4s5" (OuterVolumeSpecName: "kube-api-access-wr4s5") pod "690fc634-5a1a-4f30-bbba-264e5f3d0b5d" (UID: "690fc634-5a1a-4f30-bbba-264e5f3d0b5d"). InnerVolumeSpecName "kube-api-access-wr4s5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:23:27 crc kubenswrapper[4890]: I1125 15:23:27.776672 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/690fc634-5a1a-4f30-bbba-264e5f3d0b5d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "690fc634-5a1a-4f30-bbba-264e5f3d0b5d" (UID: "690fc634-5a1a-4f30-bbba-264e5f3d0b5d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:23:27 crc kubenswrapper[4890]: I1125 15:23:27.791385 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/690fc634-5a1a-4f30-bbba-264e5f3d0b5d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "690fc634-5a1a-4f30-bbba-264e5f3d0b5d" (UID: "690fc634-5a1a-4f30-bbba-264e5f3d0b5d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:23:27 crc kubenswrapper[4890]: I1125 15:23:27.794757 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/690fc634-5a1a-4f30-bbba-264e5f3d0b5d-config" (OuterVolumeSpecName: "config") pod "690fc634-5a1a-4f30-bbba-264e5f3d0b5d" (UID: "690fc634-5a1a-4f30-bbba-264e5f3d0b5d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:23:27 crc kubenswrapper[4890]: I1125 15:23:27.797690 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/690fc634-5a1a-4f30-bbba-264e5f3d0b5d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "690fc634-5a1a-4f30-bbba-264e5f3d0b5d" (UID: "690fc634-5a1a-4f30-bbba-264e5f3d0b5d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:23:27 crc kubenswrapper[4890]: I1125 15:23:27.802877 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/690fc634-5a1a-4f30-bbba-264e5f3d0b5d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "690fc634-5a1a-4f30-bbba-264e5f3d0b5d" (UID: "690fc634-5a1a-4f30-bbba-264e5f3d0b5d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:23:27 crc kubenswrapper[4890]: I1125 15:23:27.815120 4890 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/690fc634-5a1a-4f30-bbba-264e5f3d0b5d-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 15:23:27 crc kubenswrapper[4890]: I1125 15:23:27.815421 4890 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/690fc634-5a1a-4f30-bbba-264e5f3d0b5d-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:23:27 crc kubenswrapper[4890]: I1125 15:23:27.815485 4890 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/690fc634-5a1a-4f30-bbba-264e5f3d0b5d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 25 15:23:27 crc kubenswrapper[4890]: I1125 15:23:27.815552 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wr4s5\" (UniqueName: \"kubernetes.io/projected/690fc634-5a1a-4f30-bbba-264e5f3d0b5d-kube-api-access-wr4s5\") on node \"crc\" DevicePath \"\"" Nov 25 15:23:27 crc kubenswrapper[4890]: I1125 15:23:27.815609 4890 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/690fc634-5a1a-4f30-bbba-264e5f3d0b5d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 15:23:27 crc kubenswrapper[4890]: I1125 15:23:27.815662 4890 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/690fc634-5a1a-4f30-bbba-264e5f3d0b5d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 15:23:28 crc kubenswrapper[4890]: I1125 15:23:28.061392 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="49e416d1-e37c-4293-b007-5ec30c040757" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.190:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 25 15:23:28 crc kubenswrapper[4890]: I1125 15:23:28.061594 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="49e416d1-e37c-4293-b007-5ec30c040757" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.190:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 25 15:23:28 crc kubenswrapper[4890]: I1125 15:23:28.318700 4890 generic.go:334] "Generic (PLEG): container finished" podID="936ff525-5705-492b-8262-dc886cd0ccfa" containerID="dcace7be575ca39c462893475e6bbccce00e1f0de451fabe14cb9ef826a5dbd7" exitCode=0 Nov 25 15:23:28 crc kubenswrapper[4890]: I1125 15:23:28.318763 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"936ff525-5705-492b-8262-dc886cd0ccfa","Type":"ContainerDied","Data":"dcace7be575ca39c462893475e6bbccce00e1f0de451fabe14cb9ef826a5dbd7"} Nov 25 15:23:28 crc kubenswrapper[4890]: I1125 15:23:28.320704 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-kl5cw" Nov 25 15:23:28 crc kubenswrapper[4890]: I1125 15:23:28.321543 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-kl5cw" event={"ID":"690fc634-5a1a-4f30-bbba-264e5f3d0b5d","Type":"ContainerDied","Data":"c84b3e7c06e146df097fa57c912f654d928baca2b5df02121f2da60796087d2c"} Nov 25 15:23:28 crc kubenswrapper[4890]: I1125 15:23:28.321572 4890 scope.go:117] "RemoveContainer" containerID="07c84e1d442d2b461dc07c93f2e5691075954cc4f96c0f66bc25c474ca99f41b" Nov 25 15:23:28 crc kubenswrapper[4890]: I1125 15:23:28.322665 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 25 15:23:28 crc kubenswrapper[4890]: I1125 15:23:28.358371 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-kl5cw"] Nov 25 15:23:28 crc kubenswrapper[4890]: I1125 15:23:28.366393 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-kl5cw"] Nov 25 15:23:28 crc kubenswrapper[4890]: I1125 15:23:28.392713 4890 scope.go:117] "RemoveContainer" containerID="00c78bdfe24e45137486eb82fefad0c72a5dacebc512a5e06667f61811b2ff97" Nov 25 15:23:28 crc kubenswrapper[4890]: I1125 15:23:28.454671 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 15:23:28 crc kubenswrapper[4890]: I1125 15:23:28.636490 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/936ff525-5705-492b-8262-dc886cd0ccfa-run-httpd\") pod \"936ff525-5705-492b-8262-dc886cd0ccfa\" (UID: \"936ff525-5705-492b-8262-dc886cd0ccfa\") " Nov 25 15:23:28 crc kubenswrapper[4890]: I1125 15:23:28.636593 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/936ff525-5705-492b-8262-dc886cd0ccfa-log-httpd\") pod \"936ff525-5705-492b-8262-dc886cd0ccfa\" (UID: \"936ff525-5705-492b-8262-dc886cd0ccfa\") " Nov 25 15:23:28 crc kubenswrapper[4890]: I1125 15:23:28.636629 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/936ff525-5705-492b-8262-dc886cd0ccfa-scripts\") pod \"936ff525-5705-492b-8262-dc886cd0ccfa\" (UID: \"936ff525-5705-492b-8262-dc886cd0ccfa\") " Nov 25 15:23:28 crc kubenswrapper[4890]: I1125 15:23:28.636746 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/936ff525-5705-492b-8262-dc886cd0ccfa-sg-core-conf-yaml\") pod \"936ff525-5705-492b-8262-dc886cd0ccfa\" (UID: \"936ff525-5705-492b-8262-dc886cd0ccfa\") " Nov 25 15:23:28 crc kubenswrapper[4890]: I1125 15:23:28.636777 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lptvc\" (UniqueName: \"kubernetes.io/projected/936ff525-5705-492b-8262-dc886cd0ccfa-kube-api-access-lptvc\") pod \"936ff525-5705-492b-8262-dc886cd0ccfa\" (UID: \"936ff525-5705-492b-8262-dc886cd0ccfa\") " Nov 25 15:23:28 crc kubenswrapper[4890]: I1125 15:23:28.636864 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/936ff525-5705-492b-8262-dc886cd0ccfa-config-data\") pod \"936ff525-5705-492b-8262-dc886cd0ccfa\" (UID: \"936ff525-5705-492b-8262-dc886cd0ccfa\") " Nov 25 15:23:28 crc kubenswrapper[4890]: I1125 15:23:28.636927 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/936ff525-5705-492b-8262-dc886cd0ccfa-combined-ca-bundle\") pod \"936ff525-5705-492b-8262-dc886cd0ccfa\" (UID: \"936ff525-5705-492b-8262-dc886cd0ccfa\") " Nov 25 15:23:28 crc kubenswrapper[4890]: I1125 15:23:28.639353 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/936ff525-5705-492b-8262-dc886cd0ccfa-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "936ff525-5705-492b-8262-dc886cd0ccfa" (UID: "936ff525-5705-492b-8262-dc886cd0ccfa"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:23:28 crc kubenswrapper[4890]: I1125 15:23:28.641137 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/936ff525-5705-492b-8262-dc886cd0ccfa-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "936ff525-5705-492b-8262-dc886cd0ccfa" (UID: "936ff525-5705-492b-8262-dc886cd0ccfa"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:23:28 crc kubenswrapper[4890]: I1125 15:23:28.643090 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/936ff525-5705-492b-8262-dc886cd0ccfa-scripts" (OuterVolumeSpecName: "scripts") pod "936ff525-5705-492b-8262-dc886cd0ccfa" (UID: "936ff525-5705-492b-8262-dc886cd0ccfa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:23:28 crc kubenswrapper[4890]: I1125 15:23:28.644856 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/936ff525-5705-492b-8262-dc886cd0ccfa-kube-api-access-lptvc" (OuterVolumeSpecName: "kube-api-access-lptvc") pod "936ff525-5705-492b-8262-dc886cd0ccfa" (UID: "936ff525-5705-492b-8262-dc886cd0ccfa"). InnerVolumeSpecName "kube-api-access-lptvc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:23:28 crc kubenswrapper[4890]: I1125 15:23:28.676302 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/936ff525-5705-492b-8262-dc886cd0ccfa-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "936ff525-5705-492b-8262-dc886cd0ccfa" (UID: "936ff525-5705-492b-8262-dc886cd0ccfa"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:23:28 crc kubenswrapper[4890]: I1125 15:23:28.726010 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/936ff525-5705-492b-8262-dc886cd0ccfa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "936ff525-5705-492b-8262-dc886cd0ccfa" (UID: "936ff525-5705-492b-8262-dc886cd0ccfa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:23:28 crc kubenswrapper[4890]: I1125 15:23:28.739100 4890 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/936ff525-5705-492b-8262-dc886cd0ccfa-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 15:23:28 crc kubenswrapper[4890]: I1125 15:23:28.739144 4890 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/936ff525-5705-492b-8262-dc886cd0ccfa-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 15:23:28 crc kubenswrapper[4890]: I1125 15:23:28.739173 4890 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/936ff525-5705-492b-8262-dc886cd0ccfa-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 15:23:28 crc kubenswrapper[4890]: I1125 15:23:28.739185 4890 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/936ff525-5705-492b-8262-dc886cd0ccfa-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 25 15:23:28 crc kubenswrapper[4890]: I1125 15:23:28.739199 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lptvc\" (UniqueName: \"kubernetes.io/projected/936ff525-5705-492b-8262-dc886cd0ccfa-kube-api-access-lptvc\") on node \"crc\" DevicePath \"\"" Nov 25 15:23:28 crc kubenswrapper[4890]: I1125 15:23:28.739210 4890 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/936ff525-5705-492b-8262-dc886cd0ccfa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:23:28 crc kubenswrapper[4890]: I1125 15:23:28.762135 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/936ff525-5705-492b-8262-dc886cd0ccfa-config-data" (OuterVolumeSpecName: "config-data") pod "936ff525-5705-492b-8262-dc886cd0ccfa" (UID: "936ff525-5705-492b-8262-dc886cd0ccfa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:23:28 crc kubenswrapper[4890]: I1125 15:23:28.833065 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-t4rb6" Nov 25 15:23:28 crc kubenswrapper[4890]: I1125 15:23:28.842149 4890 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/936ff525-5705-492b-8262-dc886cd0ccfa-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 15:23:28 crc kubenswrapper[4890]: I1125 15:23:28.943147 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/872e8100-002a-4ac1-b699-b35111573d7d-config-data\") pod \"872e8100-002a-4ac1-b699-b35111573d7d\" (UID: \"872e8100-002a-4ac1-b699-b35111573d7d\") " Nov 25 15:23:28 crc kubenswrapper[4890]: I1125 15:23:28.943260 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/872e8100-002a-4ac1-b699-b35111573d7d-combined-ca-bundle\") pod \"872e8100-002a-4ac1-b699-b35111573d7d\" (UID: \"872e8100-002a-4ac1-b699-b35111573d7d\") " Nov 25 15:23:28 crc kubenswrapper[4890]: I1125 15:23:28.943297 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/872e8100-002a-4ac1-b699-b35111573d7d-scripts\") pod \"872e8100-002a-4ac1-b699-b35111573d7d\" (UID: \"872e8100-002a-4ac1-b699-b35111573d7d\") " Nov 25 15:23:28 crc kubenswrapper[4890]: I1125 15:23:28.943397 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9hq2s\" (UniqueName: \"kubernetes.io/projected/872e8100-002a-4ac1-b699-b35111573d7d-kube-api-access-9hq2s\") pod \"872e8100-002a-4ac1-b699-b35111573d7d\" (UID: \"872e8100-002a-4ac1-b699-b35111573d7d\") " Nov 25 15:23:28 crc kubenswrapper[4890]: I1125 15:23:28.950080 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/872e8100-002a-4ac1-b699-b35111573d7d-scripts" (OuterVolumeSpecName: "scripts") pod "872e8100-002a-4ac1-b699-b35111573d7d" (UID: "872e8100-002a-4ac1-b699-b35111573d7d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:23:28 crc kubenswrapper[4890]: I1125 15:23:28.950128 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/872e8100-002a-4ac1-b699-b35111573d7d-kube-api-access-9hq2s" (OuterVolumeSpecName: "kube-api-access-9hq2s") pod "872e8100-002a-4ac1-b699-b35111573d7d" (UID: "872e8100-002a-4ac1-b699-b35111573d7d"). InnerVolumeSpecName "kube-api-access-9hq2s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:23:28 crc kubenswrapper[4890]: I1125 15:23:28.972236 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/872e8100-002a-4ac1-b699-b35111573d7d-config-data" (OuterVolumeSpecName: "config-data") pod "872e8100-002a-4ac1-b699-b35111573d7d" (UID: "872e8100-002a-4ac1-b699-b35111573d7d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:23:28 crc kubenswrapper[4890]: I1125 15:23:28.979705 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/872e8100-002a-4ac1-b699-b35111573d7d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "872e8100-002a-4ac1-b699-b35111573d7d" (UID: "872e8100-002a-4ac1-b699-b35111573d7d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.046340 4890 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/872e8100-002a-4ac1-b699-b35111573d7d-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.046371 4890 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/872e8100-002a-4ac1-b699-b35111573d7d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.046382 4890 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/872e8100-002a-4ac1-b699-b35111573d7d-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.046390 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9hq2s\" (UniqueName: \"kubernetes.io/projected/872e8100-002a-4ac1-b699-b35111573d7d-kube-api-access-9hq2s\") on node \"crc\" DevicePath \"\"" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.341226 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-t4rb6" event={"ID":"872e8100-002a-4ac1-b699-b35111573d7d","Type":"ContainerDied","Data":"1ddfcb0480bd4cfa71d8707d0f61ffcd795836bde5146b735f36c1b141387fe0"} Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.341271 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ddfcb0480bd4cfa71d8707d0f61ffcd795836bde5146b735f36c1b141387fe0" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.341334 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-t4rb6" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.353218 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.353212 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"936ff525-5705-492b-8262-dc886cd0ccfa","Type":"ContainerDied","Data":"9c4bd9e2d7855c2478ee5b38022cafc88c168a8cb03c2308b367d31d0d5bec96"} Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.353421 4890 scope.go:117] "RemoveContainer" containerID="663454459ec9d503abbc83dafe5617cd37228e8a850da5b7b16f3ca8dece0567" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.394287 4890 scope.go:117] "RemoveContainer" containerID="6c5b96bfb343f6da2c3f5521049b204f2b60420ceb97afee565b4bb5a609af83" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.400786 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.408213 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.424005 4890 scope.go:117] "RemoveContainer" containerID="dcace7be575ca39c462893475e6bbccce00e1f0de451fabe14cb9ef826a5dbd7" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.426671 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:23:29 crc kubenswrapper[4890]: E1125 15:23:29.427108 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="690fc634-5a1a-4f30-bbba-264e5f3d0b5d" containerName="init" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.427211 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="690fc634-5a1a-4f30-bbba-264e5f3d0b5d" containerName="init" Nov 25 15:23:29 crc kubenswrapper[4890]: E1125 15:23:29.427289 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="936ff525-5705-492b-8262-dc886cd0ccfa" containerName="proxy-httpd" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.427343 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="936ff525-5705-492b-8262-dc886cd0ccfa" containerName="proxy-httpd" Nov 25 15:23:29 crc kubenswrapper[4890]: E1125 15:23:29.427403 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="936ff525-5705-492b-8262-dc886cd0ccfa" containerName="sg-core" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.427608 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="936ff525-5705-492b-8262-dc886cd0ccfa" containerName="sg-core" Nov 25 15:23:29 crc kubenswrapper[4890]: E1125 15:23:29.427672 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="936ff525-5705-492b-8262-dc886cd0ccfa" containerName="ceilometer-notification-agent" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.427728 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="936ff525-5705-492b-8262-dc886cd0ccfa" containerName="ceilometer-notification-agent" Nov 25 15:23:29 crc kubenswrapper[4890]: E1125 15:23:29.427791 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="872e8100-002a-4ac1-b699-b35111573d7d" containerName="nova-manage" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.427842 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="872e8100-002a-4ac1-b699-b35111573d7d" containerName="nova-manage" Nov 25 15:23:29 crc kubenswrapper[4890]: E1125 15:23:29.427901 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="690fc634-5a1a-4f30-bbba-264e5f3d0b5d" containerName="dnsmasq-dns" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.427951 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="690fc634-5a1a-4f30-bbba-264e5f3d0b5d" containerName="dnsmasq-dns" Nov 25 15:23:29 crc kubenswrapper[4890]: E1125 15:23:29.428013 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="936ff525-5705-492b-8262-dc886cd0ccfa" containerName="ceilometer-central-agent" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.428087 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="936ff525-5705-492b-8262-dc886cd0ccfa" containerName="ceilometer-central-agent" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.428346 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="936ff525-5705-492b-8262-dc886cd0ccfa" containerName="proxy-httpd" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.428410 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="690fc634-5a1a-4f30-bbba-264e5f3d0b5d" containerName="dnsmasq-dns" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.428467 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="936ff525-5705-492b-8262-dc886cd0ccfa" containerName="ceilometer-central-agent" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.428525 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="936ff525-5705-492b-8262-dc886cd0ccfa" containerName="ceilometer-notification-agent" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.428591 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="872e8100-002a-4ac1-b699-b35111573d7d" containerName="nova-manage" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.428661 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="936ff525-5705-492b-8262-dc886cd0ccfa" containerName="sg-core" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.447097 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.454432 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.455080 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.455529 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.482008 4890 scope.go:117] "RemoveContainer" containerID="e1d0efa602b582a36d30b22f424477e447da02cbf0301407dfc60beceffddfe4" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.482524 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.549312 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.549644 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="49e416d1-e37c-4293-b007-5ec30c040757" containerName="nova-api-log" containerID="cri-o://8f13efbe74ff4d27e1662ce4246ec230c23743e053662b64e118dce0f23a53a6" gracePeriod=30 Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.550190 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="49e416d1-e37c-4293-b007-5ec30c040757" containerName="nova-api-api" containerID="cri-o://884812514b8afe9a3b5eb43645839401f17cbda7213ace64048c1facc0d00a74" gracePeriod=30 Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.558819 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/154b7e43-8329-4c7d-9870-ffca704e80af-log-httpd\") pod \"ceilometer-0\" (UID: \"154b7e43-8329-4c7d-9870-ffca704e80af\") " pod="openstack/ceilometer-0" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.558873 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/154b7e43-8329-4c7d-9870-ffca704e80af-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"154b7e43-8329-4c7d-9870-ffca704e80af\") " pod="openstack/ceilometer-0" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.558897 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/154b7e43-8329-4c7d-9870-ffca704e80af-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"154b7e43-8329-4c7d-9870-ffca704e80af\") " pod="openstack/ceilometer-0" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.559009 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/154b7e43-8329-4c7d-9870-ffca704e80af-run-httpd\") pod \"ceilometer-0\" (UID: \"154b7e43-8329-4c7d-9870-ffca704e80af\") " pod="openstack/ceilometer-0" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.559071 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/154b7e43-8329-4c7d-9870-ffca704e80af-scripts\") pod \"ceilometer-0\" (UID: \"154b7e43-8329-4c7d-9870-ffca704e80af\") " pod="openstack/ceilometer-0" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.559109 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4zxj\" (UniqueName: \"kubernetes.io/projected/154b7e43-8329-4c7d-9870-ffca704e80af-kube-api-access-t4zxj\") pod \"ceilometer-0\" (UID: \"154b7e43-8329-4c7d-9870-ffca704e80af\") " pod="openstack/ceilometer-0" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.559141 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/154b7e43-8329-4c7d-9870-ffca704e80af-config-data\") pod \"ceilometer-0\" (UID: \"154b7e43-8329-4c7d-9870-ffca704e80af\") " pod="openstack/ceilometer-0" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.559268 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/154b7e43-8329-4c7d-9870-ffca704e80af-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"154b7e43-8329-4c7d-9870-ffca704e80af\") " pod="openstack/ceilometer-0" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.563069 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.563345 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="92350db6-3f3b-44c6-915c-8223296fe946" containerName="nova-scheduler-scheduler" containerID="cri-o://15321cc73c2abb27fce208f45be1bbe219056863f1de914f8bb7838f82f05ffb" gracePeriod=30 Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.586463 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.587313 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="9355e0f0-8d98-49a0-9dd4-720841673727" containerName="nova-metadata-metadata" containerID="cri-o://e250fa327ea48096afcfbf3d3a485c7eff758a9583e480b94a2b41767fdf1607" gracePeriod=30 Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.586922 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="9355e0f0-8d98-49a0-9dd4-720841673727" containerName="nova-metadata-log" containerID="cri-o://5c87c8a4cf5e8d339eae1d1cff2df87aca8d18a2acc8d4d86741296390b30367" gracePeriod=30 Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.661022 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/154b7e43-8329-4c7d-9870-ffca704e80af-run-httpd\") pod \"ceilometer-0\" (UID: \"154b7e43-8329-4c7d-9870-ffca704e80af\") " pod="openstack/ceilometer-0" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.661093 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/154b7e43-8329-4c7d-9870-ffca704e80af-scripts\") pod \"ceilometer-0\" (UID: \"154b7e43-8329-4c7d-9870-ffca704e80af\") " pod="openstack/ceilometer-0" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.661125 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4zxj\" (UniqueName: \"kubernetes.io/projected/154b7e43-8329-4c7d-9870-ffca704e80af-kube-api-access-t4zxj\") pod \"ceilometer-0\" (UID: \"154b7e43-8329-4c7d-9870-ffca704e80af\") " pod="openstack/ceilometer-0" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.661150 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/154b7e43-8329-4c7d-9870-ffca704e80af-config-data\") pod \"ceilometer-0\" (UID: \"154b7e43-8329-4c7d-9870-ffca704e80af\") " pod="openstack/ceilometer-0" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.661204 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/154b7e43-8329-4c7d-9870-ffca704e80af-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"154b7e43-8329-4c7d-9870-ffca704e80af\") " pod="openstack/ceilometer-0" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.661267 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/154b7e43-8329-4c7d-9870-ffca704e80af-log-httpd\") pod \"ceilometer-0\" (UID: \"154b7e43-8329-4c7d-9870-ffca704e80af\") " pod="openstack/ceilometer-0" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.661295 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/154b7e43-8329-4c7d-9870-ffca704e80af-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"154b7e43-8329-4c7d-9870-ffca704e80af\") " pod="openstack/ceilometer-0" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.661313 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/154b7e43-8329-4c7d-9870-ffca704e80af-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"154b7e43-8329-4c7d-9870-ffca704e80af\") " pod="openstack/ceilometer-0" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.662027 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/154b7e43-8329-4c7d-9870-ffca704e80af-run-httpd\") pod \"ceilometer-0\" (UID: \"154b7e43-8329-4c7d-9870-ffca704e80af\") " pod="openstack/ceilometer-0" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.662407 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/154b7e43-8329-4c7d-9870-ffca704e80af-log-httpd\") pod \"ceilometer-0\" (UID: \"154b7e43-8329-4c7d-9870-ffca704e80af\") " pod="openstack/ceilometer-0" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.668021 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/154b7e43-8329-4c7d-9870-ffca704e80af-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"154b7e43-8329-4c7d-9870-ffca704e80af\") " pod="openstack/ceilometer-0" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.668932 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/154b7e43-8329-4c7d-9870-ffca704e80af-scripts\") pod \"ceilometer-0\" (UID: \"154b7e43-8329-4c7d-9870-ffca704e80af\") " pod="openstack/ceilometer-0" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.668949 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/154b7e43-8329-4c7d-9870-ffca704e80af-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"154b7e43-8329-4c7d-9870-ffca704e80af\") " pod="openstack/ceilometer-0" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.672296 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/154b7e43-8329-4c7d-9870-ffca704e80af-config-data\") pod \"ceilometer-0\" (UID: \"154b7e43-8329-4c7d-9870-ffca704e80af\") " pod="openstack/ceilometer-0" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.676990 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/154b7e43-8329-4c7d-9870-ffca704e80af-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"154b7e43-8329-4c7d-9870-ffca704e80af\") " pod="openstack/ceilometer-0" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.680505 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4zxj\" (UniqueName: \"kubernetes.io/projected/154b7e43-8329-4c7d-9870-ffca704e80af-kube-api-access-t4zxj\") pod \"ceilometer-0\" (UID: \"154b7e43-8329-4c7d-9870-ffca704e80af\") " pod="openstack/ceilometer-0" Nov 25 15:23:29 crc kubenswrapper[4890]: I1125 15:23:29.780934 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 15:23:30 crc kubenswrapper[4890]: I1125 15:23:30.182472 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="690fc634-5a1a-4f30-bbba-264e5f3d0b5d" path="/var/lib/kubelet/pods/690fc634-5a1a-4f30-bbba-264e5f3d0b5d/volumes" Nov 25 15:23:30 crc kubenswrapper[4890]: I1125 15:23:30.183800 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="936ff525-5705-492b-8262-dc886cd0ccfa" path="/var/lib/kubelet/pods/936ff525-5705-492b-8262-dc886cd0ccfa/volumes" Nov 25 15:23:30 crc kubenswrapper[4890]: I1125 15:23:30.255294 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:23:30 crc kubenswrapper[4890]: I1125 15:23:30.367174 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"154b7e43-8329-4c7d-9870-ffca704e80af","Type":"ContainerStarted","Data":"2578edede8e4da833402c6ce7f04a159508d2c227f17deb95588e255413b726e"} Nov 25 15:23:30 crc kubenswrapper[4890]: I1125 15:23:30.370308 4890 generic.go:334] "Generic (PLEG): container finished" podID="49e416d1-e37c-4293-b007-5ec30c040757" containerID="8f13efbe74ff4d27e1662ce4246ec230c23743e053662b64e118dce0f23a53a6" exitCode=143 Nov 25 15:23:30 crc kubenswrapper[4890]: I1125 15:23:30.370380 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"49e416d1-e37c-4293-b007-5ec30c040757","Type":"ContainerDied","Data":"8f13efbe74ff4d27e1662ce4246ec230c23743e053662b64e118dce0f23a53a6"} Nov 25 15:23:30 crc kubenswrapper[4890]: I1125 15:23:30.372264 4890 generic.go:334] "Generic (PLEG): container finished" podID="9355e0f0-8d98-49a0-9dd4-720841673727" containerID="e250fa327ea48096afcfbf3d3a485c7eff758a9583e480b94a2b41767fdf1607" exitCode=0 Nov 25 15:23:30 crc kubenswrapper[4890]: I1125 15:23:30.372294 4890 generic.go:334] "Generic (PLEG): container finished" podID="9355e0f0-8d98-49a0-9dd4-720841673727" containerID="5c87c8a4cf5e8d339eae1d1cff2df87aca8d18a2acc8d4d86741296390b30367" exitCode=143 Nov 25 15:23:30 crc kubenswrapper[4890]: I1125 15:23:30.372328 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9355e0f0-8d98-49a0-9dd4-720841673727","Type":"ContainerDied","Data":"e250fa327ea48096afcfbf3d3a485c7eff758a9583e480b94a2b41767fdf1607"} Nov 25 15:23:30 crc kubenswrapper[4890]: I1125 15:23:30.372380 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9355e0f0-8d98-49a0-9dd4-720841673727","Type":"ContainerDied","Data":"5c87c8a4cf5e8d339eae1d1cff2df87aca8d18a2acc8d4d86741296390b30367"} Nov 25 15:23:30 crc kubenswrapper[4890]: I1125 15:23:30.784405 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 15:23:30 crc kubenswrapper[4890]: I1125 15:23:30.986301 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9355e0f0-8d98-49a0-9dd4-720841673727-nova-metadata-tls-certs\") pod \"9355e0f0-8d98-49a0-9dd4-720841673727\" (UID: \"9355e0f0-8d98-49a0-9dd4-720841673727\") " Nov 25 15:23:30 crc kubenswrapper[4890]: I1125 15:23:30.986454 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9355e0f0-8d98-49a0-9dd4-720841673727-combined-ca-bundle\") pod \"9355e0f0-8d98-49a0-9dd4-720841673727\" (UID: \"9355e0f0-8d98-49a0-9dd4-720841673727\") " Nov 25 15:23:30 crc kubenswrapper[4890]: I1125 15:23:30.986638 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nfc2z\" (UniqueName: \"kubernetes.io/projected/9355e0f0-8d98-49a0-9dd4-720841673727-kube-api-access-nfc2z\") pod \"9355e0f0-8d98-49a0-9dd4-720841673727\" (UID: \"9355e0f0-8d98-49a0-9dd4-720841673727\") " Nov 25 15:23:30 crc kubenswrapper[4890]: I1125 15:23:30.986680 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9355e0f0-8d98-49a0-9dd4-720841673727-config-data\") pod \"9355e0f0-8d98-49a0-9dd4-720841673727\" (UID: \"9355e0f0-8d98-49a0-9dd4-720841673727\") " Nov 25 15:23:30 crc kubenswrapper[4890]: I1125 15:23:30.986711 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9355e0f0-8d98-49a0-9dd4-720841673727-logs\") pod \"9355e0f0-8d98-49a0-9dd4-720841673727\" (UID: \"9355e0f0-8d98-49a0-9dd4-720841673727\") " Nov 25 15:23:30 crc kubenswrapper[4890]: I1125 15:23:30.987536 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9355e0f0-8d98-49a0-9dd4-720841673727-logs" (OuterVolumeSpecName: "logs") pod "9355e0f0-8d98-49a0-9dd4-720841673727" (UID: "9355e0f0-8d98-49a0-9dd4-720841673727"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:23:30 crc kubenswrapper[4890]: I1125 15:23:30.995482 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9355e0f0-8d98-49a0-9dd4-720841673727-kube-api-access-nfc2z" (OuterVolumeSpecName: "kube-api-access-nfc2z") pod "9355e0f0-8d98-49a0-9dd4-720841673727" (UID: "9355e0f0-8d98-49a0-9dd4-720841673727"). InnerVolumeSpecName "kube-api-access-nfc2z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:23:31 crc kubenswrapper[4890]: I1125 15:23:31.014868 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9355e0f0-8d98-49a0-9dd4-720841673727-config-data" (OuterVolumeSpecName: "config-data") pod "9355e0f0-8d98-49a0-9dd4-720841673727" (UID: "9355e0f0-8d98-49a0-9dd4-720841673727"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:23:31 crc kubenswrapper[4890]: I1125 15:23:31.019291 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9355e0f0-8d98-49a0-9dd4-720841673727-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9355e0f0-8d98-49a0-9dd4-720841673727" (UID: "9355e0f0-8d98-49a0-9dd4-720841673727"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:23:31 crc kubenswrapper[4890]: I1125 15:23:31.043477 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9355e0f0-8d98-49a0-9dd4-720841673727-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "9355e0f0-8d98-49a0-9dd4-720841673727" (UID: "9355e0f0-8d98-49a0-9dd4-720841673727"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:23:31 crc kubenswrapper[4890]: I1125 15:23:31.088567 4890 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9355e0f0-8d98-49a0-9dd4-720841673727-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 15:23:31 crc kubenswrapper[4890]: I1125 15:23:31.088611 4890 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9355e0f0-8d98-49a0-9dd4-720841673727-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:23:31 crc kubenswrapper[4890]: I1125 15:23:31.088623 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nfc2z\" (UniqueName: \"kubernetes.io/projected/9355e0f0-8d98-49a0-9dd4-720841673727-kube-api-access-nfc2z\") on node \"crc\" DevicePath \"\"" Nov 25 15:23:31 crc kubenswrapper[4890]: I1125 15:23:31.088636 4890 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9355e0f0-8d98-49a0-9dd4-720841673727-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 15:23:31 crc kubenswrapper[4890]: I1125 15:23:31.088648 4890 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9355e0f0-8d98-49a0-9dd4-720841673727-logs\") on node \"crc\" DevicePath \"\"" Nov 25 15:23:31 crc kubenswrapper[4890]: I1125 15:23:31.383079 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 15:23:31 crc kubenswrapper[4890]: I1125 15:23:31.383077 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9355e0f0-8d98-49a0-9dd4-720841673727","Type":"ContainerDied","Data":"580ef2b49d1cab70b26dd8079a4bbe7ac7e8a9f2ef238e0a0c2d246806306e09"} Nov 25 15:23:31 crc kubenswrapper[4890]: I1125 15:23:31.383150 4890 scope.go:117] "RemoveContainer" containerID="e250fa327ea48096afcfbf3d3a485c7eff758a9583e480b94a2b41767fdf1607" Nov 25 15:23:31 crc kubenswrapper[4890]: I1125 15:23:31.390355 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"154b7e43-8329-4c7d-9870-ffca704e80af","Type":"ContainerStarted","Data":"3cf7a048e23b8c441dcdc9d0edd56497a526f509d3341bf8de1c9d8622d0ff53"} Nov 25 15:23:31 crc kubenswrapper[4890]: I1125 15:23:31.420456 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 15:23:31 crc kubenswrapper[4890]: I1125 15:23:31.439464 4890 scope.go:117] "RemoveContainer" containerID="5c87c8a4cf5e8d339eae1d1cff2df87aca8d18a2acc8d4d86741296390b30367" Nov 25 15:23:31 crc kubenswrapper[4890]: I1125 15:23:31.439635 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 15:23:31 crc kubenswrapper[4890]: I1125 15:23:31.449048 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 25 15:23:31 crc kubenswrapper[4890]: E1125 15:23:31.449636 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9355e0f0-8d98-49a0-9dd4-720841673727" containerName="nova-metadata-metadata" Nov 25 15:23:31 crc kubenswrapper[4890]: I1125 15:23:31.449662 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="9355e0f0-8d98-49a0-9dd4-720841673727" containerName="nova-metadata-metadata" Nov 25 15:23:31 crc kubenswrapper[4890]: E1125 15:23:31.449682 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9355e0f0-8d98-49a0-9dd4-720841673727" containerName="nova-metadata-log" Nov 25 15:23:31 crc kubenswrapper[4890]: I1125 15:23:31.449691 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="9355e0f0-8d98-49a0-9dd4-720841673727" containerName="nova-metadata-log" Nov 25 15:23:31 crc kubenswrapper[4890]: I1125 15:23:31.449966 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="9355e0f0-8d98-49a0-9dd4-720841673727" containerName="nova-metadata-log" Nov 25 15:23:31 crc kubenswrapper[4890]: I1125 15:23:31.449984 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="9355e0f0-8d98-49a0-9dd4-720841673727" containerName="nova-metadata-metadata" Nov 25 15:23:31 crc kubenswrapper[4890]: I1125 15:23:31.451448 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 15:23:31 crc kubenswrapper[4890]: I1125 15:23:31.453671 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 25 15:23:31 crc kubenswrapper[4890]: I1125 15:23:31.462621 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 25 15:23:31 crc kubenswrapper[4890]: I1125 15:23:31.466028 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 15:23:31 crc kubenswrapper[4890]: I1125 15:23:31.496878 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb7922d1-e2c8-4ee7-a700-6e6353de7d91-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"fb7922d1-e2c8-4ee7-a700-6e6353de7d91\") " pod="openstack/nova-metadata-0" Nov 25 15:23:31 crc kubenswrapper[4890]: I1125 15:23:31.497026 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb7922d1-e2c8-4ee7-a700-6e6353de7d91-config-data\") pod \"nova-metadata-0\" (UID: \"fb7922d1-e2c8-4ee7-a700-6e6353de7d91\") " pod="openstack/nova-metadata-0" Nov 25 15:23:31 crc kubenswrapper[4890]: I1125 15:23:31.497123 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb7922d1-e2c8-4ee7-a700-6e6353de7d91-logs\") pod \"nova-metadata-0\" (UID: \"fb7922d1-e2c8-4ee7-a700-6e6353de7d91\") " pod="openstack/nova-metadata-0" Nov 25 15:23:31 crc kubenswrapper[4890]: I1125 15:23:31.497213 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb7922d1-e2c8-4ee7-a700-6e6353de7d91-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fb7922d1-e2c8-4ee7-a700-6e6353de7d91\") " pod="openstack/nova-metadata-0" Nov 25 15:23:31 crc kubenswrapper[4890]: I1125 15:23:31.497257 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsjjl\" (UniqueName: \"kubernetes.io/projected/fb7922d1-e2c8-4ee7-a700-6e6353de7d91-kube-api-access-fsjjl\") pod \"nova-metadata-0\" (UID: \"fb7922d1-e2c8-4ee7-a700-6e6353de7d91\") " pod="openstack/nova-metadata-0" Nov 25 15:23:31 crc kubenswrapper[4890]: I1125 15:23:31.598418 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb7922d1-e2c8-4ee7-a700-6e6353de7d91-logs\") pod \"nova-metadata-0\" (UID: \"fb7922d1-e2c8-4ee7-a700-6e6353de7d91\") " pod="openstack/nova-metadata-0" Nov 25 15:23:31 crc kubenswrapper[4890]: I1125 15:23:31.598478 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb7922d1-e2c8-4ee7-a700-6e6353de7d91-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fb7922d1-e2c8-4ee7-a700-6e6353de7d91\") " pod="openstack/nova-metadata-0" Nov 25 15:23:31 crc kubenswrapper[4890]: I1125 15:23:31.598503 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsjjl\" (UniqueName: \"kubernetes.io/projected/fb7922d1-e2c8-4ee7-a700-6e6353de7d91-kube-api-access-fsjjl\") pod \"nova-metadata-0\" (UID: \"fb7922d1-e2c8-4ee7-a700-6e6353de7d91\") " pod="openstack/nova-metadata-0" Nov 25 15:23:31 crc kubenswrapper[4890]: I1125 15:23:31.598550 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb7922d1-e2c8-4ee7-a700-6e6353de7d91-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"fb7922d1-e2c8-4ee7-a700-6e6353de7d91\") " pod="openstack/nova-metadata-0" Nov 25 15:23:31 crc kubenswrapper[4890]: I1125 15:23:31.598608 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb7922d1-e2c8-4ee7-a700-6e6353de7d91-config-data\") pod \"nova-metadata-0\" (UID: \"fb7922d1-e2c8-4ee7-a700-6e6353de7d91\") " pod="openstack/nova-metadata-0" Nov 25 15:23:31 crc kubenswrapper[4890]: I1125 15:23:31.598922 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb7922d1-e2c8-4ee7-a700-6e6353de7d91-logs\") pod \"nova-metadata-0\" (UID: \"fb7922d1-e2c8-4ee7-a700-6e6353de7d91\") " pod="openstack/nova-metadata-0" Nov 25 15:23:31 crc kubenswrapper[4890]: I1125 15:23:31.603863 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb7922d1-e2c8-4ee7-a700-6e6353de7d91-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"fb7922d1-e2c8-4ee7-a700-6e6353de7d91\") " pod="openstack/nova-metadata-0" Nov 25 15:23:31 crc kubenswrapper[4890]: I1125 15:23:31.604386 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb7922d1-e2c8-4ee7-a700-6e6353de7d91-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fb7922d1-e2c8-4ee7-a700-6e6353de7d91\") " pod="openstack/nova-metadata-0" Nov 25 15:23:31 crc kubenswrapper[4890]: I1125 15:23:31.610227 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb7922d1-e2c8-4ee7-a700-6e6353de7d91-config-data\") pod \"nova-metadata-0\" (UID: \"fb7922d1-e2c8-4ee7-a700-6e6353de7d91\") " pod="openstack/nova-metadata-0" Nov 25 15:23:31 crc kubenswrapper[4890]: I1125 15:23:31.630510 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsjjl\" (UniqueName: \"kubernetes.io/projected/fb7922d1-e2c8-4ee7-a700-6e6353de7d91-kube-api-access-fsjjl\") pod \"nova-metadata-0\" (UID: \"fb7922d1-e2c8-4ee7-a700-6e6353de7d91\") " pod="openstack/nova-metadata-0" Nov 25 15:23:31 crc kubenswrapper[4890]: I1125 15:23:31.779913 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 15:23:31 crc kubenswrapper[4890]: E1125 15:23:31.992482 4890 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="15321cc73c2abb27fce208f45be1bbe219056863f1de914f8bb7838f82f05ffb" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 25 15:23:31 crc kubenswrapper[4890]: E1125 15:23:31.994268 4890 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="15321cc73c2abb27fce208f45be1bbe219056863f1de914f8bb7838f82f05ffb" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 25 15:23:31 crc kubenswrapper[4890]: E1125 15:23:31.996966 4890 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="15321cc73c2abb27fce208f45be1bbe219056863f1de914f8bb7838f82f05ffb" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 25 15:23:31 crc kubenswrapper[4890]: E1125 15:23:31.997042 4890 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="92350db6-3f3b-44c6-915c-8223296fe946" containerName="nova-scheduler-scheduler" Nov 25 15:23:32 crc kubenswrapper[4890]: I1125 15:23:32.186652 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9355e0f0-8d98-49a0-9dd4-720841673727" path="/var/lib/kubelet/pods/9355e0f0-8d98-49a0-9dd4-720841673727/volumes" Nov 25 15:23:32 crc kubenswrapper[4890]: I1125 15:23:32.230639 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 15:23:32 crc kubenswrapper[4890]: W1125 15:23:32.239099 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfb7922d1_e2c8_4ee7_a700_6e6353de7d91.slice/crio-d0f33bdd3b9e241d0c07ec77857039ba1e888eeb76e5a385bfdc38caa75bc7ee WatchSource:0}: Error finding container d0f33bdd3b9e241d0c07ec77857039ba1e888eeb76e5a385bfdc38caa75bc7ee: Status 404 returned error can't find the container with id d0f33bdd3b9e241d0c07ec77857039ba1e888eeb76e5a385bfdc38caa75bc7ee Nov 25 15:23:32 crc kubenswrapper[4890]: I1125 15:23:32.402708 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fb7922d1-e2c8-4ee7-a700-6e6353de7d91","Type":"ContainerStarted","Data":"d0f33bdd3b9e241d0c07ec77857039ba1e888eeb76e5a385bfdc38caa75bc7ee"} Nov 25 15:23:33 crc kubenswrapper[4890]: I1125 15:23:33.417276 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"154b7e43-8329-4c7d-9870-ffca704e80af","Type":"ContainerStarted","Data":"c188f25928575f0e2459d821ee704903e370c8a80d57a9a9a94092152a317754"} Nov 25 15:23:33 crc kubenswrapper[4890]: I1125 15:23:33.420105 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fb7922d1-e2c8-4ee7-a700-6e6353de7d91","Type":"ContainerStarted","Data":"53a207e706c9c4a6fe645567e7859d702ccd877ceb186ecc2155100a47808c7e"} Nov 25 15:23:33 crc kubenswrapper[4890]: I1125 15:23:33.420131 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fb7922d1-e2c8-4ee7-a700-6e6353de7d91","Type":"ContainerStarted","Data":"2b9bbbdc949ac4ef3283ade2efc3a174ce0f3ed26d413db474f54f62904e12d8"} Nov 25 15:23:33 crc kubenswrapper[4890]: I1125 15:23:33.450133 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.450115529 podStartE2EDuration="2.450115529s" podCreationTimestamp="2025-11-25 15:23:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:23:33.442045056 +0000 UTC m=+1271.884507666" watchObservedRunningTime="2025-11-25 15:23:33.450115529 +0000 UTC m=+1271.892578129" Nov 25 15:23:33 crc kubenswrapper[4890]: I1125 15:23:33.951085 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.048217 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92350db6-3f3b-44c6-915c-8223296fe946-config-data\") pod \"92350db6-3f3b-44c6-915c-8223296fe946\" (UID: \"92350db6-3f3b-44c6-915c-8223296fe946\") " Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.048279 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92350db6-3f3b-44c6-915c-8223296fe946-combined-ca-bundle\") pod \"92350db6-3f3b-44c6-915c-8223296fe946\" (UID: \"92350db6-3f3b-44c6-915c-8223296fe946\") " Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.048307 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tqfrz\" (UniqueName: \"kubernetes.io/projected/92350db6-3f3b-44c6-915c-8223296fe946-kube-api-access-tqfrz\") pod \"92350db6-3f3b-44c6-915c-8223296fe946\" (UID: \"92350db6-3f3b-44c6-915c-8223296fe946\") " Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.057071 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92350db6-3f3b-44c6-915c-8223296fe946-kube-api-access-tqfrz" (OuterVolumeSpecName: "kube-api-access-tqfrz") pod "92350db6-3f3b-44c6-915c-8223296fe946" (UID: "92350db6-3f3b-44c6-915c-8223296fe946"). InnerVolumeSpecName "kube-api-access-tqfrz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.088546 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92350db6-3f3b-44c6-915c-8223296fe946-config-data" (OuterVolumeSpecName: "config-data") pod "92350db6-3f3b-44c6-915c-8223296fe946" (UID: "92350db6-3f3b-44c6-915c-8223296fe946"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.131574 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92350db6-3f3b-44c6-915c-8223296fe946-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "92350db6-3f3b-44c6-915c-8223296fe946" (UID: "92350db6-3f3b-44c6-915c-8223296fe946"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.149912 4890 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92350db6-3f3b-44c6-915c-8223296fe946-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.149942 4890 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92350db6-3f3b-44c6-915c-8223296fe946-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.149953 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tqfrz\" (UniqueName: \"kubernetes.io/projected/92350db6-3f3b-44c6-915c-8223296fe946-kube-api-access-tqfrz\") on node \"crc\" DevicePath \"\"" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.365018 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.438438 4890 generic.go:334] "Generic (PLEG): container finished" podID="92350db6-3f3b-44c6-915c-8223296fe946" containerID="15321cc73c2abb27fce208f45be1bbe219056863f1de914f8bb7838f82f05ffb" exitCode=0 Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.438495 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"92350db6-3f3b-44c6-915c-8223296fe946","Type":"ContainerDied","Data":"15321cc73c2abb27fce208f45be1bbe219056863f1de914f8bb7838f82f05ffb"} Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.438525 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"92350db6-3f3b-44c6-915c-8223296fe946","Type":"ContainerDied","Data":"4976e86c186e922cf0c220529e46409ef2648632691213e5018ab6db88632282"} Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.438545 4890 scope.go:117] "RemoveContainer" containerID="15321cc73c2abb27fce208f45be1bbe219056863f1de914f8bb7838f82f05ffb" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.438657 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.442804 4890 generic.go:334] "Generic (PLEG): container finished" podID="49e416d1-e37c-4293-b007-5ec30c040757" containerID="884812514b8afe9a3b5eb43645839401f17cbda7213ace64048c1facc0d00a74" exitCode=0 Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.442867 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.442896 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"49e416d1-e37c-4293-b007-5ec30c040757","Type":"ContainerDied","Data":"884812514b8afe9a3b5eb43645839401f17cbda7213ace64048c1facc0d00a74"} Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.442926 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"49e416d1-e37c-4293-b007-5ec30c040757","Type":"ContainerDied","Data":"64306c3cfa1b006d1e9b88a1b7b9b99bc6fdc44247d3a141caa8131c2f65489b"} Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.446420 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"154b7e43-8329-4c7d-9870-ffca704e80af","Type":"ContainerStarted","Data":"7db4830819acbf94ebff9eadbc3cd19d99ee3e5a7bd2990fbc67d168359b2645"} Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.459550 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49e416d1-e37c-4293-b007-5ec30c040757-config-data\") pod \"49e416d1-e37c-4293-b007-5ec30c040757\" (UID: \"49e416d1-e37c-4293-b007-5ec30c040757\") " Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.459632 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49e416d1-e37c-4293-b007-5ec30c040757-combined-ca-bundle\") pod \"49e416d1-e37c-4293-b007-5ec30c040757\" (UID: \"49e416d1-e37c-4293-b007-5ec30c040757\") " Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.459680 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/49e416d1-e37c-4293-b007-5ec30c040757-logs\") pod \"49e416d1-e37c-4293-b007-5ec30c040757\" (UID: \"49e416d1-e37c-4293-b007-5ec30c040757\") " Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.459719 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-82rvh\" (UniqueName: \"kubernetes.io/projected/49e416d1-e37c-4293-b007-5ec30c040757-kube-api-access-82rvh\") pod \"49e416d1-e37c-4293-b007-5ec30c040757\" (UID: \"49e416d1-e37c-4293-b007-5ec30c040757\") " Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.461131 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49e416d1-e37c-4293-b007-5ec30c040757-logs" (OuterVolumeSpecName: "logs") pod "49e416d1-e37c-4293-b007-5ec30c040757" (UID: "49e416d1-e37c-4293-b007-5ec30c040757"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.462675 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.465493 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49e416d1-e37c-4293-b007-5ec30c040757-kube-api-access-82rvh" (OuterVolumeSpecName: "kube-api-access-82rvh") pod "49e416d1-e37c-4293-b007-5ec30c040757" (UID: "49e416d1-e37c-4293-b007-5ec30c040757"). InnerVolumeSpecName "kube-api-access-82rvh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.466817 4890 scope.go:117] "RemoveContainer" containerID="15321cc73c2abb27fce208f45be1bbe219056863f1de914f8bb7838f82f05ffb" Nov 25 15:23:34 crc kubenswrapper[4890]: E1125 15:23:34.469689 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15321cc73c2abb27fce208f45be1bbe219056863f1de914f8bb7838f82f05ffb\": container with ID starting with 15321cc73c2abb27fce208f45be1bbe219056863f1de914f8bb7838f82f05ffb not found: ID does not exist" containerID="15321cc73c2abb27fce208f45be1bbe219056863f1de914f8bb7838f82f05ffb" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.469728 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15321cc73c2abb27fce208f45be1bbe219056863f1de914f8bb7838f82f05ffb"} err="failed to get container status \"15321cc73c2abb27fce208f45be1bbe219056863f1de914f8bb7838f82f05ffb\": rpc error: code = NotFound desc = could not find container \"15321cc73c2abb27fce208f45be1bbe219056863f1de914f8bb7838f82f05ffb\": container with ID starting with 15321cc73c2abb27fce208f45be1bbe219056863f1de914f8bb7838f82f05ffb not found: ID does not exist" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.469752 4890 scope.go:117] "RemoveContainer" containerID="884812514b8afe9a3b5eb43645839401f17cbda7213ace64048c1facc0d00a74" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.482197 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.492346 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49e416d1-e37c-4293-b007-5ec30c040757-config-data" (OuterVolumeSpecName: "config-data") pod "49e416d1-e37c-4293-b007-5ec30c040757" (UID: "49e416d1-e37c-4293-b007-5ec30c040757"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.492673 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 15:23:34 crc kubenswrapper[4890]: E1125 15:23:34.493298 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49e416d1-e37c-4293-b007-5ec30c040757" containerName="nova-api-api" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.493318 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="49e416d1-e37c-4293-b007-5ec30c040757" containerName="nova-api-api" Nov 25 15:23:34 crc kubenswrapper[4890]: E1125 15:23:34.493334 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92350db6-3f3b-44c6-915c-8223296fe946" containerName="nova-scheduler-scheduler" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.493341 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="92350db6-3f3b-44c6-915c-8223296fe946" containerName="nova-scheduler-scheduler" Nov 25 15:23:34 crc kubenswrapper[4890]: E1125 15:23:34.493363 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49e416d1-e37c-4293-b007-5ec30c040757" containerName="nova-api-log" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.493372 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="49e416d1-e37c-4293-b007-5ec30c040757" containerName="nova-api-log" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.493594 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="49e416d1-e37c-4293-b007-5ec30c040757" containerName="nova-api-api" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.493633 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="92350db6-3f3b-44c6-915c-8223296fe946" containerName="nova-scheduler-scheduler" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.493645 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="49e416d1-e37c-4293-b007-5ec30c040757" containerName="nova-api-log" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.494376 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.502261 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49e416d1-e37c-4293-b007-5ec30c040757-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "49e416d1-e37c-4293-b007-5ec30c040757" (UID: "49e416d1-e37c-4293-b007-5ec30c040757"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.507526 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.508777 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.520837 4890 scope.go:117] "RemoveContainer" containerID="8f13efbe74ff4d27e1662ce4246ec230c23743e053662b64e118dce0f23a53a6" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.542721 4890 scope.go:117] "RemoveContainer" containerID="884812514b8afe9a3b5eb43645839401f17cbda7213ace64048c1facc0d00a74" Nov 25 15:23:34 crc kubenswrapper[4890]: E1125 15:23:34.544016 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"884812514b8afe9a3b5eb43645839401f17cbda7213ace64048c1facc0d00a74\": container with ID starting with 884812514b8afe9a3b5eb43645839401f17cbda7213ace64048c1facc0d00a74 not found: ID does not exist" containerID="884812514b8afe9a3b5eb43645839401f17cbda7213ace64048c1facc0d00a74" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.544175 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"884812514b8afe9a3b5eb43645839401f17cbda7213ace64048c1facc0d00a74"} err="failed to get container status \"884812514b8afe9a3b5eb43645839401f17cbda7213ace64048c1facc0d00a74\": rpc error: code = NotFound desc = could not find container \"884812514b8afe9a3b5eb43645839401f17cbda7213ace64048c1facc0d00a74\": container with ID starting with 884812514b8afe9a3b5eb43645839401f17cbda7213ace64048c1facc0d00a74 not found: ID does not exist" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.544304 4890 scope.go:117] "RemoveContainer" containerID="8f13efbe74ff4d27e1662ce4246ec230c23743e053662b64e118dce0f23a53a6" Nov 25 15:23:34 crc kubenswrapper[4890]: E1125 15:23:34.544885 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f13efbe74ff4d27e1662ce4246ec230c23743e053662b64e118dce0f23a53a6\": container with ID starting with 8f13efbe74ff4d27e1662ce4246ec230c23743e053662b64e118dce0f23a53a6 not found: ID does not exist" containerID="8f13efbe74ff4d27e1662ce4246ec230c23743e053662b64e118dce0f23a53a6" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.544934 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f13efbe74ff4d27e1662ce4246ec230c23743e053662b64e118dce0f23a53a6"} err="failed to get container status \"8f13efbe74ff4d27e1662ce4246ec230c23743e053662b64e118dce0f23a53a6\": rpc error: code = NotFound desc = could not find container \"8f13efbe74ff4d27e1662ce4246ec230c23743e053662b64e118dce0f23a53a6\": container with ID starting with 8f13efbe74ff4d27e1662ce4246ec230c23743e053662b64e118dce0f23a53a6 not found: ID does not exist" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.561960 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12acc506-6412-4a9e-a80c-4b1cc761246d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"12acc506-6412-4a9e-a80c-4b1cc761246d\") " pod="openstack/nova-scheduler-0" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.562004 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cl9sg\" (UniqueName: \"kubernetes.io/projected/12acc506-6412-4a9e-a80c-4b1cc761246d-kube-api-access-cl9sg\") pod \"nova-scheduler-0\" (UID: \"12acc506-6412-4a9e-a80c-4b1cc761246d\") " pod="openstack/nova-scheduler-0" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.562043 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12acc506-6412-4a9e-a80c-4b1cc761246d-config-data\") pod \"nova-scheduler-0\" (UID: \"12acc506-6412-4a9e-a80c-4b1cc761246d\") " pod="openstack/nova-scheduler-0" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.562181 4890 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49e416d1-e37c-4293-b007-5ec30c040757-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.562192 4890 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/49e416d1-e37c-4293-b007-5ec30c040757-logs\") on node \"crc\" DevicePath \"\"" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.562201 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-82rvh\" (UniqueName: \"kubernetes.io/projected/49e416d1-e37c-4293-b007-5ec30c040757-kube-api-access-82rvh\") on node \"crc\" DevicePath \"\"" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.562211 4890 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49e416d1-e37c-4293-b007-5ec30c040757-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.664401 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cl9sg\" (UniqueName: \"kubernetes.io/projected/12acc506-6412-4a9e-a80c-4b1cc761246d-kube-api-access-cl9sg\") pod \"nova-scheduler-0\" (UID: \"12acc506-6412-4a9e-a80c-4b1cc761246d\") " pod="openstack/nova-scheduler-0" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.664518 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12acc506-6412-4a9e-a80c-4b1cc761246d-config-data\") pod \"nova-scheduler-0\" (UID: \"12acc506-6412-4a9e-a80c-4b1cc761246d\") " pod="openstack/nova-scheduler-0" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.664701 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12acc506-6412-4a9e-a80c-4b1cc761246d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"12acc506-6412-4a9e-a80c-4b1cc761246d\") " pod="openstack/nova-scheduler-0" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.669844 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12acc506-6412-4a9e-a80c-4b1cc761246d-config-data\") pod \"nova-scheduler-0\" (UID: \"12acc506-6412-4a9e-a80c-4b1cc761246d\") " pod="openstack/nova-scheduler-0" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.673498 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12acc506-6412-4a9e-a80c-4b1cc761246d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"12acc506-6412-4a9e-a80c-4b1cc761246d\") " pod="openstack/nova-scheduler-0" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.679468 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cl9sg\" (UniqueName: \"kubernetes.io/projected/12acc506-6412-4a9e-a80c-4b1cc761246d-kube-api-access-cl9sg\") pod \"nova-scheduler-0\" (UID: \"12acc506-6412-4a9e-a80c-4b1cc761246d\") " pod="openstack/nova-scheduler-0" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.796525 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.812773 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.823125 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.824732 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.827367 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.833761 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.839925 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.868153 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/324efe97-1d1d-47d9-83c0-631a31cf4d2e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"324efe97-1d1d-47d9-83c0-631a31cf4d2e\") " pod="openstack/nova-api-0" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.868418 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/324efe97-1d1d-47d9-83c0-631a31cf4d2e-logs\") pod \"nova-api-0\" (UID: \"324efe97-1d1d-47d9-83c0-631a31cf4d2e\") " pod="openstack/nova-api-0" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.868575 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lr5vm\" (UniqueName: \"kubernetes.io/projected/324efe97-1d1d-47d9-83c0-631a31cf4d2e-kube-api-access-lr5vm\") pod \"nova-api-0\" (UID: \"324efe97-1d1d-47d9-83c0-631a31cf4d2e\") " pod="openstack/nova-api-0" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.868768 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/324efe97-1d1d-47d9-83c0-631a31cf4d2e-config-data\") pod \"nova-api-0\" (UID: \"324efe97-1d1d-47d9-83c0-631a31cf4d2e\") " pod="openstack/nova-api-0" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.970665 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/324efe97-1d1d-47d9-83c0-631a31cf4d2e-config-data\") pod \"nova-api-0\" (UID: \"324efe97-1d1d-47d9-83c0-631a31cf4d2e\") " pod="openstack/nova-api-0" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.970913 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/324efe97-1d1d-47d9-83c0-631a31cf4d2e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"324efe97-1d1d-47d9-83c0-631a31cf4d2e\") " pod="openstack/nova-api-0" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.970939 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/324efe97-1d1d-47d9-83c0-631a31cf4d2e-logs\") pod \"nova-api-0\" (UID: \"324efe97-1d1d-47d9-83c0-631a31cf4d2e\") " pod="openstack/nova-api-0" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.970973 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lr5vm\" (UniqueName: \"kubernetes.io/projected/324efe97-1d1d-47d9-83c0-631a31cf4d2e-kube-api-access-lr5vm\") pod \"nova-api-0\" (UID: \"324efe97-1d1d-47d9-83c0-631a31cf4d2e\") " pod="openstack/nova-api-0" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.971664 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/324efe97-1d1d-47d9-83c0-631a31cf4d2e-logs\") pod \"nova-api-0\" (UID: \"324efe97-1d1d-47d9-83c0-631a31cf4d2e\") " pod="openstack/nova-api-0" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.984470 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/324efe97-1d1d-47d9-83c0-631a31cf4d2e-config-data\") pod \"nova-api-0\" (UID: \"324efe97-1d1d-47d9-83c0-631a31cf4d2e\") " pod="openstack/nova-api-0" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.986478 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/324efe97-1d1d-47d9-83c0-631a31cf4d2e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"324efe97-1d1d-47d9-83c0-631a31cf4d2e\") " pod="openstack/nova-api-0" Nov 25 15:23:34 crc kubenswrapper[4890]: I1125 15:23:34.994674 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lr5vm\" (UniqueName: \"kubernetes.io/projected/324efe97-1d1d-47d9-83c0-631a31cf4d2e-kube-api-access-lr5vm\") pod \"nova-api-0\" (UID: \"324efe97-1d1d-47d9-83c0-631a31cf4d2e\") " pod="openstack/nova-api-0" Nov 25 15:23:35 crc kubenswrapper[4890]: I1125 15:23:35.257179 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 15:23:35 crc kubenswrapper[4890]: I1125 15:23:35.396385 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 15:23:35 crc kubenswrapper[4890]: I1125 15:23:35.459555 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"12acc506-6412-4a9e-a80c-4b1cc761246d","Type":"ContainerStarted","Data":"1cc5a35a5b85d593f9a447a8791877cc01e4ee3875a3bf6f0782bcffc96ff089"} Nov 25 15:23:35 crc kubenswrapper[4890]: I1125 15:23:35.707928 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 15:23:35 crc kubenswrapper[4890]: W1125 15:23:35.715809 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod324efe97_1d1d_47d9_83c0_631a31cf4d2e.slice/crio-95489b8853f2c989cce2cf433d410053adb75580ff8c34e3e992c8c76110eddb WatchSource:0}: Error finding container 95489b8853f2c989cce2cf433d410053adb75580ff8c34e3e992c8c76110eddb: Status 404 returned error can't find the container with id 95489b8853f2c989cce2cf433d410053adb75580ff8c34e3e992c8c76110eddb Nov 25 15:23:35 crc kubenswrapper[4890]: I1125 15:23:35.760231 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 25 15:23:36 crc kubenswrapper[4890]: I1125 15:23:36.185551 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49e416d1-e37c-4293-b007-5ec30c040757" path="/var/lib/kubelet/pods/49e416d1-e37c-4293-b007-5ec30c040757/volumes" Nov 25 15:23:36 crc kubenswrapper[4890]: I1125 15:23:36.188712 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92350db6-3f3b-44c6-915c-8223296fe946" path="/var/lib/kubelet/pods/92350db6-3f3b-44c6-915c-8223296fe946/volumes" Nov 25 15:23:36 crc kubenswrapper[4890]: I1125 15:23:36.480903 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"324efe97-1d1d-47d9-83c0-631a31cf4d2e","Type":"ContainerStarted","Data":"36f461bec3aa616aac770d2c2fb97fd0685688dcccb1c4ed337fdd4c08fdf413"} Nov 25 15:23:36 crc kubenswrapper[4890]: I1125 15:23:36.480954 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"324efe97-1d1d-47d9-83c0-631a31cf4d2e","Type":"ContainerStarted","Data":"95489b8853f2c989cce2cf433d410053adb75580ff8c34e3e992c8c76110eddb"} Nov 25 15:23:36 crc kubenswrapper[4890]: I1125 15:23:36.486921 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"12acc506-6412-4a9e-a80c-4b1cc761246d","Type":"ContainerStarted","Data":"7364e8cda4f00d093052c4416a5150b9c81e120eb72578c5db8e133dd20d94fa"} Nov 25 15:23:36 crc kubenswrapper[4890]: I1125 15:23:36.511216 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.511196027 podStartE2EDuration="2.511196027s" podCreationTimestamp="2025-11-25 15:23:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:23:36.504695534 +0000 UTC m=+1274.947158144" watchObservedRunningTime="2025-11-25 15:23:36.511196027 +0000 UTC m=+1274.953658637" Nov 25 15:23:36 crc kubenswrapper[4890]: I1125 15:23:36.780454 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 25 15:23:36 crc kubenswrapper[4890]: I1125 15:23:36.780522 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 25 15:23:37 crc kubenswrapper[4890]: I1125 15:23:37.498622 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"154b7e43-8329-4c7d-9870-ffca704e80af","Type":"ContainerStarted","Data":"aca91755e7903d75a205a2f0efa57c85839df1ef3b758ce7e6453c9e38094964"} Nov 25 15:23:37 crc kubenswrapper[4890]: I1125 15:23:37.499853 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 25 15:23:37 crc kubenswrapper[4890]: I1125 15:23:37.502960 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"324efe97-1d1d-47d9-83c0-631a31cf4d2e","Type":"ContainerStarted","Data":"c1976273a01588d0ee2baa87932d07d78389f02d3e7b6bfc38257dd07b754dbf"} Nov 25 15:23:37 crc kubenswrapper[4890]: I1125 15:23:37.526883 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.583768682 podStartE2EDuration="8.526862601s" podCreationTimestamp="2025-11-25 15:23:29 +0000 UTC" firstStartedPulling="2025-11-25 15:23:30.25411338 +0000 UTC m=+1268.696576000" lastFinishedPulling="2025-11-25 15:23:36.197207309 +0000 UTC m=+1274.639669919" observedRunningTime="2025-11-25 15:23:37.518113281 +0000 UTC m=+1275.960575921" watchObservedRunningTime="2025-11-25 15:23:37.526862601 +0000 UTC m=+1275.969325211" Nov 25 15:23:37 crc kubenswrapper[4890]: I1125 15:23:37.551024 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.551004818 podStartE2EDuration="3.551004818s" podCreationTimestamp="2025-11-25 15:23:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:23:37.535550959 +0000 UTC m=+1275.978013579" watchObservedRunningTime="2025-11-25 15:23:37.551004818 +0000 UTC m=+1275.993467428" Nov 25 15:23:39 crc kubenswrapper[4890]: I1125 15:23:39.834591 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 25 15:23:40 crc kubenswrapper[4890]: I1125 15:23:40.531133 4890 generic.go:334] "Generic (PLEG): container finished" podID="70c6170b-9975-487c-ac2d-a2a309833658" containerID="b76d7672c9af03f6cfa1ad825433495b765e29311771d9a0c6a518019d549333" exitCode=0 Nov 25 15:23:40 crc kubenswrapper[4890]: I1125 15:23:40.531520 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-828xw" event={"ID":"70c6170b-9975-487c-ac2d-a2a309833658","Type":"ContainerDied","Data":"b76d7672c9af03f6cfa1ad825433495b765e29311771d9a0c6a518019d549333"} Nov 25 15:23:41 crc kubenswrapper[4890]: I1125 15:23:41.780095 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 25 15:23:41 crc kubenswrapper[4890]: I1125 15:23:41.780260 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 25 15:23:41 crc kubenswrapper[4890]: I1125 15:23:41.893977 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-828xw" Nov 25 15:23:42 crc kubenswrapper[4890]: I1125 15:23:42.000948 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qcbxm\" (UniqueName: \"kubernetes.io/projected/70c6170b-9975-487c-ac2d-a2a309833658-kube-api-access-qcbxm\") pod \"70c6170b-9975-487c-ac2d-a2a309833658\" (UID: \"70c6170b-9975-487c-ac2d-a2a309833658\") " Nov 25 15:23:42 crc kubenswrapper[4890]: I1125 15:23:42.001060 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70c6170b-9975-487c-ac2d-a2a309833658-scripts\") pod \"70c6170b-9975-487c-ac2d-a2a309833658\" (UID: \"70c6170b-9975-487c-ac2d-a2a309833658\") " Nov 25 15:23:42 crc kubenswrapper[4890]: I1125 15:23:42.001119 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70c6170b-9975-487c-ac2d-a2a309833658-config-data\") pod \"70c6170b-9975-487c-ac2d-a2a309833658\" (UID: \"70c6170b-9975-487c-ac2d-a2a309833658\") " Nov 25 15:23:42 crc kubenswrapper[4890]: I1125 15:23:42.001173 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70c6170b-9975-487c-ac2d-a2a309833658-combined-ca-bundle\") pod \"70c6170b-9975-487c-ac2d-a2a309833658\" (UID: \"70c6170b-9975-487c-ac2d-a2a309833658\") " Nov 25 15:23:42 crc kubenswrapper[4890]: I1125 15:23:42.007411 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70c6170b-9975-487c-ac2d-a2a309833658-scripts" (OuterVolumeSpecName: "scripts") pod "70c6170b-9975-487c-ac2d-a2a309833658" (UID: "70c6170b-9975-487c-ac2d-a2a309833658"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:23:42 crc kubenswrapper[4890]: I1125 15:23:42.007411 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70c6170b-9975-487c-ac2d-a2a309833658-kube-api-access-qcbxm" (OuterVolumeSpecName: "kube-api-access-qcbxm") pod "70c6170b-9975-487c-ac2d-a2a309833658" (UID: "70c6170b-9975-487c-ac2d-a2a309833658"). InnerVolumeSpecName "kube-api-access-qcbxm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:23:42 crc kubenswrapper[4890]: I1125 15:23:42.029679 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70c6170b-9975-487c-ac2d-a2a309833658-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "70c6170b-9975-487c-ac2d-a2a309833658" (UID: "70c6170b-9975-487c-ac2d-a2a309833658"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:23:42 crc kubenswrapper[4890]: I1125 15:23:42.032300 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70c6170b-9975-487c-ac2d-a2a309833658-config-data" (OuterVolumeSpecName: "config-data") pod "70c6170b-9975-487c-ac2d-a2a309833658" (UID: "70c6170b-9975-487c-ac2d-a2a309833658"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:23:42 crc kubenswrapper[4890]: I1125 15:23:42.104872 4890 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70c6170b-9975-487c-ac2d-a2a309833658-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:23:42 crc kubenswrapper[4890]: I1125 15:23:42.104905 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qcbxm\" (UniqueName: \"kubernetes.io/projected/70c6170b-9975-487c-ac2d-a2a309833658-kube-api-access-qcbxm\") on node \"crc\" DevicePath \"\"" Nov 25 15:23:42 crc kubenswrapper[4890]: I1125 15:23:42.104917 4890 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70c6170b-9975-487c-ac2d-a2a309833658-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 15:23:42 crc kubenswrapper[4890]: I1125 15:23:42.104925 4890 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70c6170b-9975-487c-ac2d-a2a309833658-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 15:23:42 crc kubenswrapper[4890]: I1125 15:23:42.549143 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-828xw" event={"ID":"70c6170b-9975-487c-ac2d-a2a309833658","Type":"ContainerDied","Data":"2edd2f1868d79aada5c6f4f563af558ef2fb3b1efebc4ef8844df88a7a9cef20"} Nov 25 15:23:42 crc kubenswrapper[4890]: I1125 15:23:42.549211 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2edd2f1868d79aada5c6f4f563af558ef2fb3b1efebc4ef8844df88a7a9cef20" Nov 25 15:23:42 crc kubenswrapper[4890]: I1125 15:23:42.549225 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-828xw" Nov 25 15:23:42 crc kubenswrapper[4890]: I1125 15:23:42.638706 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 25 15:23:42 crc kubenswrapper[4890]: E1125 15:23:42.639172 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70c6170b-9975-487c-ac2d-a2a309833658" containerName="nova-cell1-conductor-db-sync" Nov 25 15:23:42 crc kubenswrapper[4890]: I1125 15:23:42.639189 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="70c6170b-9975-487c-ac2d-a2a309833658" containerName="nova-cell1-conductor-db-sync" Nov 25 15:23:42 crc kubenswrapper[4890]: I1125 15:23:42.639403 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="70c6170b-9975-487c-ac2d-a2a309833658" containerName="nova-cell1-conductor-db-sync" Nov 25 15:23:42 crc kubenswrapper[4890]: I1125 15:23:42.640076 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 25 15:23:42 crc kubenswrapper[4890]: I1125 15:23:42.642477 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 25 15:23:42 crc kubenswrapper[4890]: I1125 15:23:42.662381 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 25 15:23:42 crc kubenswrapper[4890]: I1125 15:23:42.716806 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ks57h\" (UniqueName: \"kubernetes.io/projected/59c66e1f-c917-441e-99d4-10980ac6891e-kube-api-access-ks57h\") pod \"nova-cell1-conductor-0\" (UID: \"59c66e1f-c917-441e-99d4-10980ac6891e\") " pod="openstack/nova-cell1-conductor-0" Nov 25 15:23:42 crc kubenswrapper[4890]: I1125 15:23:42.716987 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59c66e1f-c917-441e-99d4-10980ac6891e-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"59c66e1f-c917-441e-99d4-10980ac6891e\") " pod="openstack/nova-cell1-conductor-0" Nov 25 15:23:42 crc kubenswrapper[4890]: I1125 15:23:42.717392 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59c66e1f-c917-441e-99d4-10980ac6891e-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"59c66e1f-c917-441e-99d4-10980ac6891e\") " pod="openstack/nova-cell1-conductor-0" Nov 25 15:23:42 crc kubenswrapper[4890]: I1125 15:23:42.796321 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="fb7922d1-e2c8-4ee7-a700-6e6353de7d91" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.196:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 15:23:42 crc kubenswrapper[4890]: I1125 15:23:42.796365 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="fb7922d1-e2c8-4ee7-a700-6e6353de7d91" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.196:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 15:23:42 crc kubenswrapper[4890]: I1125 15:23:42.819557 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59c66e1f-c917-441e-99d4-10980ac6891e-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"59c66e1f-c917-441e-99d4-10980ac6891e\") " pod="openstack/nova-cell1-conductor-0" Nov 25 15:23:42 crc kubenswrapper[4890]: I1125 15:23:42.819669 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ks57h\" (UniqueName: \"kubernetes.io/projected/59c66e1f-c917-441e-99d4-10980ac6891e-kube-api-access-ks57h\") pod \"nova-cell1-conductor-0\" (UID: \"59c66e1f-c917-441e-99d4-10980ac6891e\") " pod="openstack/nova-cell1-conductor-0" Nov 25 15:23:42 crc kubenswrapper[4890]: I1125 15:23:42.819710 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59c66e1f-c917-441e-99d4-10980ac6891e-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"59c66e1f-c917-441e-99d4-10980ac6891e\") " pod="openstack/nova-cell1-conductor-0" Nov 25 15:23:42 crc kubenswrapper[4890]: I1125 15:23:42.834195 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59c66e1f-c917-441e-99d4-10980ac6891e-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"59c66e1f-c917-441e-99d4-10980ac6891e\") " pod="openstack/nova-cell1-conductor-0" Nov 25 15:23:42 crc kubenswrapper[4890]: I1125 15:23:42.838108 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59c66e1f-c917-441e-99d4-10980ac6891e-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"59c66e1f-c917-441e-99d4-10980ac6891e\") " pod="openstack/nova-cell1-conductor-0" Nov 25 15:23:42 crc kubenswrapper[4890]: I1125 15:23:42.838502 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ks57h\" (UniqueName: \"kubernetes.io/projected/59c66e1f-c917-441e-99d4-10980ac6891e-kube-api-access-ks57h\") pod \"nova-cell1-conductor-0\" (UID: \"59c66e1f-c917-441e-99d4-10980ac6891e\") " pod="openstack/nova-cell1-conductor-0" Nov 25 15:23:42 crc kubenswrapper[4890]: I1125 15:23:42.958834 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 25 15:23:43 crc kubenswrapper[4890]: I1125 15:23:43.451183 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 25 15:23:43 crc kubenswrapper[4890]: W1125 15:23:43.457097 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod59c66e1f_c917_441e_99d4_10980ac6891e.slice/crio-eecbe65d08e1f66bec374348bf1eea1e9720bcfa3370e4267fae86f3b9f0b2aa WatchSource:0}: Error finding container eecbe65d08e1f66bec374348bf1eea1e9720bcfa3370e4267fae86f3b9f0b2aa: Status 404 returned error can't find the container with id eecbe65d08e1f66bec374348bf1eea1e9720bcfa3370e4267fae86f3b9f0b2aa Nov 25 15:23:43 crc kubenswrapper[4890]: I1125 15:23:43.559399 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"59c66e1f-c917-441e-99d4-10980ac6891e","Type":"ContainerStarted","Data":"eecbe65d08e1f66bec374348bf1eea1e9720bcfa3370e4267fae86f3b9f0b2aa"} Nov 25 15:23:44 crc kubenswrapper[4890]: I1125 15:23:44.569041 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"59c66e1f-c917-441e-99d4-10980ac6891e","Type":"ContainerStarted","Data":"9e216da18622fe16fa2038c525a1d13bdaf38e6154beba795fccdad53647ec87"} Nov 25 15:23:44 crc kubenswrapper[4890]: I1125 15:23:44.570470 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Nov 25 15:23:44 crc kubenswrapper[4890]: I1125 15:23:44.594934 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.59491877 podStartE2EDuration="2.59491877s" podCreationTimestamp="2025-11-25 15:23:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:23:44.586819746 +0000 UTC m=+1283.029282356" watchObservedRunningTime="2025-11-25 15:23:44.59491877 +0000 UTC m=+1283.037381380" Nov 25 15:23:44 crc kubenswrapper[4890]: I1125 15:23:44.835075 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 25 15:23:44 crc kubenswrapper[4890]: I1125 15:23:44.869842 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 25 15:23:45 crc kubenswrapper[4890]: I1125 15:23:45.257903 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 25 15:23:45 crc kubenswrapper[4890]: I1125 15:23:45.257976 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 25 15:23:45 crc kubenswrapper[4890]: I1125 15:23:45.610972 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 25 15:23:46 crc kubenswrapper[4890]: I1125 15:23:46.340350 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="324efe97-1d1d-47d9-83c0-631a31cf4d2e" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.198:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 25 15:23:46 crc kubenswrapper[4890]: I1125 15:23:46.340388 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="324efe97-1d1d-47d9-83c0-631a31cf4d2e" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.198:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 25 15:23:51 crc kubenswrapper[4890]: I1125 15:23:51.787589 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 25 15:23:51 crc kubenswrapper[4890]: I1125 15:23:51.790745 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 25 15:23:51 crc kubenswrapper[4890]: I1125 15:23:51.794126 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 25 15:23:52 crc kubenswrapper[4890]: I1125 15:23:52.599423 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 25 15:23:52 crc kubenswrapper[4890]: I1125 15:23:52.640263 4890 generic.go:334] "Generic (PLEG): container finished" podID="f7874ec1-b9cb-4fc3-a5af-002ceb718857" containerID="86bba2dc4137e2dc190867b6156e8f0a083e36324a8432d17e4da208e31170c7" exitCode=137 Nov 25 15:23:52 crc kubenswrapper[4890]: I1125 15:23:52.640300 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 25 15:23:52 crc kubenswrapper[4890]: I1125 15:23:52.640340 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"f7874ec1-b9cb-4fc3-a5af-002ceb718857","Type":"ContainerDied","Data":"86bba2dc4137e2dc190867b6156e8f0a083e36324a8432d17e4da208e31170c7"} Nov 25 15:23:52 crc kubenswrapper[4890]: I1125 15:23:52.640387 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"f7874ec1-b9cb-4fc3-a5af-002ceb718857","Type":"ContainerDied","Data":"c1ce1936e66e601c7026e7f2e98633c46322a184ebb215d6739126f049f88146"} Nov 25 15:23:52 crc kubenswrapper[4890]: I1125 15:23:52.640409 4890 scope.go:117] "RemoveContainer" containerID="86bba2dc4137e2dc190867b6156e8f0a083e36324a8432d17e4da208e31170c7" Nov 25 15:23:52 crc kubenswrapper[4890]: I1125 15:23:52.647151 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 25 15:23:52 crc kubenswrapper[4890]: I1125 15:23:52.675637 4890 scope.go:117] "RemoveContainer" containerID="86bba2dc4137e2dc190867b6156e8f0a083e36324a8432d17e4da208e31170c7" Nov 25 15:23:52 crc kubenswrapper[4890]: E1125 15:23:52.676007 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86bba2dc4137e2dc190867b6156e8f0a083e36324a8432d17e4da208e31170c7\": container with ID starting with 86bba2dc4137e2dc190867b6156e8f0a083e36324a8432d17e4da208e31170c7 not found: ID does not exist" containerID="86bba2dc4137e2dc190867b6156e8f0a083e36324a8432d17e4da208e31170c7" Nov 25 15:23:52 crc kubenswrapper[4890]: I1125 15:23:52.676050 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86bba2dc4137e2dc190867b6156e8f0a083e36324a8432d17e4da208e31170c7"} err="failed to get container status \"86bba2dc4137e2dc190867b6156e8f0a083e36324a8432d17e4da208e31170c7\": rpc error: code = NotFound desc = could not find container \"86bba2dc4137e2dc190867b6156e8f0a083e36324a8432d17e4da208e31170c7\": container with ID starting with 86bba2dc4137e2dc190867b6156e8f0a083e36324a8432d17e4da208e31170c7 not found: ID does not exist" Nov 25 15:23:52 crc kubenswrapper[4890]: I1125 15:23:52.711665 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lp6cx\" (UniqueName: \"kubernetes.io/projected/f7874ec1-b9cb-4fc3-a5af-002ceb718857-kube-api-access-lp6cx\") pod \"f7874ec1-b9cb-4fc3-a5af-002ceb718857\" (UID: \"f7874ec1-b9cb-4fc3-a5af-002ceb718857\") " Nov 25 15:23:52 crc kubenswrapper[4890]: I1125 15:23:52.712757 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7874ec1-b9cb-4fc3-a5af-002ceb718857-combined-ca-bundle\") pod \"f7874ec1-b9cb-4fc3-a5af-002ceb718857\" (UID: \"f7874ec1-b9cb-4fc3-a5af-002ceb718857\") " Nov 25 15:23:52 crc kubenswrapper[4890]: I1125 15:23:52.713440 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7874ec1-b9cb-4fc3-a5af-002ceb718857-config-data\") pod \"f7874ec1-b9cb-4fc3-a5af-002ceb718857\" (UID: \"f7874ec1-b9cb-4fc3-a5af-002ceb718857\") " Nov 25 15:23:52 crc kubenswrapper[4890]: I1125 15:23:52.718786 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7874ec1-b9cb-4fc3-a5af-002ceb718857-kube-api-access-lp6cx" (OuterVolumeSpecName: "kube-api-access-lp6cx") pod "f7874ec1-b9cb-4fc3-a5af-002ceb718857" (UID: "f7874ec1-b9cb-4fc3-a5af-002ceb718857"). InnerVolumeSpecName "kube-api-access-lp6cx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:23:52 crc kubenswrapper[4890]: I1125 15:23:52.761439 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7874ec1-b9cb-4fc3-a5af-002ceb718857-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f7874ec1-b9cb-4fc3-a5af-002ceb718857" (UID: "f7874ec1-b9cb-4fc3-a5af-002ceb718857"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:23:52 crc kubenswrapper[4890]: I1125 15:23:52.761702 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7874ec1-b9cb-4fc3-a5af-002ceb718857-config-data" (OuterVolumeSpecName: "config-data") pod "f7874ec1-b9cb-4fc3-a5af-002ceb718857" (UID: "f7874ec1-b9cb-4fc3-a5af-002ceb718857"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:23:52 crc kubenswrapper[4890]: I1125 15:23:52.815934 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lp6cx\" (UniqueName: \"kubernetes.io/projected/f7874ec1-b9cb-4fc3-a5af-002ceb718857-kube-api-access-lp6cx\") on node \"crc\" DevicePath \"\"" Nov 25 15:23:52 crc kubenswrapper[4890]: I1125 15:23:52.815989 4890 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7874ec1-b9cb-4fc3-a5af-002ceb718857-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:23:52 crc kubenswrapper[4890]: I1125 15:23:52.816000 4890 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7874ec1-b9cb-4fc3-a5af-002ceb718857-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 15:23:52 crc kubenswrapper[4890]: I1125 15:23:52.988906 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 15:23:53 crc kubenswrapper[4890]: I1125 15:23:53.009583 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Nov 25 15:23:53 crc kubenswrapper[4890]: I1125 15:23:53.019456 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 15:23:53 crc kubenswrapper[4890]: I1125 15:23:53.030482 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 15:23:53 crc kubenswrapper[4890]: E1125 15:23:53.031336 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7874ec1-b9cb-4fc3-a5af-002ceb718857" containerName="nova-cell1-novncproxy-novncproxy" Nov 25 15:23:53 crc kubenswrapper[4890]: I1125 15:23:53.031456 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7874ec1-b9cb-4fc3-a5af-002ceb718857" containerName="nova-cell1-novncproxy-novncproxy" Nov 25 15:23:53 crc kubenswrapper[4890]: I1125 15:23:53.031820 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7874ec1-b9cb-4fc3-a5af-002ceb718857" containerName="nova-cell1-novncproxy-novncproxy" Nov 25 15:23:53 crc kubenswrapper[4890]: I1125 15:23:53.039555 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 25 15:23:53 crc kubenswrapper[4890]: I1125 15:23:53.042489 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 25 15:23:53 crc kubenswrapper[4890]: I1125 15:23:53.042789 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Nov 25 15:23:53 crc kubenswrapper[4890]: I1125 15:23:53.043395 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Nov 25 15:23:53 crc kubenswrapper[4890]: I1125 15:23:53.043631 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 15:23:53 crc kubenswrapper[4890]: I1125 15:23:53.222239 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85887546-bd08-4f9d-826a-ec30d6dd3dee-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"85887546-bd08-4f9d-826a-ec30d6dd3dee\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 15:23:53 crc kubenswrapper[4890]: I1125 15:23:53.222339 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85887546-bd08-4f9d-826a-ec30d6dd3dee-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"85887546-bd08-4f9d-826a-ec30d6dd3dee\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 15:23:53 crc kubenswrapper[4890]: I1125 15:23:53.222379 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6v2z\" (UniqueName: \"kubernetes.io/projected/85887546-bd08-4f9d-826a-ec30d6dd3dee-kube-api-access-s6v2z\") pod \"nova-cell1-novncproxy-0\" (UID: \"85887546-bd08-4f9d-826a-ec30d6dd3dee\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 15:23:53 crc kubenswrapper[4890]: I1125 15:23:53.222398 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/85887546-bd08-4f9d-826a-ec30d6dd3dee-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"85887546-bd08-4f9d-826a-ec30d6dd3dee\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 15:23:53 crc kubenswrapper[4890]: I1125 15:23:53.222519 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/85887546-bd08-4f9d-826a-ec30d6dd3dee-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"85887546-bd08-4f9d-826a-ec30d6dd3dee\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 15:23:53 crc kubenswrapper[4890]: I1125 15:23:53.323927 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/85887546-bd08-4f9d-826a-ec30d6dd3dee-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"85887546-bd08-4f9d-826a-ec30d6dd3dee\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 15:23:53 crc kubenswrapper[4890]: I1125 15:23:53.324008 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85887546-bd08-4f9d-826a-ec30d6dd3dee-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"85887546-bd08-4f9d-826a-ec30d6dd3dee\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 15:23:53 crc kubenswrapper[4890]: I1125 15:23:53.324081 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85887546-bd08-4f9d-826a-ec30d6dd3dee-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"85887546-bd08-4f9d-826a-ec30d6dd3dee\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 15:23:53 crc kubenswrapper[4890]: I1125 15:23:53.324128 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6v2z\" (UniqueName: \"kubernetes.io/projected/85887546-bd08-4f9d-826a-ec30d6dd3dee-kube-api-access-s6v2z\") pod \"nova-cell1-novncproxy-0\" (UID: \"85887546-bd08-4f9d-826a-ec30d6dd3dee\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 15:23:53 crc kubenswrapper[4890]: I1125 15:23:53.324154 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/85887546-bd08-4f9d-826a-ec30d6dd3dee-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"85887546-bd08-4f9d-826a-ec30d6dd3dee\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 15:23:53 crc kubenswrapper[4890]: I1125 15:23:53.327829 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/85887546-bd08-4f9d-826a-ec30d6dd3dee-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"85887546-bd08-4f9d-826a-ec30d6dd3dee\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 15:23:53 crc kubenswrapper[4890]: I1125 15:23:53.331011 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/85887546-bd08-4f9d-826a-ec30d6dd3dee-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"85887546-bd08-4f9d-826a-ec30d6dd3dee\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 15:23:53 crc kubenswrapper[4890]: I1125 15:23:53.331615 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85887546-bd08-4f9d-826a-ec30d6dd3dee-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"85887546-bd08-4f9d-826a-ec30d6dd3dee\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 15:23:53 crc kubenswrapper[4890]: I1125 15:23:53.336273 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85887546-bd08-4f9d-826a-ec30d6dd3dee-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"85887546-bd08-4f9d-826a-ec30d6dd3dee\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 15:23:53 crc kubenswrapper[4890]: I1125 15:23:53.348963 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6v2z\" (UniqueName: \"kubernetes.io/projected/85887546-bd08-4f9d-826a-ec30d6dd3dee-kube-api-access-s6v2z\") pod \"nova-cell1-novncproxy-0\" (UID: \"85887546-bd08-4f9d-826a-ec30d6dd3dee\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 15:23:53 crc kubenswrapper[4890]: I1125 15:23:53.361854 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 25 15:23:53 crc kubenswrapper[4890]: I1125 15:23:53.843504 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 15:23:53 crc kubenswrapper[4890]: W1125 15:23:53.849011 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod85887546_bd08_4f9d_826a_ec30d6dd3dee.slice/crio-030ad6bd6b23dbc08d62b6754b7f4c733f968cd90ccdd60570de160a079fdac2 WatchSource:0}: Error finding container 030ad6bd6b23dbc08d62b6754b7f4c733f968cd90ccdd60570de160a079fdac2: Status 404 returned error can't find the container with id 030ad6bd6b23dbc08d62b6754b7f4c733f968cd90ccdd60570de160a079fdac2 Nov 25 15:23:54 crc kubenswrapper[4890]: I1125 15:23:54.184961 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7874ec1-b9cb-4fc3-a5af-002ceb718857" path="/var/lib/kubelet/pods/f7874ec1-b9cb-4fc3-a5af-002ceb718857/volumes" Nov 25 15:23:54 crc kubenswrapper[4890]: I1125 15:23:54.658719 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"85887546-bd08-4f9d-826a-ec30d6dd3dee","Type":"ContainerStarted","Data":"96f07f2e1f79e425ed4b6b410885290f8c7ea1b3ca0b6494fc53f269cecd1ac8"} Nov 25 15:23:54 crc kubenswrapper[4890]: I1125 15:23:54.659074 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"85887546-bd08-4f9d-826a-ec30d6dd3dee","Type":"ContainerStarted","Data":"030ad6bd6b23dbc08d62b6754b7f4c733f968cd90ccdd60570de160a079fdac2"} Nov 25 15:23:54 crc kubenswrapper[4890]: I1125 15:23:54.673876 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.673858755 podStartE2EDuration="2.673858755s" podCreationTimestamp="2025-11-25 15:23:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:23:54.673717931 +0000 UTC m=+1293.116180551" watchObservedRunningTime="2025-11-25 15:23:54.673858755 +0000 UTC m=+1293.116321355" Nov 25 15:23:55 crc kubenswrapper[4890]: I1125 15:23:55.262023 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 25 15:23:55 crc kubenswrapper[4890]: I1125 15:23:55.262934 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 25 15:23:55 crc kubenswrapper[4890]: I1125 15:23:55.265506 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 25 15:23:55 crc kubenswrapper[4890]: I1125 15:23:55.266625 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 25 15:23:55 crc kubenswrapper[4890]: I1125 15:23:55.667805 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 25 15:23:55 crc kubenswrapper[4890]: I1125 15:23:55.671205 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 25 15:23:55 crc kubenswrapper[4890]: I1125 15:23:55.846548 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-4fhbm"] Nov 25 15:23:55 crc kubenswrapper[4890]: I1125 15:23:55.848589 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-4fhbm" Nov 25 15:23:55 crc kubenswrapper[4890]: I1125 15:23:55.867722 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-4fhbm"] Nov 25 15:23:55 crc kubenswrapper[4890]: I1125 15:23:55.985036 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/96c4c8fb-37d7-4415-aef1-6da42e54fbc2-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-4fhbm\" (UID: \"96c4c8fb-37d7-4415-aef1-6da42e54fbc2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-4fhbm" Nov 25 15:23:55 crc kubenswrapper[4890]: I1125 15:23:55.985146 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/96c4c8fb-37d7-4415-aef1-6da42e54fbc2-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-4fhbm\" (UID: \"96c4c8fb-37d7-4415-aef1-6da42e54fbc2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-4fhbm" Nov 25 15:23:55 crc kubenswrapper[4890]: I1125 15:23:55.985300 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jf74x\" (UniqueName: \"kubernetes.io/projected/96c4c8fb-37d7-4415-aef1-6da42e54fbc2-kube-api-access-jf74x\") pod \"dnsmasq-dns-89c5cd4d5-4fhbm\" (UID: \"96c4c8fb-37d7-4415-aef1-6da42e54fbc2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-4fhbm" Nov 25 15:23:55 crc kubenswrapper[4890]: I1125 15:23:55.985496 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96c4c8fb-37d7-4415-aef1-6da42e54fbc2-config\") pod \"dnsmasq-dns-89c5cd4d5-4fhbm\" (UID: \"96c4c8fb-37d7-4415-aef1-6da42e54fbc2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-4fhbm" Nov 25 15:23:55 crc kubenswrapper[4890]: I1125 15:23:55.985756 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/96c4c8fb-37d7-4415-aef1-6da42e54fbc2-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-4fhbm\" (UID: \"96c4c8fb-37d7-4415-aef1-6da42e54fbc2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-4fhbm" Nov 25 15:23:55 crc kubenswrapper[4890]: I1125 15:23:55.986412 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/96c4c8fb-37d7-4415-aef1-6da42e54fbc2-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-4fhbm\" (UID: \"96c4c8fb-37d7-4415-aef1-6da42e54fbc2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-4fhbm" Nov 25 15:23:56 crc kubenswrapper[4890]: I1125 15:23:56.088716 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96c4c8fb-37d7-4415-aef1-6da42e54fbc2-config\") pod \"dnsmasq-dns-89c5cd4d5-4fhbm\" (UID: \"96c4c8fb-37d7-4415-aef1-6da42e54fbc2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-4fhbm" Nov 25 15:23:56 crc kubenswrapper[4890]: I1125 15:23:56.089196 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/96c4c8fb-37d7-4415-aef1-6da42e54fbc2-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-4fhbm\" (UID: \"96c4c8fb-37d7-4415-aef1-6da42e54fbc2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-4fhbm" Nov 25 15:23:56 crc kubenswrapper[4890]: I1125 15:23:56.089360 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/96c4c8fb-37d7-4415-aef1-6da42e54fbc2-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-4fhbm\" (UID: \"96c4c8fb-37d7-4415-aef1-6da42e54fbc2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-4fhbm" Nov 25 15:23:56 crc kubenswrapper[4890]: I1125 15:23:56.089511 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/96c4c8fb-37d7-4415-aef1-6da42e54fbc2-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-4fhbm\" (UID: \"96c4c8fb-37d7-4415-aef1-6da42e54fbc2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-4fhbm" Nov 25 15:23:56 crc kubenswrapper[4890]: I1125 15:23:56.089642 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/96c4c8fb-37d7-4415-aef1-6da42e54fbc2-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-4fhbm\" (UID: \"96c4c8fb-37d7-4415-aef1-6da42e54fbc2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-4fhbm" Nov 25 15:23:56 crc kubenswrapper[4890]: I1125 15:23:56.089768 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jf74x\" (UniqueName: \"kubernetes.io/projected/96c4c8fb-37d7-4415-aef1-6da42e54fbc2-kube-api-access-jf74x\") pod \"dnsmasq-dns-89c5cd4d5-4fhbm\" (UID: \"96c4c8fb-37d7-4415-aef1-6da42e54fbc2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-4fhbm" Nov 25 15:23:56 crc kubenswrapper[4890]: I1125 15:23:56.090128 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/96c4c8fb-37d7-4415-aef1-6da42e54fbc2-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-4fhbm\" (UID: \"96c4c8fb-37d7-4415-aef1-6da42e54fbc2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-4fhbm" Nov 25 15:23:56 crc kubenswrapper[4890]: I1125 15:23:56.090565 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/96c4c8fb-37d7-4415-aef1-6da42e54fbc2-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-4fhbm\" (UID: \"96c4c8fb-37d7-4415-aef1-6da42e54fbc2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-4fhbm" Nov 25 15:23:56 crc kubenswrapper[4890]: I1125 15:23:56.090486 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/96c4c8fb-37d7-4415-aef1-6da42e54fbc2-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-4fhbm\" (UID: \"96c4c8fb-37d7-4415-aef1-6da42e54fbc2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-4fhbm" Nov 25 15:23:56 crc kubenswrapper[4890]: I1125 15:23:56.090234 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/96c4c8fb-37d7-4415-aef1-6da42e54fbc2-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-4fhbm\" (UID: \"96c4c8fb-37d7-4415-aef1-6da42e54fbc2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-4fhbm" Nov 25 15:23:56 crc kubenswrapper[4890]: I1125 15:23:56.090704 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96c4c8fb-37d7-4415-aef1-6da42e54fbc2-config\") pod \"dnsmasq-dns-89c5cd4d5-4fhbm\" (UID: \"96c4c8fb-37d7-4415-aef1-6da42e54fbc2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-4fhbm" Nov 25 15:23:56 crc kubenswrapper[4890]: I1125 15:23:56.114447 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jf74x\" (UniqueName: \"kubernetes.io/projected/96c4c8fb-37d7-4415-aef1-6da42e54fbc2-kube-api-access-jf74x\") pod \"dnsmasq-dns-89c5cd4d5-4fhbm\" (UID: \"96c4c8fb-37d7-4415-aef1-6da42e54fbc2\") " pod="openstack/dnsmasq-dns-89c5cd4d5-4fhbm" Nov 25 15:23:56 crc kubenswrapper[4890]: I1125 15:23:56.235005 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-4fhbm" Nov 25 15:23:56 crc kubenswrapper[4890]: I1125 15:23:56.449515 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 15:23:56 crc kubenswrapper[4890]: I1125 15:23:56.449844 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 15:23:56 crc kubenswrapper[4890]: I1125 15:23:56.733102 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-4fhbm"] Nov 25 15:23:57 crc kubenswrapper[4890]: I1125 15:23:57.687034 4890 generic.go:334] "Generic (PLEG): container finished" podID="96c4c8fb-37d7-4415-aef1-6da42e54fbc2" containerID="ec93458d0b0b9fe23fe22cd76764503b41d8aa51f583e59ce8d0dfede25477eb" exitCode=0 Nov 25 15:23:57 crc kubenswrapper[4890]: I1125 15:23:57.688679 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-4fhbm" event={"ID":"96c4c8fb-37d7-4415-aef1-6da42e54fbc2","Type":"ContainerDied","Data":"ec93458d0b0b9fe23fe22cd76764503b41d8aa51f583e59ce8d0dfede25477eb"} Nov 25 15:23:57 crc kubenswrapper[4890]: I1125 15:23:57.688719 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-4fhbm" event={"ID":"96c4c8fb-37d7-4415-aef1-6da42e54fbc2","Type":"ContainerStarted","Data":"17c99d83e22cd428c3c04a271604974a773fd54a98646ae74486c93321491708"} Nov 25 15:23:58 crc kubenswrapper[4890]: I1125 15:23:58.263668 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:23:58 crc kubenswrapper[4890]: I1125 15:23:58.264675 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="154b7e43-8329-4c7d-9870-ffca704e80af" containerName="ceilometer-central-agent" containerID="cri-o://3cf7a048e23b8c441dcdc9d0edd56497a526f509d3341bf8de1c9d8622d0ff53" gracePeriod=30 Nov 25 15:23:58 crc kubenswrapper[4890]: I1125 15:23:58.264753 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="154b7e43-8329-4c7d-9870-ffca704e80af" containerName="sg-core" containerID="cri-o://7db4830819acbf94ebff9eadbc3cd19d99ee3e5a7bd2990fbc67d168359b2645" gracePeriod=30 Nov 25 15:23:58 crc kubenswrapper[4890]: I1125 15:23:58.264752 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="154b7e43-8329-4c7d-9870-ffca704e80af" containerName="proxy-httpd" containerID="cri-o://aca91755e7903d75a205a2f0efa57c85839df1ef3b758ce7e6453c9e38094964" gracePeriod=30 Nov 25 15:23:58 crc kubenswrapper[4890]: I1125 15:23:58.264824 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="154b7e43-8329-4c7d-9870-ffca704e80af" containerName="ceilometer-notification-agent" containerID="cri-o://c188f25928575f0e2459d821ee704903e370c8a80d57a9a9a94092152a317754" gracePeriod=30 Nov 25 15:23:58 crc kubenswrapper[4890]: I1125 15:23:58.284712 4890 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="154b7e43-8329-4c7d-9870-ffca704e80af" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.195:3000/\": EOF" Nov 25 15:23:58 crc kubenswrapper[4890]: I1125 15:23:58.362225 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 25 15:23:58 crc kubenswrapper[4890]: I1125 15:23:58.553379 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 25 15:23:58 crc kubenswrapper[4890]: I1125 15:23:58.699690 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-4fhbm" event={"ID":"96c4c8fb-37d7-4415-aef1-6da42e54fbc2","Type":"ContainerStarted","Data":"60206e73ae858d7c0dc08e3962ac6d824a7cb1a06b0480372710c505233f663d"} Nov 25 15:23:58 crc kubenswrapper[4890]: I1125 15:23:58.699801 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-89c5cd4d5-4fhbm" Nov 25 15:23:58 crc kubenswrapper[4890]: I1125 15:23:58.703336 4890 generic.go:334] "Generic (PLEG): container finished" podID="154b7e43-8329-4c7d-9870-ffca704e80af" containerID="aca91755e7903d75a205a2f0efa57c85839df1ef3b758ce7e6453c9e38094964" exitCode=0 Nov 25 15:23:58 crc kubenswrapper[4890]: I1125 15:23:58.703371 4890 generic.go:334] "Generic (PLEG): container finished" podID="154b7e43-8329-4c7d-9870-ffca704e80af" containerID="7db4830819acbf94ebff9eadbc3cd19d99ee3e5a7bd2990fbc67d168359b2645" exitCode=2 Nov 25 15:23:58 crc kubenswrapper[4890]: I1125 15:23:58.703379 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"154b7e43-8329-4c7d-9870-ffca704e80af","Type":"ContainerDied","Data":"aca91755e7903d75a205a2f0efa57c85839df1ef3b758ce7e6453c9e38094964"} Nov 25 15:23:58 crc kubenswrapper[4890]: I1125 15:23:58.703431 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"154b7e43-8329-4c7d-9870-ffca704e80af","Type":"ContainerDied","Data":"7db4830819acbf94ebff9eadbc3cd19d99ee3e5a7bd2990fbc67d168359b2645"} Nov 25 15:23:58 crc kubenswrapper[4890]: I1125 15:23:58.703586 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="324efe97-1d1d-47d9-83c0-631a31cf4d2e" containerName="nova-api-log" containerID="cri-o://36f461bec3aa616aac770d2c2fb97fd0685688dcccb1c4ed337fdd4c08fdf413" gracePeriod=30 Nov 25 15:23:58 crc kubenswrapper[4890]: I1125 15:23:58.703651 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="324efe97-1d1d-47d9-83c0-631a31cf4d2e" containerName="nova-api-api" containerID="cri-o://c1976273a01588d0ee2baa87932d07d78389f02d3e7b6bfc38257dd07b754dbf" gracePeriod=30 Nov 25 15:23:58 crc kubenswrapper[4890]: I1125 15:23:58.750251 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-89c5cd4d5-4fhbm" podStartSLOduration=3.750222268 podStartE2EDuration="3.750222268s" podCreationTimestamp="2025-11-25 15:23:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:23:58.737471848 +0000 UTC m=+1297.179934468" watchObservedRunningTime="2025-11-25 15:23:58.750222268 +0000 UTC m=+1297.192684878" Nov 25 15:23:59 crc kubenswrapper[4890]: I1125 15:23:59.725372 4890 generic.go:334] "Generic (PLEG): container finished" podID="154b7e43-8329-4c7d-9870-ffca704e80af" containerID="c188f25928575f0e2459d821ee704903e370c8a80d57a9a9a94092152a317754" exitCode=0 Nov 25 15:23:59 crc kubenswrapper[4890]: I1125 15:23:59.725734 4890 generic.go:334] "Generic (PLEG): container finished" podID="154b7e43-8329-4c7d-9870-ffca704e80af" containerID="3cf7a048e23b8c441dcdc9d0edd56497a526f509d3341bf8de1c9d8622d0ff53" exitCode=0 Nov 25 15:23:59 crc kubenswrapper[4890]: I1125 15:23:59.725474 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"154b7e43-8329-4c7d-9870-ffca704e80af","Type":"ContainerDied","Data":"c188f25928575f0e2459d821ee704903e370c8a80d57a9a9a94092152a317754"} Nov 25 15:23:59 crc kubenswrapper[4890]: I1125 15:23:59.725826 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"154b7e43-8329-4c7d-9870-ffca704e80af","Type":"ContainerDied","Data":"3cf7a048e23b8c441dcdc9d0edd56497a526f509d3341bf8de1c9d8622d0ff53"} Nov 25 15:23:59 crc kubenswrapper[4890]: I1125 15:23:59.730523 4890 generic.go:334] "Generic (PLEG): container finished" podID="324efe97-1d1d-47d9-83c0-631a31cf4d2e" containerID="36f461bec3aa616aac770d2c2fb97fd0685688dcccb1c4ed337fdd4c08fdf413" exitCode=143 Nov 25 15:23:59 crc kubenswrapper[4890]: I1125 15:23:59.731613 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"324efe97-1d1d-47d9-83c0-631a31cf4d2e","Type":"ContainerDied","Data":"36f461bec3aa616aac770d2c2fb97fd0685688dcccb1c4ed337fdd4c08fdf413"} Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.066630 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.175996 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/154b7e43-8329-4c7d-9870-ffca704e80af-log-httpd\") pod \"154b7e43-8329-4c7d-9870-ffca704e80af\" (UID: \"154b7e43-8329-4c7d-9870-ffca704e80af\") " Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.176066 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/154b7e43-8329-4c7d-9870-ffca704e80af-run-httpd\") pod \"154b7e43-8329-4c7d-9870-ffca704e80af\" (UID: \"154b7e43-8329-4c7d-9870-ffca704e80af\") " Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.176134 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/154b7e43-8329-4c7d-9870-ffca704e80af-scripts\") pod \"154b7e43-8329-4c7d-9870-ffca704e80af\" (UID: \"154b7e43-8329-4c7d-9870-ffca704e80af\") " Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.176190 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/154b7e43-8329-4c7d-9870-ffca704e80af-ceilometer-tls-certs\") pod \"154b7e43-8329-4c7d-9870-ffca704e80af\" (UID: \"154b7e43-8329-4c7d-9870-ffca704e80af\") " Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.176231 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/154b7e43-8329-4c7d-9870-ffca704e80af-combined-ca-bundle\") pod \"154b7e43-8329-4c7d-9870-ffca704e80af\" (UID: \"154b7e43-8329-4c7d-9870-ffca704e80af\") " Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.176397 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/154b7e43-8329-4c7d-9870-ffca704e80af-sg-core-conf-yaml\") pod \"154b7e43-8329-4c7d-9870-ffca704e80af\" (UID: \"154b7e43-8329-4c7d-9870-ffca704e80af\") " Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.176437 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t4zxj\" (UniqueName: \"kubernetes.io/projected/154b7e43-8329-4c7d-9870-ffca704e80af-kube-api-access-t4zxj\") pod \"154b7e43-8329-4c7d-9870-ffca704e80af\" (UID: \"154b7e43-8329-4c7d-9870-ffca704e80af\") " Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.176469 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/154b7e43-8329-4c7d-9870-ffca704e80af-config-data\") pod \"154b7e43-8329-4c7d-9870-ffca704e80af\" (UID: \"154b7e43-8329-4c7d-9870-ffca704e80af\") " Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.180080 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/154b7e43-8329-4c7d-9870-ffca704e80af-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "154b7e43-8329-4c7d-9870-ffca704e80af" (UID: "154b7e43-8329-4c7d-9870-ffca704e80af"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.181364 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/154b7e43-8329-4c7d-9870-ffca704e80af-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "154b7e43-8329-4c7d-9870-ffca704e80af" (UID: "154b7e43-8329-4c7d-9870-ffca704e80af"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.187398 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/154b7e43-8329-4c7d-9870-ffca704e80af-kube-api-access-t4zxj" (OuterVolumeSpecName: "kube-api-access-t4zxj") pod "154b7e43-8329-4c7d-9870-ffca704e80af" (UID: "154b7e43-8329-4c7d-9870-ffca704e80af"). InnerVolumeSpecName "kube-api-access-t4zxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.193421 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/154b7e43-8329-4c7d-9870-ffca704e80af-scripts" (OuterVolumeSpecName: "scripts") pod "154b7e43-8329-4c7d-9870-ffca704e80af" (UID: "154b7e43-8329-4c7d-9870-ffca704e80af"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.219411 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/154b7e43-8329-4c7d-9870-ffca704e80af-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "154b7e43-8329-4c7d-9870-ffca704e80af" (UID: "154b7e43-8329-4c7d-9870-ffca704e80af"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.269996 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/154b7e43-8329-4c7d-9870-ffca704e80af-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "154b7e43-8329-4c7d-9870-ffca704e80af" (UID: "154b7e43-8329-4c7d-9870-ffca704e80af"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.279894 4890 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/154b7e43-8329-4c7d-9870-ffca704e80af-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.279927 4890 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/154b7e43-8329-4c7d-9870-ffca704e80af-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.279936 4890 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/154b7e43-8329-4c7d-9870-ffca704e80af-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.279946 4890 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/154b7e43-8329-4c7d-9870-ffca704e80af-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.279956 4890 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/154b7e43-8329-4c7d-9870-ffca704e80af-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.279965 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t4zxj\" (UniqueName: \"kubernetes.io/projected/154b7e43-8329-4c7d-9870-ffca704e80af-kube-api-access-t4zxj\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.285780 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/154b7e43-8329-4c7d-9870-ffca704e80af-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "154b7e43-8329-4c7d-9870-ffca704e80af" (UID: "154b7e43-8329-4c7d-9870-ffca704e80af"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.312789 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/154b7e43-8329-4c7d-9870-ffca704e80af-config-data" (OuterVolumeSpecName: "config-data") pod "154b7e43-8329-4c7d-9870-ffca704e80af" (UID: "154b7e43-8329-4c7d-9870-ffca704e80af"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.381991 4890 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/154b7e43-8329-4c7d-9870-ffca704e80af-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.382732 4890 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/154b7e43-8329-4c7d-9870-ffca704e80af-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.742134 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"154b7e43-8329-4c7d-9870-ffca704e80af","Type":"ContainerDied","Data":"2578edede8e4da833402c6ce7f04a159508d2c227f17deb95588e255413b726e"} Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.742227 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.742265 4890 scope.go:117] "RemoveContainer" containerID="aca91755e7903d75a205a2f0efa57c85839df1ef3b758ce7e6453c9e38094964" Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.768964 4890 scope.go:117] "RemoveContainer" containerID="7db4830819acbf94ebff9eadbc3cd19d99ee3e5a7bd2990fbc67d168359b2645" Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.789428 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.808202 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.819882 4890 scope.go:117] "RemoveContainer" containerID="c188f25928575f0e2459d821ee704903e370c8a80d57a9a9a94092152a317754" Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.826091 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:24:00 crc kubenswrapper[4890]: E1125 15:24:00.827095 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="154b7e43-8329-4c7d-9870-ffca704e80af" containerName="ceilometer-notification-agent" Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.827118 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="154b7e43-8329-4c7d-9870-ffca704e80af" containerName="ceilometer-notification-agent" Nov 25 15:24:00 crc kubenswrapper[4890]: E1125 15:24:00.827138 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="154b7e43-8329-4c7d-9870-ffca704e80af" containerName="ceilometer-central-agent" Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.827146 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="154b7e43-8329-4c7d-9870-ffca704e80af" containerName="ceilometer-central-agent" Nov 25 15:24:00 crc kubenswrapper[4890]: E1125 15:24:00.827171 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="154b7e43-8329-4c7d-9870-ffca704e80af" containerName="proxy-httpd" Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.827178 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="154b7e43-8329-4c7d-9870-ffca704e80af" containerName="proxy-httpd" Nov 25 15:24:00 crc kubenswrapper[4890]: E1125 15:24:00.827192 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="154b7e43-8329-4c7d-9870-ffca704e80af" containerName="sg-core" Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.827199 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="154b7e43-8329-4c7d-9870-ffca704e80af" containerName="sg-core" Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.827413 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="154b7e43-8329-4c7d-9870-ffca704e80af" containerName="ceilometer-central-agent" Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.827426 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="154b7e43-8329-4c7d-9870-ffca704e80af" containerName="proxy-httpd" Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.827438 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="154b7e43-8329-4c7d-9870-ffca704e80af" containerName="sg-core" Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.827453 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="154b7e43-8329-4c7d-9870-ffca704e80af" containerName="ceilometer-notification-agent" Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.829569 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.832584 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.832768 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.833472 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.845435 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.857371 4890 scope.go:117] "RemoveContainer" containerID="3cf7a048e23b8c441dcdc9d0edd56497a526f509d3341bf8de1c9d8622d0ff53" Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.994365 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f27f06e7-28d2-403a-92c8-b1d858d90619-config-data\") pod \"ceilometer-0\" (UID: \"f27f06e7-28d2-403a-92c8-b1d858d90619\") " pod="openstack/ceilometer-0" Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.994440 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f27f06e7-28d2-403a-92c8-b1d858d90619-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f27f06e7-28d2-403a-92c8-b1d858d90619\") " pod="openstack/ceilometer-0" Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.994466 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f27f06e7-28d2-403a-92c8-b1d858d90619-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f27f06e7-28d2-403a-92c8-b1d858d90619\") " pod="openstack/ceilometer-0" Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.994512 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f27f06e7-28d2-403a-92c8-b1d858d90619-run-httpd\") pod \"ceilometer-0\" (UID: \"f27f06e7-28d2-403a-92c8-b1d858d90619\") " pod="openstack/ceilometer-0" Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.994552 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrhkf\" (UniqueName: \"kubernetes.io/projected/f27f06e7-28d2-403a-92c8-b1d858d90619-kube-api-access-xrhkf\") pod \"ceilometer-0\" (UID: \"f27f06e7-28d2-403a-92c8-b1d858d90619\") " pod="openstack/ceilometer-0" Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.994582 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f27f06e7-28d2-403a-92c8-b1d858d90619-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f27f06e7-28d2-403a-92c8-b1d858d90619\") " pod="openstack/ceilometer-0" Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.994651 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f27f06e7-28d2-403a-92c8-b1d858d90619-log-httpd\") pod \"ceilometer-0\" (UID: \"f27f06e7-28d2-403a-92c8-b1d858d90619\") " pod="openstack/ceilometer-0" Nov 25 15:24:00 crc kubenswrapper[4890]: I1125 15:24:00.994679 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f27f06e7-28d2-403a-92c8-b1d858d90619-scripts\") pod \"ceilometer-0\" (UID: \"f27f06e7-28d2-403a-92c8-b1d858d90619\") " pod="openstack/ceilometer-0" Nov 25 15:24:01 crc kubenswrapper[4890]: I1125 15:24:01.096376 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f27f06e7-28d2-403a-92c8-b1d858d90619-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f27f06e7-28d2-403a-92c8-b1d858d90619\") " pod="openstack/ceilometer-0" Nov 25 15:24:01 crc kubenswrapper[4890]: I1125 15:24:01.096430 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f27f06e7-28d2-403a-92c8-b1d858d90619-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f27f06e7-28d2-403a-92c8-b1d858d90619\") " pod="openstack/ceilometer-0" Nov 25 15:24:01 crc kubenswrapper[4890]: I1125 15:24:01.096498 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f27f06e7-28d2-403a-92c8-b1d858d90619-run-httpd\") pod \"ceilometer-0\" (UID: \"f27f06e7-28d2-403a-92c8-b1d858d90619\") " pod="openstack/ceilometer-0" Nov 25 15:24:01 crc kubenswrapper[4890]: I1125 15:24:01.096555 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrhkf\" (UniqueName: \"kubernetes.io/projected/f27f06e7-28d2-403a-92c8-b1d858d90619-kube-api-access-xrhkf\") pod \"ceilometer-0\" (UID: \"f27f06e7-28d2-403a-92c8-b1d858d90619\") " pod="openstack/ceilometer-0" Nov 25 15:24:01 crc kubenswrapper[4890]: I1125 15:24:01.096591 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f27f06e7-28d2-403a-92c8-b1d858d90619-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f27f06e7-28d2-403a-92c8-b1d858d90619\") " pod="openstack/ceilometer-0" Nov 25 15:24:01 crc kubenswrapper[4890]: I1125 15:24:01.096668 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f27f06e7-28d2-403a-92c8-b1d858d90619-log-httpd\") pod \"ceilometer-0\" (UID: \"f27f06e7-28d2-403a-92c8-b1d858d90619\") " pod="openstack/ceilometer-0" Nov 25 15:24:01 crc kubenswrapper[4890]: I1125 15:24:01.096698 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f27f06e7-28d2-403a-92c8-b1d858d90619-scripts\") pod \"ceilometer-0\" (UID: \"f27f06e7-28d2-403a-92c8-b1d858d90619\") " pod="openstack/ceilometer-0" Nov 25 15:24:01 crc kubenswrapper[4890]: I1125 15:24:01.096733 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f27f06e7-28d2-403a-92c8-b1d858d90619-config-data\") pod \"ceilometer-0\" (UID: \"f27f06e7-28d2-403a-92c8-b1d858d90619\") " pod="openstack/ceilometer-0" Nov 25 15:24:01 crc kubenswrapper[4890]: I1125 15:24:01.097625 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f27f06e7-28d2-403a-92c8-b1d858d90619-run-httpd\") pod \"ceilometer-0\" (UID: \"f27f06e7-28d2-403a-92c8-b1d858d90619\") " pod="openstack/ceilometer-0" Nov 25 15:24:01 crc kubenswrapper[4890]: I1125 15:24:01.097765 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f27f06e7-28d2-403a-92c8-b1d858d90619-log-httpd\") pod \"ceilometer-0\" (UID: \"f27f06e7-28d2-403a-92c8-b1d858d90619\") " pod="openstack/ceilometer-0" Nov 25 15:24:01 crc kubenswrapper[4890]: I1125 15:24:01.103050 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f27f06e7-28d2-403a-92c8-b1d858d90619-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f27f06e7-28d2-403a-92c8-b1d858d90619\") " pod="openstack/ceilometer-0" Nov 25 15:24:01 crc kubenswrapper[4890]: I1125 15:24:01.103081 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f27f06e7-28d2-403a-92c8-b1d858d90619-scripts\") pod \"ceilometer-0\" (UID: \"f27f06e7-28d2-403a-92c8-b1d858d90619\") " pod="openstack/ceilometer-0" Nov 25 15:24:01 crc kubenswrapper[4890]: I1125 15:24:01.105874 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f27f06e7-28d2-403a-92c8-b1d858d90619-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f27f06e7-28d2-403a-92c8-b1d858d90619\") " pod="openstack/ceilometer-0" Nov 25 15:24:01 crc kubenswrapper[4890]: I1125 15:24:01.107741 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f27f06e7-28d2-403a-92c8-b1d858d90619-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f27f06e7-28d2-403a-92c8-b1d858d90619\") " pod="openstack/ceilometer-0" Nov 25 15:24:01 crc kubenswrapper[4890]: I1125 15:24:01.116720 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrhkf\" (UniqueName: \"kubernetes.io/projected/f27f06e7-28d2-403a-92c8-b1d858d90619-kube-api-access-xrhkf\") pod \"ceilometer-0\" (UID: \"f27f06e7-28d2-403a-92c8-b1d858d90619\") " pod="openstack/ceilometer-0" Nov 25 15:24:01 crc kubenswrapper[4890]: I1125 15:24:01.117224 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f27f06e7-28d2-403a-92c8-b1d858d90619-config-data\") pod \"ceilometer-0\" (UID: \"f27f06e7-28d2-403a-92c8-b1d858d90619\") " pod="openstack/ceilometer-0" Nov 25 15:24:01 crc kubenswrapper[4890]: I1125 15:24:01.157672 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 15:24:01 crc kubenswrapper[4890]: I1125 15:24:01.620590 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 15:24:01 crc kubenswrapper[4890]: I1125 15:24:01.751372 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f27f06e7-28d2-403a-92c8-b1d858d90619","Type":"ContainerStarted","Data":"53ae920c527bd08fabbb57d6325282f6d8e46d5e28b9b6fec390cd31910fb6c2"} Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.187686 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="154b7e43-8329-4c7d-9870-ffca704e80af" path="/var/lib/kubelet/pods/154b7e43-8329-4c7d-9870-ffca704e80af/volumes" Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.336145 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.451154 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/324efe97-1d1d-47d9-83c0-631a31cf4d2e-config-data\") pod \"324efe97-1d1d-47d9-83c0-631a31cf4d2e\" (UID: \"324efe97-1d1d-47d9-83c0-631a31cf4d2e\") " Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.451533 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lr5vm\" (UniqueName: \"kubernetes.io/projected/324efe97-1d1d-47d9-83c0-631a31cf4d2e-kube-api-access-lr5vm\") pod \"324efe97-1d1d-47d9-83c0-631a31cf4d2e\" (UID: \"324efe97-1d1d-47d9-83c0-631a31cf4d2e\") " Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.451618 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/324efe97-1d1d-47d9-83c0-631a31cf4d2e-logs\") pod \"324efe97-1d1d-47d9-83c0-631a31cf4d2e\" (UID: \"324efe97-1d1d-47d9-83c0-631a31cf4d2e\") " Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.451724 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/324efe97-1d1d-47d9-83c0-631a31cf4d2e-combined-ca-bundle\") pod \"324efe97-1d1d-47d9-83c0-631a31cf4d2e\" (UID: \"324efe97-1d1d-47d9-83c0-631a31cf4d2e\") " Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.452730 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/324efe97-1d1d-47d9-83c0-631a31cf4d2e-logs" (OuterVolumeSpecName: "logs") pod "324efe97-1d1d-47d9-83c0-631a31cf4d2e" (UID: "324efe97-1d1d-47d9-83c0-631a31cf4d2e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.462337 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/324efe97-1d1d-47d9-83c0-631a31cf4d2e-kube-api-access-lr5vm" (OuterVolumeSpecName: "kube-api-access-lr5vm") pod "324efe97-1d1d-47d9-83c0-631a31cf4d2e" (UID: "324efe97-1d1d-47d9-83c0-631a31cf4d2e"). InnerVolumeSpecName "kube-api-access-lr5vm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.496722 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/324efe97-1d1d-47d9-83c0-631a31cf4d2e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "324efe97-1d1d-47d9-83c0-631a31cf4d2e" (UID: "324efe97-1d1d-47d9-83c0-631a31cf4d2e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.515647 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/324efe97-1d1d-47d9-83c0-631a31cf4d2e-config-data" (OuterVolumeSpecName: "config-data") pod "324efe97-1d1d-47d9-83c0-631a31cf4d2e" (UID: "324efe97-1d1d-47d9-83c0-631a31cf4d2e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.555012 4890 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/324efe97-1d1d-47d9-83c0-631a31cf4d2e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.555044 4890 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/324efe97-1d1d-47d9-83c0-631a31cf4d2e-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.555066 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lr5vm\" (UniqueName: \"kubernetes.io/projected/324efe97-1d1d-47d9-83c0-631a31cf4d2e-kube-api-access-lr5vm\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.555078 4890 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/324efe97-1d1d-47d9-83c0-631a31cf4d2e-logs\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.763003 4890 generic.go:334] "Generic (PLEG): container finished" podID="324efe97-1d1d-47d9-83c0-631a31cf4d2e" containerID="c1976273a01588d0ee2baa87932d07d78389f02d3e7b6bfc38257dd07b754dbf" exitCode=0 Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.763041 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"324efe97-1d1d-47d9-83c0-631a31cf4d2e","Type":"ContainerDied","Data":"c1976273a01588d0ee2baa87932d07d78389f02d3e7b6bfc38257dd07b754dbf"} Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.763065 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"324efe97-1d1d-47d9-83c0-631a31cf4d2e","Type":"ContainerDied","Data":"95489b8853f2c989cce2cf433d410053adb75580ff8c34e3e992c8c76110eddb"} Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.763079 4890 scope.go:117] "RemoveContainer" containerID="c1976273a01588d0ee2baa87932d07d78389f02d3e7b6bfc38257dd07b754dbf" Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.763278 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.796496 4890 scope.go:117] "RemoveContainer" containerID="36f461bec3aa616aac770d2c2fb97fd0685688dcccb1c4ed337fdd4c08fdf413" Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.809304 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.822547 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.834581 4890 scope.go:117] "RemoveContainer" containerID="c1976273a01588d0ee2baa87932d07d78389f02d3e7b6bfc38257dd07b754dbf" Nov 25 15:24:02 crc kubenswrapper[4890]: E1125 15:24:02.835186 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1976273a01588d0ee2baa87932d07d78389f02d3e7b6bfc38257dd07b754dbf\": container with ID starting with c1976273a01588d0ee2baa87932d07d78389f02d3e7b6bfc38257dd07b754dbf not found: ID does not exist" containerID="c1976273a01588d0ee2baa87932d07d78389f02d3e7b6bfc38257dd07b754dbf" Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.835217 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1976273a01588d0ee2baa87932d07d78389f02d3e7b6bfc38257dd07b754dbf"} err="failed to get container status \"c1976273a01588d0ee2baa87932d07d78389f02d3e7b6bfc38257dd07b754dbf\": rpc error: code = NotFound desc = could not find container \"c1976273a01588d0ee2baa87932d07d78389f02d3e7b6bfc38257dd07b754dbf\": container with ID starting with c1976273a01588d0ee2baa87932d07d78389f02d3e7b6bfc38257dd07b754dbf not found: ID does not exist" Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.835238 4890 scope.go:117] "RemoveContainer" containerID="36f461bec3aa616aac770d2c2fb97fd0685688dcccb1c4ed337fdd4c08fdf413" Nov 25 15:24:02 crc kubenswrapper[4890]: E1125 15:24:02.835741 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36f461bec3aa616aac770d2c2fb97fd0685688dcccb1c4ed337fdd4c08fdf413\": container with ID starting with 36f461bec3aa616aac770d2c2fb97fd0685688dcccb1c4ed337fdd4c08fdf413 not found: ID does not exist" containerID="36f461bec3aa616aac770d2c2fb97fd0685688dcccb1c4ed337fdd4c08fdf413" Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.835765 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36f461bec3aa616aac770d2c2fb97fd0685688dcccb1c4ed337fdd4c08fdf413"} err="failed to get container status \"36f461bec3aa616aac770d2c2fb97fd0685688dcccb1c4ed337fdd4c08fdf413\": rpc error: code = NotFound desc = could not find container \"36f461bec3aa616aac770d2c2fb97fd0685688dcccb1c4ed337fdd4c08fdf413\": container with ID starting with 36f461bec3aa616aac770d2c2fb97fd0685688dcccb1c4ed337fdd4c08fdf413 not found: ID does not exist" Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.836357 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 25 15:24:02 crc kubenswrapper[4890]: E1125 15:24:02.836793 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="324efe97-1d1d-47d9-83c0-631a31cf4d2e" containerName="nova-api-api" Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.836806 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="324efe97-1d1d-47d9-83c0-631a31cf4d2e" containerName="nova-api-api" Nov 25 15:24:02 crc kubenswrapper[4890]: E1125 15:24:02.836831 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="324efe97-1d1d-47d9-83c0-631a31cf4d2e" containerName="nova-api-log" Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.836839 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="324efe97-1d1d-47d9-83c0-631a31cf4d2e" containerName="nova-api-log" Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.836996 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="324efe97-1d1d-47d9-83c0-631a31cf4d2e" containerName="nova-api-log" Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.837013 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="324efe97-1d1d-47d9-83c0-631a31cf4d2e" containerName="nova-api-api" Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.837992 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.846046 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.856791 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.856969 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.857127 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.870182 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d89315fb-6d71-4209-a3ea-a46d6a8e7691-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d89315fb-6d71-4209-a3ea-a46d6a8e7691\") " pod="openstack/nova-api-0" Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.872394 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d89315fb-6d71-4209-a3ea-a46d6a8e7691-internal-tls-certs\") pod \"nova-api-0\" (UID: \"d89315fb-6d71-4209-a3ea-a46d6a8e7691\") " pod="openstack/nova-api-0" Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.872626 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d89315fb-6d71-4209-a3ea-a46d6a8e7691-logs\") pod \"nova-api-0\" (UID: \"d89315fb-6d71-4209-a3ea-a46d6a8e7691\") " pod="openstack/nova-api-0" Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.872900 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wr2ts\" (UniqueName: \"kubernetes.io/projected/d89315fb-6d71-4209-a3ea-a46d6a8e7691-kube-api-access-wr2ts\") pod \"nova-api-0\" (UID: \"d89315fb-6d71-4209-a3ea-a46d6a8e7691\") " pod="openstack/nova-api-0" Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.873051 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d89315fb-6d71-4209-a3ea-a46d6a8e7691-public-tls-certs\") pod \"nova-api-0\" (UID: \"d89315fb-6d71-4209-a3ea-a46d6a8e7691\") " pod="openstack/nova-api-0" Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.873154 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d89315fb-6d71-4209-a3ea-a46d6a8e7691-config-data\") pod \"nova-api-0\" (UID: \"d89315fb-6d71-4209-a3ea-a46d6a8e7691\") " pod="openstack/nova-api-0" Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.974718 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d89315fb-6d71-4209-a3ea-a46d6a8e7691-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d89315fb-6d71-4209-a3ea-a46d6a8e7691\") " pod="openstack/nova-api-0" Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.974807 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d89315fb-6d71-4209-a3ea-a46d6a8e7691-internal-tls-certs\") pod \"nova-api-0\" (UID: \"d89315fb-6d71-4209-a3ea-a46d6a8e7691\") " pod="openstack/nova-api-0" Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.974838 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d89315fb-6d71-4209-a3ea-a46d6a8e7691-logs\") pod \"nova-api-0\" (UID: \"d89315fb-6d71-4209-a3ea-a46d6a8e7691\") " pod="openstack/nova-api-0" Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.974913 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wr2ts\" (UniqueName: \"kubernetes.io/projected/d89315fb-6d71-4209-a3ea-a46d6a8e7691-kube-api-access-wr2ts\") pod \"nova-api-0\" (UID: \"d89315fb-6d71-4209-a3ea-a46d6a8e7691\") " pod="openstack/nova-api-0" Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.974952 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d89315fb-6d71-4209-a3ea-a46d6a8e7691-public-tls-certs\") pod \"nova-api-0\" (UID: \"d89315fb-6d71-4209-a3ea-a46d6a8e7691\") " pod="openstack/nova-api-0" Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.974969 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d89315fb-6d71-4209-a3ea-a46d6a8e7691-config-data\") pod \"nova-api-0\" (UID: \"d89315fb-6d71-4209-a3ea-a46d6a8e7691\") " pod="openstack/nova-api-0" Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.975858 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d89315fb-6d71-4209-a3ea-a46d6a8e7691-logs\") pod \"nova-api-0\" (UID: \"d89315fb-6d71-4209-a3ea-a46d6a8e7691\") " pod="openstack/nova-api-0" Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.982650 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d89315fb-6d71-4209-a3ea-a46d6a8e7691-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d89315fb-6d71-4209-a3ea-a46d6a8e7691\") " pod="openstack/nova-api-0" Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.982669 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d89315fb-6d71-4209-a3ea-a46d6a8e7691-config-data\") pod \"nova-api-0\" (UID: \"d89315fb-6d71-4209-a3ea-a46d6a8e7691\") " pod="openstack/nova-api-0" Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.984523 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d89315fb-6d71-4209-a3ea-a46d6a8e7691-public-tls-certs\") pod \"nova-api-0\" (UID: \"d89315fb-6d71-4209-a3ea-a46d6a8e7691\") " pod="openstack/nova-api-0" Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.992712 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d89315fb-6d71-4209-a3ea-a46d6a8e7691-internal-tls-certs\") pod \"nova-api-0\" (UID: \"d89315fb-6d71-4209-a3ea-a46d6a8e7691\") " pod="openstack/nova-api-0" Nov 25 15:24:02 crc kubenswrapper[4890]: I1125 15:24:02.997219 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wr2ts\" (UniqueName: \"kubernetes.io/projected/d89315fb-6d71-4209-a3ea-a46d6a8e7691-kube-api-access-wr2ts\") pod \"nova-api-0\" (UID: \"d89315fb-6d71-4209-a3ea-a46d6a8e7691\") " pod="openstack/nova-api-0" Nov 25 15:24:03 crc kubenswrapper[4890]: I1125 15:24:03.178493 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 15:24:03 crc kubenswrapper[4890]: I1125 15:24:03.362149 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Nov 25 15:24:03 crc kubenswrapper[4890]: I1125 15:24:03.396344 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Nov 25 15:24:03 crc kubenswrapper[4890]: I1125 15:24:03.690774 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 15:24:03 crc kubenswrapper[4890]: I1125 15:24:03.773927 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d89315fb-6d71-4209-a3ea-a46d6a8e7691","Type":"ContainerStarted","Data":"67ca661595e44ff01859de973a82d8adee6c9305980a409f14c9739c5caa6371"} Nov 25 15:24:03 crc kubenswrapper[4890]: I1125 15:24:03.777880 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f27f06e7-28d2-403a-92c8-b1d858d90619","Type":"ContainerStarted","Data":"c77d2521310f2e0f2f21c84b482c5c9e5bd11d4d35cbb493d86819cae9fa5cf1"} Nov 25 15:24:03 crc kubenswrapper[4890]: I1125 15:24:03.777927 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f27f06e7-28d2-403a-92c8-b1d858d90619","Type":"ContainerStarted","Data":"de669cc3cedb45b92231c0eb4147d1acc0dc4701c7b755baebc0d5ef42665bb1"} Nov 25 15:24:03 crc kubenswrapper[4890]: I1125 15:24:03.796393 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Nov 25 15:24:04 crc kubenswrapper[4890]: I1125 15:24:04.014538 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-s2qx4"] Nov 25 15:24:04 crc kubenswrapper[4890]: I1125 15:24:04.016031 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-s2qx4" Nov 25 15:24:04 crc kubenswrapper[4890]: I1125 15:24:04.021462 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Nov 25 15:24:04 crc kubenswrapper[4890]: I1125 15:24:04.021657 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Nov 25 15:24:04 crc kubenswrapper[4890]: I1125 15:24:04.041449 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-s2qx4"] Nov 25 15:24:04 crc kubenswrapper[4890]: I1125 15:24:04.098512 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/519d05c4-07a9-4749-a8ae-072cdd3fd0af-config-data\") pod \"nova-cell1-cell-mapping-s2qx4\" (UID: \"519d05c4-07a9-4749-a8ae-072cdd3fd0af\") " pod="openstack/nova-cell1-cell-mapping-s2qx4" Nov 25 15:24:04 crc kubenswrapper[4890]: I1125 15:24:04.098566 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2q54\" (UniqueName: \"kubernetes.io/projected/519d05c4-07a9-4749-a8ae-072cdd3fd0af-kube-api-access-j2q54\") pod \"nova-cell1-cell-mapping-s2qx4\" (UID: \"519d05c4-07a9-4749-a8ae-072cdd3fd0af\") " pod="openstack/nova-cell1-cell-mapping-s2qx4" Nov 25 15:24:04 crc kubenswrapper[4890]: I1125 15:24:04.098670 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/519d05c4-07a9-4749-a8ae-072cdd3fd0af-scripts\") pod \"nova-cell1-cell-mapping-s2qx4\" (UID: \"519d05c4-07a9-4749-a8ae-072cdd3fd0af\") " pod="openstack/nova-cell1-cell-mapping-s2qx4" Nov 25 15:24:04 crc kubenswrapper[4890]: I1125 15:24:04.098797 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/519d05c4-07a9-4749-a8ae-072cdd3fd0af-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-s2qx4\" (UID: \"519d05c4-07a9-4749-a8ae-072cdd3fd0af\") " pod="openstack/nova-cell1-cell-mapping-s2qx4" Nov 25 15:24:04 crc kubenswrapper[4890]: I1125 15:24:04.200298 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/519d05c4-07a9-4749-a8ae-072cdd3fd0af-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-s2qx4\" (UID: \"519d05c4-07a9-4749-a8ae-072cdd3fd0af\") " pod="openstack/nova-cell1-cell-mapping-s2qx4" Nov 25 15:24:04 crc kubenswrapper[4890]: I1125 15:24:04.200410 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/519d05c4-07a9-4749-a8ae-072cdd3fd0af-config-data\") pod \"nova-cell1-cell-mapping-s2qx4\" (UID: \"519d05c4-07a9-4749-a8ae-072cdd3fd0af\") " pod="openstack/nova-cell1-cell-mapping-s2qx4" Nov 25 15:24:04 crc kubenswrapper[4890]: I1125 15:24:04.200444 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2q54\" (UniqueName: \"kubernetes.io/projected/519d05c4-07a9-4749-a8ae-072cdd3fd0af-kube-api-access-j2q54\") pod \"nova-cell1-cell-mapping-s2qx4\" (UID: \"519d05c4-07a9-4749-a8ae-072cdd3fd0af\") " pod="openstack/nova-cell1-cell-mapping-s2qx4" Nov 25 15:24:04 crc kubenswrapper[4890]: I1125 15:24:04.200509 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/519d05c4-07a9-4749-a8ae-072cdd3fd0af-scripts\") pod \"nova-cell1-cell-mapping-s2qx4\" (UID: \"519d05c4-07a9-4749-a8ae-072cdd3fd0af\") " pod="openstack/nova-cell1-cell-mapping-s2qx4" Nov 25 15:24:04 crc kubenswrapper[4890]: I1125 15:24:04.201436 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="324efe97-1d1d-47d9-83c0-631a31cf4d2e" path="/var/lib/kubelet/pods/324efe97-1d1d-47d9-83c0-631a31cf4d2e/volumes" Nov 25 15:24:04 crc kubenswrapper[4890]: I1125 15:24:04.205753 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/519d05c4-07a9-4749-a8ae-072cdd3fd0af-scripts\") pod \"nova-cell1-cell-mapping-s2qx4\" (UID: \"519d05c4-07a9-4749-a8ae-072cdd3fd0af\") " pod="openstack/nova-cell1-cell-mapping-s2qx4" Nov 25 15:24:04 crc kubenswrapper[4890]: I1125 15:24:04.206228 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/519d05c4-07a9-4749-a8ae-072cdd3fd0af-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-s2qx4\" (UID: \"519d05c4-07a9-4749-a8ae-072cdd3fd0af\") " pod="openstack/nova-cell1-cell-mapping-s2qx4" Nov 25 15:24:04 crc kubenswrapper[4890]: I1125 15:24:04.206535 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/519d05c4-07a9-4749-a8ae-072cdd3fd0af-config-data\") pod \"nova-cell1-cell-mapping-s2qx4\" (UID: \"519d05c4-07a9-4749-a8ae-072cdd3fd0af\") " pod="openstack/nova-cell1-cell-mapping-s2qx4" Nov 25 15:24:04 crc kubenswrapper[4890]: I1125 15:24:04.228755 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2q54\" (UniqueName: \"kubernetes.io/projected/519d05c4-07a9-4749-a8ae-072cdd3fd0af-kube-api-access-j2q54\") pod \"nova-cell1-cell-mapping-s2qx4\" (UID: \"519d05c4-07a9-4749-a8ae-072cdd3fd0af\") " pod="openstack/nova-cell1-cell-mapping-s2qx4" Nov 25 15:24:04 crc kubenswrapper[4890]: I1125 15:24:04.415804 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-s2qx4" Nov 25 15:24:04 crc kubenswrapper[4890]: I1125 15:24:04.798464 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d89315fb-6d71-4209-a3ea-a46d6a8e7691","Type":"ContainerStarted","Data":"ed7b44eec03eedda56084acc550e69c1560a3b14ae04652eb8dfc16d1db9a592"} Nov 25 15:24:04 crc kubenswrapper[4890]: I1125 15:24:04.799149 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d89315fb-6d71-4209-a3ea-a46d6a8e7691","Type":"ContainerStarted","Data":"39185f3ea75420ac3f496d66c9c18f68adfd5be09cbd2ef739cce24171d2ea53"} Nov 25 15:24:04 crc kubenswrapper[4890]: I1125 15:24:04.803326 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f27f06e7-28d2-403a-92c8-b1d858d90619","Type":"ContainerStarted","Data":"b2db324d11a730211b640292456230e7ba88eb48b8de7d994fe58fe6fc478885"} Nov 25 15:24:04 crc kubenswrapper[4890]: I1125 15:24:04.833075 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.833055577 podStartE2EDuration="2.833055577s" podCreationTimestamp="2025-11-25 15:24:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:24:04.822641135 +0000 UTC m=+1303.265103745" watchObservedRunningTime="2025-11-25 15:24:04.833055577 +0000 UTC m=+1303.275518187" Nov 25 15:24:04 crc kubenswrapper[4890]: I1125 15:24:04.941953 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-s2qx4"] Nov 25 15:24:05 crc kubenswrapper[4890]: I1125 15:24:05.814416 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-s2qx4" event={"ID":"519d05c4-07a9-4749-a8ae-072cdd3fd0af","Type":"ContainerStarted","Data":"1557774a1642615323744cb5147878a67bf6499e0cacf33a93edbb5e592537bc"} Nov 25 15:24:05 crc kubenswrapper[4890]: I1125 15:24:05.814974 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-s2qx4" event={"ID":"519d05c4-07a9-4749-a8ae-072cdd3fd0af","Type":"ContainerStarted","Data":"0a29b3e5c0ec5a91dab7874c867b2038a072c746745673f1b41d3b193d6564a7"} Nov 25 15:24:05 crc kubenswrapper[4890]: I1125 15:24:05.822633 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f27f06e7-28d2-403a-92c8-b1d858d90619","Type":"ContainerStarted","Data":"52471a5385ef1782f3a3a05a0641f78c0754dff41cd46dbf9f69332464a23969"} Nov 25 15:24:05 crc kubenswrapper[4890]: I1125 15:24:05.822680 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 25 15:24:05 crc kubenswrapper[4890]: I1125 15:24:05.832176 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-s2qx4" podStartSLOduration=2.832138645 podStartE2EDuration="2.832138645s" podCreationTimestamp="2025-11-25 15:24:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:24:05.826618117 +0000 UTC m=+1304.269080727" watchObservedRunningTime="2025-11-25 15:24:05.832138645 +0000 UTC m=+1304.274601255" Nov 25 15:24:05 crc kubenswrapper[4890]: I1125 15:24:05.857232 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.093388072 podStartE2EDuration="5.857207365s" podCreationTimestamp="2025-11-25 15:24:00 +0000 UTC" firstStartedPulling="2025-11-25 15:24:01.624483433 +0000 UTC m=+1300.066946043" lastFinishedPulling="2025-11-25 15:24:05.388302726 +0000 UTC m=+1303.830765336" observedRunningTime="2025-11-25 15:24:05.84546793 +0000 UTC m=+1304.287930550" watchObservedRunningTime="2025-11-25 15:24:05.857207365 +0000 UTC m=+1304.299669975" Nov 25 15:24:06 crc kubenswrapper[4890]: I1125 15:24:06.236346 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-89c5cd4d5-4fhbm" Nov 25 15:24:06 crc kubenswrapper[4890]: I1125 15:24:06.313446 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-kjvns"] Nov 25 15:24:06 crc kubenswrapper[4890]: I1125 15:24:06.313703 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-757b4f8459-kjvns" podUID="9060524b-48eb-42ba-b1e7-5a852d89d94f" containerName="dnsmasq-dns" containerID="cri-o://9a1c99eefeb5490ba9f1a7c457143c55e8a76e9cf439ecb289dfeec1c9681662" gracePeriod=10 Nov 25 15:24:06 crc kubenswrapper[4890]: I1125 15:24:06.834992 4890 generic.go:334] "Generic (PLEG): container finished" podID="9060524b-48eb-42ba-b1e7-5a852d89d94f" containerID="9a1c99eefeb5490ba9f1a7c457143c55e8a76e9cf439ecb289dfeec1c9681662" exitCode=0 Nov 25 15:24:06 crc kubenswrapper[4890]: I1125 15:24:06.835947 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-kjvns" event={"ID":"9060524b-48eb-42ba-b1e7-5a852d89d94f","Type":"ContainerDied","Data":"9a1c99eefeb5490ba9f1a7c457143c55e8a76e9cf439ecb289dfeec1c9681662"} Nov 25 15:24:06 crc kubenswrapper[4890]: I1125 15:24:06.835991 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-kjvns" event={"ID":"9060524b-48eb-42ba-b1e7-5a852d89d94f","Type":"ContainerDied","Data":"d5f2da4c503149a55d72aab22d036a67883e3df32d7c9dc4e4a2c1a7b60e8f81"} Nov 25 15:24:06 crc kubenswrapper[4890]: I1125 15:24:06.836007 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d5f2da4c503149a55d72aab22d036a67883e3df32d7c9dc4e4a2c1a7b60e8f81" Nov 25 15:24:06 crc kubenswrapper[4890]: I1125 15:24:06.883789 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-kjvns" Nov 25 15:24:06 crc kubenswrapper[4890]: I1125 15:24:06.950954 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9060524b-48eb-42ba-b1e7-5a852d89d94f-config\") pod \"9060524b-48eb-42ba-b1e7-5a852d89d94f\" (UID: \"9060524b-48eb-42ba-b1e7-5a852d89d94f\") " Nov 25 15:24:06 crc kubenswrapper[4890]: I1125 15:24:06.951041 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m28qt\" (UniqueName: \"kubernetes.io/projected/9060524b-48eb-42ba-b1e7-5a852d89d94f-kube-api-access-m28qt\") pod \"9060524b-48eb-42ba-b1e7-5a852d89d94f\" (UID: \"9060524b-48eb-42ba-b1e7-5a852d89d94f\") " Nov 25 15:24:06 crc kubenswrapper[4890]: I1125 15:24:06.951202 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9060524b-48eb-42ba-b1e7-5a852d89d94f-dns-swift-storage-0\") pod \"9060524b-48eb-42ba-b1e7-5a852d89d94f\" (UID: \"9060524b-48eb-42ba-b1e7-5a852d89d94f\") " Nov 25 15:24:06 crc kubenswrapper[4890]: I1125 15:24:06.951270 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9060524b-48eb-42ba-b1e7-5a852d89d94f-dns-svc\") pod \"9060524b-48eb-42ba-b1e7-5a852d89d94f\" (UID: \"9060524b-48eb-42ba-b1e7-5a852d89d94f\") " Nov 25 15:24:06 crc kubenswrapper[4890]: I1125 15:24:06.951321 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9060524b-48eb-42ba-b1e7-5a852d89d94f-ovsdbserver-sb\") pod \"9060524b-48eb-42ba-b1e7-5a852d89d94f\" (UID: \"9060524b-48eb-42ba-b1e7-5a852d89d94f\") " Nov 25 15:24:06 crc kubenswrapper[4890]: I1125 15:24:06.951360 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9060524b-48eb-42ba-b1e7-5a852d89d94f-ovsdbserver-nb\") pod \"9060524b-48eb-42ba-b1e7-5a852d89d94f\" (UID: \"9060524b-48eb-42ba-b1e7-5a852d89d94f\") " Nov 25 15:24:06 crc kubenswrapper[4890]: I1125 15:24:06.979015 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9060524b-48eb-42ba-b1e7-5a852d89d94f-kube-api-access-m28qt" (OuterVolumeSpecName: "kube-api-access-m28qt") pod "9060524b-48eb-42ba-b1e7-5a852d89d94f" (UID: "9060524b-48eb-42ba-b1e7-5a852d89d94f"). InnerVolumeSpecName "kube-api-access-m28qt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:24:07 crc kubenswrapper[4890]: I1125 15:24:07.007380 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9060524b-48eb-42ba-b1e7-5a852d89d94f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9060524b-48eb-42ba-b1e7-5a852d89d94f" (UID: "9060524b-48eb-42ba-b1e7-5a852d89d94f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:24:07 crc kubenswrapper[4890]: I1125 15:24:07.007779 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9060524b-48eb-42ba-b1e7-5a852d89d94f-config" (OuterVolumeSpecName: "config") pod "9060524b-48eb-42ba-b1e7-5a852d89d94f" (UID: "9060524b-48eb-42ba-b1e7-5a852d89d94f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:24:07 crc kubenswrapper[4890]: I1125 15:24:07.008548 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9060524b-48eb-42ba-b1e7-5a852d89d94f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9060524b-48eb-42ba-b1e7-5a852d89d94f" (UID: "9060524b-48eb-42ba-b1e7-5a852d89d94f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:24:07 crc kubenswrapper[4890]: I1125 15:24:07.020770 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9060524b-48eb-42ba-b1e7-5a852d89d94f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9060524b-48eb-42ba-b1e7-5a852d89d94f" (UID: "9060524b-48eb-42ba-b1e7-5a852d89d94f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:24:07 crc kubenswrapper[4890]: I1125 15:24:07.038016 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9060524b-48eb-42ba-b1e7-5a852d89d94f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9060524b-48eb-42ba-b1e7-5a852d89d94f" (UID: "9060524b-48eb-42ba-b1e7-5a852d89d94f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:24:07 crc kubenswrapper[4890]: I1125 15:24:07.053593 4890 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9060524b-48eb-42ba-b1e7-5a852d89d94f-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:07 crc kubenswrapper[4890]: I1125 15:24:07.053624 4890 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9060524b-48eb-42ba-b1e7-5a852d89d94f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:07 crc kubenswrapper[4890]: I1125 15:24:07.053636 4890 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9060524b-48eb-42ba-b1e7-5a852d89d94f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:07 crc kubenswrapper[4890]: I1125 15:24:07.053645 4890 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9060524b-48eb-42ba-b1e7-5a852d89d94f-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:07 crc kubenswrapper[4890]: I1125 15:24:07.053654 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m28qt\" (UniqueName: \"kubernetes.io/projected/9060524b-48eb-42ba-b1e7-5a852d89d94f-kube-api-access-m28qt\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:07 crc kubenswrapper[4890]: I1125 15:24:07.053664 4890 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9060524b-48eb-42ba-b1e7-5a852d89d94f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:07 crc kubenswrapper[4890]: I1125 15:24:07.844061 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-kjvns" Nov 25 15:24:07 crc kubenswrapper[4890]: I1125 15:24:07.880122 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-kjvns"] Nov 25 15:24:07 crc kubenswrapper[4890]: I1125 15:24:07.888393 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-kjvns"] Nov 25 15:24:08 crc kubenswrapper[4890]: I1125 15:24:08.189505 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9060524b-48eb-42ba-b1e7-5a852d89d94f" path="/var/lib/kubelet/pods/9060524b-48eb-42ba-b1e7-5a852d89d94f/volumes" Nov 25 15:24:10 crc kubenswrapper[4890]: I1125 15:24:10.872437 4890 generic.go:334] "Generic (PLEG): container finished" podID="519d05c4-07a9-4749-a8ae-072cdd3fd0af" containerID="1557774a1642615323744cb5147878a67bf6499e0cacf33a93edbb5e592537bc" exitCode=0 Nov 25 15:24:10 crc kubenswrapper[4890]: I1125 15:24:10.872693 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-s2qx4" event={"ID":"519d05c4-07a9-4749-a8ae-072cdd3fd0af","Type":"ContainerDied","Data":"1557774a1642615323744cb5147878a67bf6499e0cacf33a93edbb5e592537bc"} Nov 25 15:24:12 crc kubenswrapper[4890]: I1125 15:24:12.270811 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-s2qx4" Nov 25 15:24:12 crc kubenswrapper[4890]: I1125 15:24:12.349554 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/519d05c4-07a9-4749-a8ae-072cdd3fd0af-config-data\") pod \"519d05c4-07a9-4749-a8ae-072cdd3fd0af\" (UID: \"519d05c4-07a9-4749-a8ae-072cdd3fd0af\") " Nov 25 15:24:12 crc kubenswrapper[4890]: I1125 15:24:12.349620 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/519d05c4-07a9-4749-a8ae-072cdd3fd0af-scripts\") pod \"519d05c4-07a9-4749-a8ae-072cdd3fd0af\" (UID: \"519d05c4-07a9-4749-a8ae-072cdd3fd0af\") " Nov 25 15:24:12 crc kubenswrapper[4890]: I1125 15:24:12.349680 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2q54\" (UniqueName: \"kubernetes.io/projected/519d05c4-07a9-4749-a8ae-072cdd3fd0af-kube-api-access-j2q54\") pod \"519d05c4-07a9-4749-a8ae-072cdd3fd0af\" (UID: \"519d05c4-07a9-4749-a8ae-072cdd3fd0af\") " Nov 25 15:24:12 crc kubenswrapper[4890]: I1125 15:24:12.349707 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/519d05c4-07a9-4749-a8ae-072cdd3fd0af-combined-ca-bundle\") pod \"519d05c4-07a9-4749-a8ae-072cdd3fd0af\" (UID: \"519d05c4-07a9-4749-a8ae-072cdd3fd0af\") " Nov 25 15:24:12 crc kubenswrapper[4890]: I1125 15:24:12.355551 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/519d05c4-07a9-4749-a8ae-072cdd3fd0af-scripts" (OuterVolumeSpecName: "scripts") pod "519d05c4-07a9-4749-a8ae-072cdd3fd0af" (UID: "519d05c4-07a9-4749-a8ae-072cdd3fd0af"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:24:12 crc kubenswrapper[4890]: I1125 15:24:12.356558 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/519d05c4-07a9-4749-a8ae-072cdd3fd0af-kube-api-access-j2q54" (OuterVolumeSpecName: "kube-api-access-j2q54") pod "519d05c4-07a9-4749-a8ae-072cdd3fd0af" (UID: "519d05c4-07a9-4749-a8ae-072cdd3fd0af"). InnerVolumeSpecName "kube-api-access-j2q54". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:24:12 crc kubenswrapper[4890]: I1125 15:24:12.380533 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/519d05c4-07a9-4749-a8ae-072cdd3fd0af-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "519d05c4-07a9-4749-a8ae-072cdd3fd0af" (UID: "519d05c4-07a9-4749-a8ae-072cdd3fd0af"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:24:12 crc kubenswrapper[4890]: I1125 15:24:12.382246 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/519d05c4-07a9-4749-a8ae-072cdd3fd0af-config-data" (OuterVolumeSpecName: "config-data") pod "519d05c4-07a9-4749-a8ae-072cdd3fd0af" (UID: "519d05c4-07a9-4749-a8ae-072cdd3fd0af"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:24:12 crc kubenswrapper[4890]: I1125 15:24:12.452364 4890 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/519d05c4-07a9-4749-a8ae-072cdd3fd0af-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:12 crc kubenswrapper[4890]: I1125 15:24:12.452408 4890 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/519d05c4-07a9-4749-a8ae-072cdd3fd0af-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:12 crc kubenswrapper[4890]: I1125 15:24:12.452419 4890 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/519d05c4-07a9-4749-a8ae-072cdd3fd0af-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:12 crc kubenswrapper[4890]: I1125 15:24:12.452431 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2q54\" (UniqueName: \"kubernetes.io/projected/519d05c4-07a9-4749-a8ae-072cdd3fd0af-kube-api-access-j2q54\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:12 crc kubenswrapper[4890]: I1125 15:24:12.891712 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-s2qx4" event={"ID":"519d05c4-07a9-4749-a8ae-072cdd3fd0af","Type":"ContainerDied","Data":"0a29b3e5c0ec5a91dab7874c867b2038a072c746745673f1b41d3b193d6564a7"} Nov 25 15:24:12 crc kubenswrapper[4890]: I1125 15:24:12.891753 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a29b3e5c0ec5a91dab7874c867b2038a072c746745673f1b41d3b193d6564a7" Nov 25 15:24:12 crc kubenswrapper[4890]: I1125 15:24:12.891766 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-s2qx4" Nov 25 15:24:13 crc kubenswrapper[4890]: I1125 15:24:13.073270 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 25 15:24:13 crc kubenswrapper[4890]: I1125 15:24:13.073552 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d89315fb-6d71-4209-a3ea-a46d6a8e7691" containerName="nova-api-log" containerID="cri-o://39185f3ea75420ac3f496d66c9c18f68adfd5be09cbd2ef739cce24171d2ea53" gracePeriod=30 Nov 25 15:24:13 crc kubenswrapper[4890]: I1125 15:24:13.073623 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d89315fb-6d71-4209-a3ea-a46d6a8e7691" containerName="nova-api-api" containerID="cri-o://ed7b44eec03eedda56084acc550e69c1560a3b14ae04652eb8dfc16d1db9a592" gracePeriod=30 Nov 25 15:24:13 crc kubenswrapper[4890]: I1125 15:24:13.149186 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 15:24:13 crc kubenswrapper[4890]: I1125 15:24:13.149748 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="12acc506-6412-4a9e-a80c-4b1cc761246d" containerName="nova-scheduler-scheduler" containerID="cri-o://7364e8cda4f00d093052c4416a5150b9c81e120eb72578c5db8e133dd20d94fa" gracePeriod=30 Nov 25 15:24:13 crc kubenswrapper[4890]: I1125 15:24:13.158876 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 15:24:13 crc kubenswrapper[4890]: I1125 15:24:13.159138 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="fb7922d1-e2c8-4ee7-a700-6e6353de7d91" containerName="nova-metadata-log" containerID="cri-o://2b9bbbdc949ac4ef3283ade2efc3a174ce0f3ed26d413db474f54f62904e12d8" gracePeriod=30 Nov 25 15:24:13 crc kubenswrapper[4890]: I1125 15:24:13.159611 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="fb7922d1-e2c8-4ee7-a700-6e6353de7d91" containerName="nova-metadata-metadata" containerID="cri-o://53a207e706c9c4a6fe645567e7859d702ccd877ceb186ecc2155100a47808c7e" gracePeriod=30 Nov 25 15:24:13 crc kubenswrapper[4890]: I1125 15:24:13.734776 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 15:24:13 crc kubenswrapper[4890]: I1125 15:24:13.880913 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d89315fb-6d71-4209-a3ea-a46d6a8e7691-public-tls-certs\") pod \"d89315fb-6d71-4209-a3ea-a46d6a8e7691\" (UID: \"d89315fb-6d71-4209-a3ea-a46d6a8e7691\") " Nov 25 15:24:13 crc kubenswrapper[4890]: I1125 15:24:13.880968 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d89315fb-6d71-4209-a3ea-a46d6a8e7691-internal-tls-certs\") pod \"d89315fb-6d71-4209-a3ea-a46d6a8e7691\" (UID: \"d89315fb-6d71-4209-a3ea-a46d6a8e7691\") " Nov 25 15:24:13 crc kubenswrapper[4890]: I1125 15:24:13.881070 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wr2ts\" (UniqueName: \"kubernetes.io/projected/d89315fb-6d71-4209-a3ea-a46d6a8e7691-kube-api-access-wr2ts\") pod \"d89315fb-6d71-4209-a3ea-a46d6a8e7691\" (UID: \"d89315fb-6d71-4209-a3ea-a46d6a8e7691\") " Nov 25 15:24:13 crc kubenswrapper[4890]: I1125 15:24:13.881091 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d89315fb-6d71-4209-a3ea-a46d6a8e7691-combined-ca-bundle\") pod \"d89315fb-6d71-4209-a3ea-a46d6a8e7691\" (UID: \"d89315fb-6d71-4209-a3ea-a46d6a8e7691\") " Nov 25 15:24:13 crc kubenswrapper[4890]: I1125 15:24:13.881145 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d89315fb-6d71-4209-a3ea-a46d6a8e7691-logs\") pod \"d89315fb-6d71-4209-a3ea-a46d6a8e7691\" (UID: \"d89315fb-6d71-4209-a3ea-a46d6a8e7691\") " Nov 25 15:24:13 crc kubenswrapper[4890]: I1125 15:24:13.881265 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d89315fb-6d71-4209-a3ea-a46d6a8e7691-config-data\") pod \"d89315fb-6d71-4209-a3ea-a46d6a8e7691\" (UID: \"d89315fb-6d71-4209-a3ea-a46d6a8e7691\") " Nov 25 15:24:13 crc kubenswrapper[4890]: I1125 15:24:13.882657 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d89315fb-6d71-4209-a3ea-a46d6a8e7691-logs" (OuterVolumeSpecName: "logs") pod "d89315fb-6d71-4209-a3ea-a46d6a8e7691" (UID: "d89315fb-6d71-4209-a3ea-a46d6a8e7691"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:24:13 crc kubenswrapper[4890]: I1125 15:24:13.889009 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d89315fb-6d71-4209-a3ea-a46d6a8e7691-kube-api-access-wr2ts" (OuterVolumeSpecName: "kube-api-access-wr2ts") pod "d89315fb-6d71-4209-a3ea-a46d6a8e7691" (UID: "d89315fb-6d71-4209-a3ea-a46d6a8e7691"). InnerVolumeSpecName "kube-api-access-wr2ts". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:24:13 crc kubenswrapper[4890]: I1125 15:24:13.910021 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d89315fb-6d71-4209-a3ea-a46d6a8e7691-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d89315fb-6d71-4209-a3ea-a46d6a8e7691" (UID: "d89315fb-6d71-4209-a3ea-a46d6a8e7691"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:24:13 crc kubenswrapper[4890]: I1125 15:24:13.920413 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d89315fb-6d71-4209-a3ea-a46d6a8e7691-config-data" (OuterVolumeSpecName: "config-data") pod "d89315fb-6d71-4209-a3ea-a46d6a8e7691" (UID: "d89315fb-6d71-4209-a3ea-a46d6a8e7691"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:24:13 crc kubenswrapper[4890]: I1125 15:24:13.920496 4890 generic.go:334] "Generic (PLEG): container finished" podID="d89315fb-6d71-4209-a3ea-a46d6a8e7691" containerID="ed7b44eec03eedda56084acc550e69c1560a3b14ae04652eb8dfc16d1db9a592" exitCode=0 Nov 25 15:24:13 crc kubenswrapper[4890]: I1125 15:24:13.920528 4890 generic.go:334] "Generic (PLEG): container finished" podID="d89315fb-6d71-4209-a3ea-a46d6a8e7691" containerID="39185f3ea75420ac3f496d66c9c18f68adfd5be09cbd2ef739cce24171d2ea53" exitCode=143 Nov 25 15:24:13 crc kubenswrapper[4890]: I1125 15:24:13.920570 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 15:24:13 crc kubenswrapper[4890]: I1125 15:24:13.920595 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d89315fb-6d71-4209-a3ea-a46d6a8e7691","Type":"ContainerDied","Data":"ed7b44eec03eedda56084acc550e69c1560a3b14ae04652eb8dfc16d1db9a592"} Nov 25 15:24:13 crc kubenswrapper[4890]: I1125 15:24:13.920626 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d89315fb-6d71-4209-a3ea-a46d6a8e7691","Type":"ContainerDied","Data":"39185f3ea75420ac3f496d66c9c18f68adfd5be09cbd2ef739cce24171d2ea53"} Nov 25 15:24:13 crc kubenswrapper[4890]: I1125 15:24:13.920638 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d89315fb-6d71-4209-a3ea-a46d6a8e7691","Type":"ContainerDied","Data":"67ca661595e44ff01859de973a82d8adee6c9305980a409f14c9739c5caa6371"} Nov 25 15:24:13 crc kubenswrapper[4890]: I1125 15:24:13.920655 4890 scope.go:117] "RemoveContainer" containerID="ed7b44eec03eedda56084acc550e69c1560a3b14ae04652eb8dfc16d1db9a592" Nov 25 15:24:13 crc kubenswrapper[4890]: I1125 15:24:13.925501 4890 generic.go:334] "Generic (PLEG): container finished" podID="fb7922d1-e2c8-4ee7-a700-6e6353de7d91" containerID="2b9bbbdc949ac4ef3283ade2efc3a174ce0f3ed26d413db474f54f62904e12d8" exitCode=143 Nov 25 15:24:13 crc kubenswrapper[4890]: I1125 15:24:13.925532 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fb7922d1-e2c8-4ee7-a700-6e6353de7d91","Type":"ContainerDied","Data":"2b9bbbdc949ac4ef3283ade2efc3a174ce0f3ed26d413db474f54f62904e12d8"} Nov 25 15:24:13 crc kubenswrapper[4890]: I1125 15:24:13.940895 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d89315fb-6d71-4209-a3ea-a46d6a8e7691-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "d89315fb-6d71-4209-a3ea-a46d6a8e7691" (UID: "d89315fb-6d71-4209-a3ea-a46d6a8e7691"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:24:13 crc kubenswrapper[4890]: I1125 15:24:13.943556 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d89315fb-6d71-4209-a3ea-a46d6a8e7691-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d89315fb-6d71-4209-a3ea-a46d6a8e7691" (UID: "d89315fb-6d71-4209-a3ea-a46d6a8e7691"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:24:13 crc kubenswrapper[4890]: I1125 15:24:13.985257 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wr2ts\" (UniqueName: \"kubernetes.io/projected/d89315fb-6d71-4209-a3ea-a46d6a8e7691-kube-api-access-wr2ts\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:13 crc kubenswrapper[4890]: I1125 15:24:13.985333 4890 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d89315fb-6d71-4209-a3ea-a46d6a8e7691-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:13 crc kubenswrapper[4890]: I1125 15:24:13.985343 4890 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d89315fb-6d71-4209-a3ea-a46d6a8e7691-logs\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:13 crc kubenswrapper[4890]: I1125 15:24:13.985353 4890 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d89315fb-6d71-4209-a3ea-a46d6a8e7691-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:13 crc kubenswrapper[4890]: I1125 15:24:13.985363 4890 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d89315fb-6d71-4209-a3ea-a46d6a8e7691-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:13 crc kubenswrapper[4890]: I1125 15:24:13.985371 4890 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d89315fb-6d71-4209-a3ea-a46d6a8e7691-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.028604 4890 scope.go:117] "RemoveContainer" containerID="39185f3ea75420ac3f496d66c9c18f68adfd5be09cbd2ef739cce24171d2ea53" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.048379 4890 scope.go:117] "RemoveContainer" containerID="ed7b44eec03eedda56084acc550e69c1560a3b14ae04652eb8dfc16d1db9a592" Nov 25 15:24:14 crc kubenswrapper[4890]: E1125 15:24:14.049235 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed7b44eec03eedda56084acc550e69c1560a3b14ae04652eb8dfc16d1db9a592\": container with ID starting with ed7b44eec03eedda56084acc550e69c1560a3b14ae04652eb8dfc16d1db9a592 not found: ID does not exist" containerID="ed7b44eec03eedda56084acc550e69c1560a3b14ae04652eb8dfc16d1db9a592" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.049274 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed7b44eec03eedda56084acc550e69c1560a3b14ae04652eb8dfc16d1db9a592"} err="failed to get container status \"ed7b44eec03eedda56084acc550e69c1560a3b14ae04652eb8dfc16d1db9a592\": rpc error: code = NotFound desc = could not find container \"ed7b44eec03eedda56084acc550e69c1560a3b14ae04652eb8dfc16d1db9a592\": container with ID starting with ed7b44eec03eedda56084acc550e69c1560a3b14ae04652eb8dfc16d1db9a592 not found: ID does not exist" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.049303 4890 scope.go:117] "RemoveContainer" containerID="39185f3ea75420ac3f496d66c9c18f68adfd5be09cbd2ef739cce24171d2ea53" Nov 25 15:24:14 crc kubenswrapper[4890]: E1125 15:24:14.049656 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39185f3ea75420ac3f496d66c9c18f68adfd5be09cbd2ef739cce24171d2ea53\": container with ID starting with 39185f3ea75420ac3f496d66c9c18f68adfd5be09cbd2ef739cce24171d2ea53 not found: ID does not exist" containerID="39185f3ea75420ac3f496d66c9c18f68adfd5be09cbd2ef739cce24171d2ea53" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.049682 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39185f3ea75420ac3f496d66c9c18f68adfd5be09cbd2ef739cce24171d2ea53"} err="failed to get container status \"39185f3ea75420ac3f496d66c9c18f68adfd5be09cbd2ef739cce24171d2ea53\": rpc error: code = NotFound desc = could not find container \"39185f3ea75420ac3f496d66c9c18f68adfd5be09cbd2ef739cce24171d2ea53\": container with ID starting with 39185f3ea75420ac3f496d66c9c18f68adfd5be09cbd2ef739cce24171d2ea53 not found: ID does not exist" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.049700 4890 scope.go:117] "RemoveContainer" containerID="ed7b44eec03eedda56084acc550e69c1560a3b14ae04652eb8dfc16d1db9a592" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.049942 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed7b44eec03eedda56084acc550e69c1560a3b14ae04652eb8dfc16d1db9a592"} err="failed to get container status \"ed7b44eec03eedda56084acc550e69c1560a3b14ae04652eb8dfc16d1db9a592\": rpc error: code = NotFound desc = could not find container \"ed7b44eec03eedda56084acc550e69c1560a3b14ae04652eb8dfc16d1db9a592\": container with ID starting with ed7b44eec03eedda56084acc550e69c1560a3b14ae04652eb8dfc16d1db9a592 not found: ID does not exist" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.049966 4890 scope.go:117] "RemoveContainer" containerID="39185f3ea75420ac3f496d66c9c18f68adfd5be09cbd2ef739cce24171d2ea53" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.050452 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39185f3ea75420ac3f496d66c9c18f68adfd5be09cbd2ef739cce24171d2ea53"} err="failed to get container status \"39185f3ea75420ac3f496d66c9c18f68adfd5be09cbd2ef739cce24171d2ea53\": rpc error: code = NotFound desc = could not find container \"39185f3ea75420ac3f496d66c9c18f68adfd5be09cbd2ef739cce24171d2ea53\": container with ID starting with 39185f3ea75420ac3f496d66c9c18f68adfd5be09cbd2ef739cce24171d2ea53 not found: ID does not exist" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.278183 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.289315 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.311888 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 25 15:24:14 crc kubenswrapper[4890]: E1125 15:24:14.312281 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9060524b-48eb-42ba-b1e7-5a852d89d94f" containerName="dnsmasq-dns" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.312314 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="9060524b-48eb-42ba-b1e7-5a852d89d94f" containerName="dnsmasq-dns" Nov 25 15:24:14 crc kubenswrapper[4890]: E1125 15:24:14.312340 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d89315fb-6d71-4209-a3ea-a46d6a8e7691" containerName="nova-api-api" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.312349 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="d89315fb-6d71-4209-a3ea-a46d6a8e7691" containerName="nova-api-api" Nov 25 15:24:14 crc kubenswrapper[4890]: E1125 15:24:14.312375 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="519d05c4-07a9-4749-a8ae-072cdd3fd0af" containerName="nova-manage" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.312383 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="519d05c4-07a9-4749-a8ae-072cdd3fd0af" containerName="nova-manage" Nov 25 15:24:14 crc kubenswrapper[4890]: E1125 15:24:14.312398 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9060524b-48eb-42ba-b1e7-5a852d89d94f" containerName="init" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.312405 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="9060524b-48eb-42ba-b1e7-5a852d89d94f" containerName="init" Nov 25 15:24:14 crc kubenswrapper[4890]: E1125 15:24:14.312437 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d89315fb-6d71-4209-a3ea-a46d6a8e7691" containerName="nova-api-log" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.312444 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="d89315fb-6d71-4209-a3ea-a46d6a8e7691" containerName="nova-api-log" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.312633 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="519d05c4-07a9-4749-a8ae-072cdd3fd0af" containerName="nova-manage" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.312651 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="9060524b-48eb-42ba-b1e7-5a852d89d94f" containerName="dnsmasq-dns" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.312666 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="d89315fb-6d71-4209-a3ea-a46d6a8e7691" containerName="nova-api-api" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.312678 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="d89315fb-6d71-4209-a3ea-a46d6a8e7691" containerName="nova-api-log" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.313663 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.320244 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.320674 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.320847 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.321778 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.411315 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.514380 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c6b60ed-19f6-472a-92d6-2a1e06c92ea6-config-data\") pod \"nova-api-0\" (UID: \"3c6b60ed-19f6-472a-92d6-2a1e06c92ea6\") " pod="openstack/nova-api-0" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.514507 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c6b60ed-19f6-472a-92d6-2a1e06c92ea6-internal-tls-certs\") pod \"nova-api-0\" (UID: \"3c6b60ed-19f6-472a-92d6-2a1e06c92ea6\") " pod="openstack/nova-api-0" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.514609 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c6b60ed-19f6-472a-92d6-2a1e06c92ea6-public-tls-certs\") pod \"nova-api-0\" (UID: \"3c6b60ed-19f6-472a-92d6-2a1e06c92ea6\") " pod="openstack/nova-api-0" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.514677 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2592z\" (UniqueName: \"kubernetes.io/projected/3c6b60ed-19f6-472a-92d6-2a1e06c92ea6-kube-api-access-2592z\") pod \"nova-api-0\" (UID: \"3c6b60ed-19f6-472a-92d6-2a1e06c92ea6\") " pod="openstack/nova-api-0" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.514815 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c6b60ed-19f6-472a-92d6-2a1e06c92ea6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3c6b60ed-19f6-472a-92d6-2a1e06c92ea6\") " pod="openstack/nova-api-0" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.514894 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c6b60ed-19f6-472a-92d6-2a1e06c92ea6-logs\") pod \"nova-api-0\" (UID: \"3c6b60ed-19f6-472a-92d6-2a1e06c92ea6\") " pod="openstack/nova-api-0" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.616292 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cl9sg\" (UniqueName: \"kubernetes.io/projected/12acc506-6412-4a9e-a80c-4b1cc761246d-kube-api-access-cl9sg\") pod \"12acc506-6412-4a9e-a80c-4b1cc761246d\" (UID: \"12acc506-6412-4a9e-a80c-4b1cc761246d\") " Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.616353 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12acc506-6412-4a9e-a80c-4b1cc761246d-combined-ca-bundle\") pod \"12acc506-6412-4a9e-a80c-4b1cc761246d\" (UID: \"12acc506-6412-4a9e-a80c-4b1cc761246d\") " Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.616417 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12acc506-6412-4a9e-a80c-4b1cc761246d-config-data\") pod \"12acc506-6412-4a9e-a80c-4b1cc761246d\" (UID: \"12acc506-6412-4a9e-a80c-4b1cc761246d\") " Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.616680 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c6b60ed-19f6-472a-92d6-2a1e06c92ea6-public-tls-certs\") pod \"nova-api-0\" (UID: \"3c6b60ed-19f6-472a-92d6-2a1e06c92ea6\") " pod="openstack/nova-api-0" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.617875 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2592z\" (UniqueName: \"kubernetes.io/projected/3c6b60ed-19f6-472a-92d6-2a1e06c92ea6-kube-api-access-2592z\") pod \"nova-api-0\" (UID: \"3c6b60ed-19f6-472a-92d6-2a1e06c92ea6\") " pod="openstack/nova-api-0" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.617954 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c6b60ed-19f6-472a-92d6-2a1e06c92ea6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3c6b60ed-19f6-472a-92d6-2a1e06c92ea6\") " pod="openstack/nova-api-0" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.618013 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c6b60ed-19f6-472a-92d6-2a1e06c92ea6-logs\") pod \"nova-api-0\" (UID: \"3c6b60ed-19f6-472a-92d6-2a1e06c92ea6\") " pod="openstack/nova-api-0" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.618455 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c6b60ed-19f6-472a-92d6-2a1e06c92ea6-config-data\") pod \"nova-api-0\" (UID: \"3c6b60ed-19f6-472a-92d6-2a1e06c92ea6\") " pod="openstack/nova-api-0" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.618525 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c6b60ed-19f6-472a-92d6-2a1e06c92ea6-internal-tls-certs\") pod \"nova-api-0\" (UID: \"3c6b60ed-19f6-472a-92d6-2a1e06c92ea6\") " pod="openstack/nova-api-0" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.619449 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c6b60ed-19f6-472a-92d6-2a1e06c92ea6-logs\") pod \"nova-api-0\" (UID: \"3c6b60ed-19f6-472a-92d6-2a1e06c92ea6\") " pod="openstack/nova-api-0" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.621347 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12acc506-6412-4a9e-a80c-4b1cc761246d-kube-api-access-cl9sg" (OuterVolumeSpecName: "kube-api-access-cl9sg") pod "12acc506-6412-4a9e-a80c-4b1cc761246d" (UID: "12acc506-6412-4a9e-a80c-4b1cc761246d"). InnerVolumeSpecName "kube-api-access-cl9sg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.623815 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c6b60ed-19f6-472a-92d6-2a1e06c92ea6-public-tls-certs\") pod \"nova-api-0\" (UID: \"3c6b60ed-19f6-472a-92d6-2a1e06c92ea6\") " pod="openstack/nova-api-0" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.624319 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c6b60ed-19f6-472a-92d6-2a1e06c92ea6-internal-tls-certs\") pod \"nova-api-0\" (UID: \"3c6b60ed-19f6-472a-92d6-2a1e06c92ea6\") " pod="openstack/nova-api-0" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.625253 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c6b60ed-19f6-472a-92d6-2a1e06c92ea6-config-data\") pod \"nova-api-0\" (UID: \"3c6b60ed-19f6-472a-92d6-2a1e06c92ea6\") " pod="openstack/nova-api-0" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.625904 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c6b60ed-19f6-472a-92d6-2a1e06c92ea6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3c6b60ed-19f6-472a-92d6-2a1e06c92ea6\") " pod="openstack/nova-api-0" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.641398 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2592z\" (UniqueName: \"kubernetes.io/projected/3c6b60ed-19f6-472a-92d6-2a1e06c92ea6-kube-api-access-2592z\") pod \"nova-api-0\" (UID: \"3c6b60ed-19f6-472a-92d6-2a1e06c92ea6\") " pod="openstack/nova-api-0" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.657444 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12acc506-6412-4a9e-a80c-4b1cc761246d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "12acc506-6412-4a9e-a80c-4b1cc761246d" (UID: "12acc506-6412-4a9e-a80c-4b1cc761246d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.667634 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12acc506-6412-4a9e-a80c-4b1cc761246d-config-data" (OuterVolumeSpecName: "config-data") pod "12acc506-6412-4a9e-a80c-4b1cc761246d" (UID: "12acc506-6412-4a9e-a80c-4b1cc761246d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.720426 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cl9sg\" (UniqueName: \"kubernetes.io/projected/12acc506-6412-4a9e-a80c-4b1cc761246d-kube-api-access-cl9sg\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.720474 4890 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12acc506-6412-4a9e-a80c-4b1cc761246d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.720487 4890 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12acc506-6412-4a9e-a80c-4b1cc761246d-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.730622 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.950634 4890 generic.go:334] "Generic (PLEG): container finished" podID="12acc506-6412-4a9e-a80c-4b1cc761246d" containerID="7364e8cda4f00d093052c4416a5150b9c81e120eb72578c5db8e133dd20d94fa" exitCode=0 Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.950677 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"12acc506-6412-4a9e-a80c-4b1cc761246d","Type":"ContainerDied","Data":"7364e8cda4f00d093052c4416a5150b9c81e120eb72578c5db8e133dd20d94fa"} Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.950722 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"12acc506-6412-4a9e-a80c-4b1cc761246d","Type":"ContainerDied","Data":"1cc5a35a5b85d593f9a447a8791877cc01e4ee3875a3bf6f0782bcffc96ff089"} Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.950741 4890 scope.go:117] "RemoveContainer" containerID="7364e8cda4f00d093052c4416a5150b9c81e120eb72578c5db8e133dd20d94fa" Nov 25 15:24:14 crc kubenswrapper[4890]: I1125 15:24:14.950864 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 15:24:15 crc kubenswrapper[4890]: I1125 15:24:15.008234 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 15:24:15 crc kubenswrapper[4890]: I1125 15:24:15.020870 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 15:24:15 crc kubenswrapper[4890]: I1125 15:24:15.022847 4890 scope.go:117] "RemoveContainer" containerID="7364e8cda4f00d093052c4416a5150b9c81e120eb72578c5db8e133dd20d94fa" Nov 25 15:24:15 crc kubenswrapper[4890]: E1125 15:24:15.023589 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7364e8cda4f00d093052c4416a5150b9c81e120eb72578c5db8e133dd20d94fa\": container with ID starting with 7364e8cda4f00d093052c4416a5150b9c81e120eb72578c5db8e133dd20d94fa not found: ID does not exist" containerID="7364e8cda4f00d093052c4416a5150b9c81e120eb72578c5db8e133dd20d94fa" Nov 25 15:24:15 crc kubenswrapper[4890]: I1125 15:24:15.023639 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7364e8cda4f00d093052c4416a5150b9c81e120eb72578c5db8e133dd20d94fa"} err="failed to get container status \"7364e8cda4f00d093052c4416a5150b9c81e120eb72578c5db8e133dd20d94fa\": rpc error: code = NotFound desc = could not find container \"7364e8cda4f00d093052c4416a5150b9c81e120eb72578c5db8e133dd20d94fa\": container with ID starting with 7364e8cda4f00d093052c4416a5150b9c81e120eb72578c5db8e133dd20d94fa not found: ID does not exist" Nov 25 15:24:15 crc kubenswrapper[4890]: I1125 15:24:15.034277 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 15:24:15 crc kubenswrapper[4890]: E1125 15:24:15.034831 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12acc506-6412-4a9e-a80c-4b1cc761246d" containerName="nova-scheduler-scheduler" Nov 25 15:24:15 crc kubenswrapper[4890]: I1125 15:24:15.034857 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="12acc506-6412-4a9e-a80c-4b1cc761246d" containerName="nova-scheduler-scheduler" Nov 25 15:24:15 crc kubenswrapper[4890]: I1125 15:24:15.035062 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="12acc506-6412-4a9e-a80c-4b1cc761246d" containerName="nova-scheduler-scheduler" Nov 25 15:24:15 crc kubenswrapper[4890]: I1125 15:24:15.035764 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 15:24:15 crc kubenswrapper[4890]: I1125 15:24:15.040819 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 25 15:24:15 crc kubenswrapper[4890]: I1125 15:24:15.042783 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 15:24:15 crc kubenswrapper[4890]: E1125 15:24:15.189217 4890 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod12acc506_6412_4a9e_a80c_4b1cc761246d.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod12acc506_6412_4a9e_a80c_4b1cc761246d.slice/crio-1cc5a35a5b85d593f9a447a8791877cc01e4ee3875a3bf6f0782bcffc96ff089\": RecentStats: unable to find data in memory cache]" Nov 25 15:24:15 crc kubenswrapper[4890]: I1125 15:24:15.244234 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8df24e08-7452-4c77-9360-b555fdecf971-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8df24e08-7452-4c77-9360-b555fdecf971\") " pod="openstack/nova-scheduler-0" Nov 25 15:24:15 crc kubenswrapper[4890]: I1125 15:24:15.245471 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pq657\" (UniqueName: \"kubernetes.io/projected/8df24e08-7452-4c77-9360-b555fdecf971-kube-api-access-pq657\") pod \"nova-scheduler-0\" (UID: \"8df24e08-7452-4c77-9360-b555fdecf971\") " pod="openstack/nova-scheduler-0" Nov 25 15:24:15 crc kubenswrapper[4890]: I1125 15:24:15.245529 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8df24e08-7452-4c77-9360-b555fdecf971-config-data\") pod \"nova-scheduler-0\" (UID: \"8df24e08-7452-4c77-9360-b555fdecf971\") " pod="openstack/nova-scheduler-0" Nov 25 15:24:15 crc kubenswrapper[4890]: I1125 15:24:15.252986 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 15:24:15 crc kubenswrapper[4890]: I1125 15:24:15.347633 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pq657\" (UniqueName: \"kubernetes.io/projected/8df24e08-7452-4c77-9360-b555fdecf971-kube-api-access-pq657\") pod \"nova-scheduler-0\" (UID: \"8df24e08-7452-4c77-9360-b555fdecf971\") " pod="openstack/nova-scheduler-0" Nov 25 15:24:15 crc kubenswrapper[4890]: I1125 15:24:15.347719 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8df24e08-7452-4c77-9360-b555fdecf971-config-data\") pod \"nova-scheduler-0\" (UID: \"8df24e08-7452-4c77-9360-b555fdecf971\") " pod="openstack/nova-scheduler-0" Nov 25 15:24:15 crc kubenswrapper[4890]: I1125 15:24:15.347999 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8df24e08-7452-4c77-9360-b555fdecf971-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8df24e08-7452-4c77-9360-b555fdecf971\") " pod="openstack/nova-scheduler-0" Nov 25 15:24:15 crc kubenswrapper[4890]: I1125 15:24:15.354969 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8df24e08-7452-4c77-9360-b555fdecf971-config-data\") pod \"nova-scheduler-0\" (UID: \"8df24e08-7452-4c77-9360-b555fdecf971\") " pod="openstack/nova-scheduler-0" Nov 25 15:24:15 crc kubenswrapper[4890]: I1125 15:24:15.355311 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8df24e08-7452-4c77-9360-b555fdecf971-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8df24e08-7452-4c77-9360-b555fdecf971\") " pod="openstack/nova-scheduler-0" Nov 25 15:24:15 crc kubenswrapper[4890]: I1125 15:24:15.377744 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pq657\" (UniqueName: \"kubernetes.io/projected/8df24e08-7452-4c77-9360-b555fdecf971-kube-api-access-pq657\") pod \"nova-scheduler-0\" (UID: \"8df24e08-7452-4c77-9360-b555fdecf971\") " pod="openstack/nova-scheduler-0" Nov 25 15:24:15 crc kubenswrapper[4890]: I1125 15:24:15.659764 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 15:24:15 crc kubenswrapper[4890]: I1125 15:24:15.975759 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3c6b60ed-19f6-472a-92d6-2a1e06c92ea6","Type":"ContainerStarted","Data":"02bf310486d9ef307a1cc1780774c88ca56867a62df13d74a0b8d2243c199c6a"} Nov 25 15:24:15 crc kubenswrapper[4890]: I1125 15:24:15.976205 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3c6b60ed-19f6-472a-92d6-2a1e06c92ea6","Type":"ContainerStarted","Data":"3d765ed34f7a5c340169b6dd690452f07317e88d4ac0439f15789066b74989d5"} Nov 25 15:24:15 crc kubenswrapper[4890]: I1125 15:24:15.976223 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3c6b60ed-19f6-472a-92d6-2a1e06c92ea6","Type":"ContainerStarted","Data":"a92c74ddf7079eed402c68bacc109796742acaece31209fa35623e75f876e64e"} Nov 25 15:24:16 crc kubenswrapper[4890]: I1125 15:24:16.000001 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=1.999979395 podStartE2EDuration="1.999979395s" podCreationTimestamp="2025-11-25 15:24:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:24:15.99381844 +0000 UTC m=+1314.436281070" watchObservedRunningTime="2025-11-25 15:24:15.999979395 +0000 UTC m=+1314.442442005" Nov 25 15:24:16 crc kubenswrapper[4890]: I1125 15:24:16.149521 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 15:24:16 crc kubenswrapper[4890]: I1125 15:24:16.183278 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12acc506-6412-4a9e-a80c-4b1cc761246d" path="/var/lib/kubelet/pods/12acc506-6412-4a9e-a80c-4b1cc761246d/volumes" Nov 25 15:24:16 crc kubenswrapper[4890]: I1125 15:24:16.183955 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d89315fb-6d71-4209-a3ea-a46d6a8e7691" path="/var/lib/kubelet/pods/d89315fb-6d71-4209-a3ea-a46d6a8e7691/volumes" Nov 25 15:24:16 crc kubenswrapper[4890]: I1125 15:24:16.780584 4890 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="fb7922d1-e2c8-4ee7-a700-6e6353de7d91" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.196:8775/\": dial tcp 10.217.0.196:8775: connect: connection refused" Nov 25 15:24:16 crc kubenswrapper[4890]: I1125 15:24:16.780631 4890 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="fb7922d1-e2c8-4ee7-a700-6e6353de7d91" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.196:8775/\": dial tcp 10.217.0.196:8775: connect: connection refused" Nov 25 15:24:16 crc kubenswrapper[4890]: I1125 15:24:16.985924 4890 generic.go:334] "Generic (PLEG): container finished" podID="fb7922d1-e2c8-4ee7-a700-6e6353de7d91" containerID="53a207e706c9c4a6fe645567e7859d702ccd877ceb186ecc2155100a47808c7e" exitCode=0 Nov 25 15:24:16 crc kubenswrapper[4890]: I1125 15:24:16.985988 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fb7922d1-e2c8-4ee7-a700-6e6353de7d91","Type":"ContainerDied","Data":"53a207e706c9c4a6fe645567e7859d702ccd877ceb186ecc2155100a47808c7e"} Nov 25 15:24:16 crc kubenswrapper[4890]: I1125 15:24:16.988353 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8df24e08-7452-4c77-9360-b555fdecf971","Type":"ContainerStarted","Data":"ef9c5b79650382b13b3f6ceadebc39c2739ffa74590c68ee7c535a2f079080f0"} Nov 25 15:24:16 crc kubenswrapper[4890]: I1125 15:24:16.988381 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8df24e08-7452-4c77-9360-b555fdecf971","Type":"ContainerStarted","Data":"4176687e6ea442d9a7970805bbf70110bcc83bcf8175fcc152c9a48a67a97f60"} Nov 25 15:24:17 crc kubenswrapper[4890]: I1125 15:24:17.011132 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.011110545 podStartE2EDuration="3.011110545s" podCreationTimestamp="2025-11-25 15:24:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:24:17.002770945 +0000 UTC m=+1315.445233565" watchObservedRunningTime="2025-11-25 15:24:17.011110545 +0000 UTC m=+1315.453573155" Nov 25 15:24:17 crc kubenswrapper[4890]: I1125 15:24:17.395261 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 15:24:17 crc kubenswrapper[4890]: I1125 15:24:17.589023 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb7922d1-e2c8-4ee7-a700-6e6353de7d91-combined-ca-bundle\") pod \"fb7922d1-e2c8-4ee7-a700-6e6353de7d91\" (UID: \"fb7922d1-e2c8-4ee7-a700-6e6353de7d91\") " Nov 25 15:24:17 crc kubenswrapper[4890]: I1125 15:24:17.589081 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fsjjl\" (UniqueName: \"kubernetes.io/projected/fb7922d1-e2c8-4ee7-a700-6e6353de7d91-kube-api-access-fsjjl\") pod \"fb7922d1-e2c8-4ee7-a700-6e6353de7d91\" (UID: \"fb7922d1-e2c8-4ee7-a700-6e6353de7d91\") " Nov 25 15:24:17 crc kubenswrapper[4890]: I1125 15:24:17.589284 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb7922d1-e2c8-4ee7-a700-6e6353de7d91-nova-metadata-tls-certs\") pod \"fb7922d1-e2c8-4ee7-a700-6e6353de7d91\" (UID: \"fb7922d1-e2c8-4ee7-a700-6e6353de7d91\") " Nov 25 15:24:17 crc kubenswrapper[4890]: I1125 15:24:17.589374 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb7922d1-e2c8-4ee7-a700-6e6353de7d91-config-data\") pod \"fb7922d1-e2c8-4ee7-a700-6e6353de7d91\" (UID: \"fb7922d1-e2c8-4ee7-a700-6e6353de7d91\") " Nov 25 15:24:17 crc kubenswrapper[4890]: I1125 15:24:17.589473 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb7922d1-e2c8-4ee7-a700-6e6353de7d91-logs\") pod \"fb7922d1-e2c8-4ee7-a700-6e6353de7d91\" (UID: \"fb7922d1-e2c8-4ee7-a700-6e6353de7d91\") " Nov 25 15:24:17 crc kubenswrapper[4890]: I1125 15:24:17.590026 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb7922d1-e2c8-4ee7-a700-6e6353de7d91-logs" (OuterVolumeSpecName: "logs") pod "fb7922d1-e2c8-4ee7-a700-6e6353de7d91" (UID: "fb7922d1-e2c8-4ee7-a700-6e6353de7d91"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:24:17 crc kubenswrapper[4890]: I1125 15:24:17.601416 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb7922d1-e2c8-4ee7-a700-6e6353de7d91-kube-api-access-fsjjl" (OuterVolumeSpecName: "kube-api-access-fsjjl") pod "fb7922d1-e2c8-4ee7-a700-6e6353de7d91" (UID: "fb7922d1-e2c8-4ee7-a700-6e6353de7d91"). InnerVolumeSpecName "kube-api-access-fsjjl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:24:17 crc kubenswrapper[4890]: I1125 15:24:17.622768 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb7922d1-e2c8-4ee7-a700-6e6353de7d91-config-data" (OuterVolumeSpecName: "config-data") pod "fb7922d1-e2c8-4ee7-a700-6e6353de7d91" (UID: "fb7922d1-e2c8-4ee7-a700-6e6353de7d91"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:24:17 crc kubenswrapper[4890]: I1125 15:24:17.627786 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb7922d1-e2c8-4ee7-a700-6e6353de7d91-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fb7922d1-e2c8-4ee7-a700-6e6353de7d91" (UID: "fb7922d1-e2c8-4ee7-a700-6e6353de7d91"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:24:17 crc kubenswrapper[4890]: I1125 15:24:17.670778 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb7922d1-e2c8-4ee7-a700-6e6353de7d91-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "fb7922d1-e2c8-4ee7-a700-6e6353de7d91" (UID: "fb7922d1-e2c8-4ee7-a700-6e6353de7d91"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:24:17 crc kubenswrapper[4890]: I1125 15:24:17.692395 4890 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb7922d1-e2c8-4ee7-a700-6e6353de7d91-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:17 crc kubenswrapper[4890]: I1125 15:24:17.692626 4890 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb7922d1-e2c8-4ee7-a700-6e6353de7d91-logs\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:17 crc kubenswrapper[4890]: I1125 15:24:17.692641 4890 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb7922d1-e2c8-4ee7-a700-6e6353de7d91-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:17 crc kubenswrapper[4890]: I1125 15:24:17.692655 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fsjjl\" (UniqueName: \"kubernetes.io/projected/fb7922d1-e2c8-4ee7-a700-6e6353de7d91-kube-api-access-fsjjl\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:17 crc kubenswrapper[4890]: I1125 15:24:17.692667 4890 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb7922d1-e2c8-4ee7-a700-6e6353de7d91-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:18 crc kubenswrapper[4890]: I1125 15:24:18.014655 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 15:24:18 crc kubenswrapper[4890]: I1125 15:24:18.014668 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fb7922d1-e2c8-4ee7-a700-6e6353de7d91","Type":"ContainerDied","Data":"d0f33bdd3b9e241d0c07ec77857039ba1e888eeb76e5a385bfdc38caa75bc7ee"} Nov 25 15:24:18 crc kubenswrapper[4890]: I1125 15:24:18.014776 4890 scope.go:117] "RemoveContainer" containerID="53a207e706c9c4a6fe645567e7859d702ccd877ceb186ecc2155100a47808c7e" Nov 25 15:24:18 crc kubenswrapper[4890]: I1125 15:24:18.064337 4890 scope.go:117] "RemoveContainer" containerID="2b9bbbdc949ac4ef3283ade2efc3a174ce0f3ed26d413db474f54f62904e12d8" Nov 25 15:24:18 crc kubenswrapper[4890]: I1125 15:24:18.068511 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 15:24:18 crc kubenswrapper[4890]: I1125 15:24:18.075275 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 15:24:18 crc kubenswrapper[4890]: I1125 15:24:18.090641 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 25 15:24:18 crc kubenswrapper[4890]: E1125 15:24:18.091098 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb7922d1-e2c8-4ee7-a700-6e6353de7d91" containerName="nova-metadata-log" Nov 25 15:24:18 crc kubenswrapper[4890]: I1125 15:24:18.091115 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb7922d1-e2c8-4ee7-a700-6e6353de7d91" containerName="nova-metadata-log" Nov 25 15:24:18 crc kubenswrapper[4890]: E1125 15:24:18.091129 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb7922d1-e2c8-4ee7-a700-6e6353de7d91" containerName="nova-metadata-metadata" Nov 25 15:24:18 crc kubenswrapper[4890]: I1125 15:24:18.091136 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb7922d1-e2c8-4ee7-a700-6e6353de7d91" containerName="nova-metadata-metadata" Nov 25 15:24:18 crc kubenswrapper[4890]: I1125 15:24:18.091367 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb7922d1-e2c8-4ee7-a700-6e6353de7d91" containerName="nova-metadata-metadata" Nov 25 15:24:18 crc kubenswrapper[4890]: I1125 15:24:18.091402 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb7922d1-e2c8-4ee7-a700-6e6353de7d91" containerName="nova-metadata-log" Nov 25 15:24:18 crc kubenswrapper[4890]: I1125 15:24:18.093273 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 15:24:18 crc kubenswrapper[4890]: I1125 15:24:18.095536 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 25 15:24:18 crc kubenswrapper[4890]: I1125 15:24:18.095792 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 25 15:24:18 crc kubenswrapper[4890]: I1125 15:24:18.111258 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 15:24:18 crc kubenswrapper[4890]: I1125 15:24:18.187331 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb7922d1-e2c8-4ee7-a700-6e6353de7d91" path="/var/lib/kubelet/pods/fb7922d1-e2c8-4ee7-a700-6e6353de7d91/volumes" Nov 25 15:24:18 crc kubenswrapper[4890]: I1125 15:24:18.219677 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab6971fb-880d-4e68-96f6-bb23533105cd-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ab6971fb-880d-4e68-96f6-bb23533105cd\") " pod="openstack/nova-metadata-0" Nov 25 15:24:18 crc kubenswrapper[4890]: I1125 15:24:18.219929 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfx5v\" (UniqueName: \"kubernetes.io/projected/ab6971fb-880d-4e68-96f6-bb23533105cd-kube-api-access-xfx5v\") pod \"nova-metadata-0\" (UID: \"ab6971fb-880d-4e68-96f6-bb23533105cd\") " pod="openstack/nova-metadata-0" Nov 25 15:24:18 crc kubenswrapper[4890]: I1125 15:24:18.220022 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab6971fb-880d-4e68-96f6-bb23533105cd-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ab6971fb-880d-4e68-96f6-bb23533105cd\") " pod="openstack/nova-metadata-0" Nov 25 15:24:18 crc kubenswrapper[4890]: I1125 15:24:18.220122 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab6971fb-880d-4e68-96f6-bb23533105cd-config-data\") pod \"nova-metadata-0\" (UID: \"ab6971fb-880d-4e68-96f6-bb23533105cd\") " pod="openstack/nova-metadata-0" Nov 25 15:24:18 crc kubenswrapper[4890]: I1125 15:24:18.220147 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab6971fb-880d-4e68-96f6-bb23533105cd-logs\") pod \"nova-metadata-0\" (UID: \"ab6971fb-880d-4e68-96f6-bb23533105cd\") " pod="openstack/nova-metadata-0" Nov 25 15:24:18 crc kubenswrapper[4890]: I1125 15:24:18.322787 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab6971fb-880d-4e68-96f6-bb23533105cd-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ab6971fb-880d-4e68-96f6-bb23533105cd\") " pod="openstack/nova-metadata-0" Nov 25 15:24:18 crc kubenswrapper[4890]: I1125 15:24:18.322936 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab6971fb-880d-4e68-96f6-bb23533105cd-config-data\") pod \"nova-metadata-0\" (UID: \"ab6971fb-880d-4e68-96f6-bb23533105cd\") " pod="openstack/nova-metadata-0" Nov 25 15:24:18 crc kubenswrapper[4890]: I1125 15:24:18.322961 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab6971fb-880d-4e68-96f6-bb23533105cd-logs\") pod \"nova-metadata-0\" (UID: \"ab6971fb-880d-4e68-96f6-bb23533105cd\") " pod="openstack/nova-metadata-0" Nov 25 15:24:18 crc kubenswrapper[4890]: I1125 15:24:18.322995 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab6971fb-880d-4e68-96f6-bb23533105cd-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ab6971fb-880d-4e68-96f6-bb23533105cd\") " pod="openstack/nova-metadata-0" Nov 25 15:24:18 crc kubenswrapper[4890]: I1125 15:24:18.323062 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfx5v\" (UniqueName: \"kubernetes.io/projected/ab6971fb-880d-4e68-96f6-bb23533105cd-kube-api-access-xfx5v\") pod \"nova-metadata-0\" (UID: \"ab6971fb-880d-4e68-96f6-bb23533105cd\") " pod="openstack/nova-metadata-0" Nov 25 15:24:18 crc kubenswrapper[4890]: I1125 15:24:18.323864 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab6971fb-880d-4e68-96f6-bb23533105cd-logs\") pod \"nova-metadata-0\" (UID: \"ab6971fb-880d-4e68-96f6-bb23533105cd\") " pod="openstack/nova-metadata-0" Nov 25 15:24:18 crc kubenswrapper[4890]: I1125 15:24:18.328142 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab6971fb-880d-4e68-96f6-bb23533105cd-config-data\") pod \"nova-metadata-0\" (UID: \"ab6971fb-880d-4e68-96f6-bb23533105cd\") " pod="openstack/nova-metadata-0" Nov 25 15:24:18 crc kubenswrapper[4890]: I1125 15:24:18.328225 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab6971fb-880d-4e68-96f6-bb23533105cd-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ab6971fb-880d-4e68-96f6-bb23533105cd\") " pod="openstack/nova-metadata-0" Nov 25 15:24:18 crc kubenswrapper[4890]: I1125 15:24:18.329243 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab6971fb-880d-4e68-96f6-bb23533105cd-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ab6971fb-880d-4e68-96f6-bb23533105cd\") " pod="openstack/nova-metadata-0" Nov 25 15:24:18 crc kubenswrapper[4890]: I1125 15:24:18.341630 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfx5v\" (UniqueName: \"kubernetes.io/projected/ab6971fb-880d-4e68-96f6-bb23533105cd-kube-api-access-xfx5v\") pod \"nova-metadata-0\" (UID: \"ab6971fb-880d-4e68-96f6-bb23533105cd\") " pod="openstack/nova-metadata-0" Nov 25 15:24:18 crc kubenswrapper[4890]: I1125 15:24:18.416730 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 15:24:22 crc kubenswrapper[4890]: I1125 15:24:18.880053 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 15:24:22 crc kubenswrapper[4890]: I1125 15:24:19.026339 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ab6971fb-880d-4e68-96f6-bb23533105cd","Type":"ContainerStarted","Data":"4ca18201a47eec779dbca032bca569c3dc65917ecaa4a8c1d82d91787a1510f2"} Nov 25 15:24:22 crc kubenswrapper[4890]: I1125 15:24:20.042557 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ab6971fb-880d-4e68-96f6-bb23533105cd","Type":"ContainerStarted","Data":"7d8555e1e00560fd9ed89fd1c1ff3dc36afa99c090e0d8feed4592358954bdf5"} Nov 25 15:24:22 crc kubenswrapper[4890]: I1125 15:24:20.042864 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ab6971fb-880d-4e68-96f6-bb23533105cd","Type":"ContainerStarted","Data":"42e84d8f20c9ab61f6f6e5de8aef1b994910c78b8bf6c63efb437b6176e8fe4f"} Nov 25 15:24:22 crc kubenswrapper[4890]: I1125 15:24:20.067568 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.067548647 podStartE2EDuration="2.067548647s" podCreationTimestamp="2025-11-25 15:24:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:24:20.058444749 +0000 UTC m=+1318.500907359" watchObservedRunningTime="2025-11-25 15:24:20.067548647 +0000 UTC m=+1318.510011257" Nov 25 15:24:22 crc kubenswrapper[4890]: I1125 15:24:20.660960 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 25 15:24:23 crc kubenswrapper[4890]: I1125 15:24:23.417103 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 25 15:24:23 crc kubenswrapper[4890]: I1125 15:24:23.417584 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 25 15:24:24 crc kubenswrapper[4890]: I1125 15:24:24.731495 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 25 15:24:24 crc kubenswrapper[4890]: I1125 15:24:24.731876 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 25 15:24:25 crc kubenswrapper[4890]: I1125 15:24:25.661555 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 25 15:24:25 crc kubenswrapper[4890]: I1125 15:24:25.688116 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 25 15:24:25 crc kubenswrapper[4890]: I1125 15:24:25.746353 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3c6b60ed-19f6-472a-92d6-2a1e06c92ea6" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.205:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 15:24:25 crc kubenswrapper[4890]: I1125 15:24:25.746355 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3c6b60ed-19f6-472a-92d6-2a1e06c92ea6" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.205:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 15:24:26 crc kubenswrapper[4890]: I1125 15:24:26.129556 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 25 15:24:26 crc kubenswrapper[4890]: I1125 15:24:26.447128 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 15:24:26 crc kubenswrapper[4890]: I1125 15:24:26.447218 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 15:24:28 crc kubenswrapper[4890]: I1125 15:24:28.417677 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 25 15:24:28 crc kubenswrapper[4890]: I1125 15:24:28.419760 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 25 15:24:29 crc kubenswrapper[4890]: I1125 15:24:29.430327 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="ab6971fb-880d-4e68-96f6-bb23533105cd" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.207:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 15:24:29 crc kubenswrapper[4890]: I1125 15:24:29.430402 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="ab6971fb-880d-4e68-96f6-bb23533105cd" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.207:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 15:24:29 crc kubenswrapper[4890]: I1125 15:24:29.783711 4890 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="154b7e43-8329-4c7d-9870-ffca704e80af" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.195:3000/\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 25 15:24:31 crc kubenswrapper[4890]: I1125 15:24:31.166025 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 25 15:24:34 crc kubenswrapper[4890]: I1125 15:24:34.737903 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 25 15:24:34 crc kubenswrapper[4890]: I1125 15:24:34.738420 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 25 15:24:34 crc kubenswrapper[4890]: I1125 15:24:34.738577 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 25 15:24:34 crc kubenswrapper[4890]: I1125 15:24:34.738615 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 25 15:24:34 crc kubenswrapper[4890]: I1125 15:24:34.746304 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 25 15:24:34 crc kubenswrapper[4890]: I1125 15:24:34.747937 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 25 15:24:37 crc kubenswrapper[4890]: I1125 15:24:37.868450 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 25 15:24:37 crc kubenswrapper[4890]: I1125 15:24:37.870372 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 15:24:37 crc kubenswrapper[4890]: I1125 15:24:37.876219 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 25 15:24:37 crc kubenswrapper[4890]: I1125 15:24:37.876233 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 25 15:24:37 crc kubenswrapper[4890]: I1125 15:24:37.879593 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 25 15:24:38 crc kubenswrapper[4890]: I1125 15:24:38.025104 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/93c59329-d272-468f-800d-56c2ae34fe50-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"93c59329-d272-468f-800d-56c2ae34fe50\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 15:24:38 crc kubenswrapper[4890]: I1125 15:24:38.025155 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/93c59329-d272-468f-800d-56c2ae34fe50-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"93c59329-d272-468f-800d-56c2ae34fe50\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 15:24:38 crc kubenswrapper[4890]: I1125 15:24:38.127779 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/93c59329-d272-468f-800d-56c2ae34fe50-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"93c59329-d272-468f-800d-56c2ae34fe50\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 15:24:38 crc kubenswrapper[4890]: I1125 15:24:38.127856 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/93c59329-d272-468f-800d-56c2ae34fe50-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"93c59329-d272-468f-800d-56c2ae34fe50\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 15:24:38 crc kubenswrapper[4890]: I1125 15:24:38.128121 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/93c59329-d272-468f-800d-56c2ae34fe50-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"93c59329-d272-468f-800d-56c2ae34fe50\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 15:24:38 crc kubenswrapper[4890]: I1125 15:24:38.152555 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/93c59329-d272-468f-800d-56c2ae34fe50-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"93c59329-d272-468f-800d-56c2ae34fe50\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 15:24:38 crc kubenswrapper[4890]: I1125 15:24:38.206308 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 15:24:38 crc kubenswrapper[4890]: I1125 15:24:38.424846 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 25 15:24:38 crc kubenswrapper[4890]: I1125 15:24:38.425635 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 25 15:24:38 crc kubenswrapper[4890]: I1125 15:24:38.434724 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 25 15:24:38 crc kubenswrapper[4890]: I1125 15:24:38.441493 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 25 15:24:38 crc kubenswrapper[4890]: I1125 15:24:38.810785 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 25 15:24:38 crc kubenswrapper[4890]: W1125 15:24:38.813769 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod93c59329_d272_468f_800d_56c2ae34fe50.slice/crio-72a4545161a5da128d5849fb9ad619db50a353ebad5027eab796250c02559186 WatchSource:0}: Error finding container 72a4545161a5da128d5849fb9ad619db50a353ebad5027eab796250c02559186: Status 404 returned error can't find the container with id 72a4545161a5da128d5849fb9ad619db50a353ebad5027eab796250c02559186 Nov 25 15:24:39 crc kubenswrapper[4890]: I1125 15:24:39.260335 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"93c59329-d272-468f-800d-56c2ae34fe50","Type":"ContainerStarted","Data":"62bfccfc2562190f42aef8b53c639d0d276f91bd01401797332b1123cc99d97f"} Nov 25 15:24:39 crc kubenswrapper[4890]: I1125 15:24:39.260389 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"93c59329-d272-468f-800d-56c2ae34fe50","Type":"ContainerStarted","Data":"72a4545161a5da128d5849fb9ad619db50a353ebad5027eab796250c02559186"} Nov 25 15:24:39 crc kubenswrapper[4890]: I1125 15:24:39.279706 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=2.2796886179999998 podStartE2EDuration="2.279688618s" podCreationTimestamp="2025-11-25 15:24:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:24:39.271863381 +0000 UTC m=+1337.714326001" watchObservedRunningTime="2025-11-25 15:24:39.279688618 +0000 UTC m=+1337.722151228" Nov 25 15:24:40 crc kubenswrapper[4890]: I1125 15:24:40.269466 4890 generic.go:334] "Generic (PLEG): container finished" podID="93c59329-d272-468f-800d-56c2ae34fe50" containerID="62bfccfc2562190f42aef8b53c639d0d276f91bd01401797332b1123cc99d97f" exitCode=0 Nov 25 15:24:40 crc kubenswrapper[4890]: I1125 15:24:40.269556 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"93c59329-d272-468f-800d-56c2ae34fe50","Type":"ContainerDied","Data":"62bfccfc2562190f42aef8b53c639d0d276f91bd01401797332b1123cc99d97f"} Nov 25 15:24:41 crc kubenswrapper[4890]: I1125 15:24:41.573587 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 15:24:41 crc kubenswrapper[4890]: I1125 15:24:41.703853 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/93c59329-d272-468f-800d-56c2ae34fe50-kube-api-access\") pod \"93c59329-d272-468f-800d-56c2ae34fe50\" (UID: \"93c59329-d272-468f-800d-56c2ae34fe50\") " Nov 25 15:24:41 crc kubenswrapper[4890]: I1125 15:24:41.703916 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/93c59329-d272-468f-800d-56c2ae34fe50-kubelet-dir\") pod \"93c59329-d272-468f-800d-56c2ae34fe50\" (UID: \"93c59329-d272-468f-800d-56c2ae34fe50\") " Nov 25 15:24:41 crc kubenswrapper[4890]: I1125 15:24:41.704050 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93c59329-d272-468f-800d-56c2ae34fe50-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "93c59329-d272-468f-800d-56c2ae34fe50" (UID: "93c59329-d272-468f-800d-56c2ae34fe50"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 15:24:41 crc kubenswrapper[4890]: I1125 15:24:41.704433 4890 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/93c59329-d272-468f-800d-56c2ae34fe50-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:41 crc kubenswrapper[4890]: I1125 15:24:41.711069 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93c59329-d272-468f-800d-56c2ae34fe50-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "93c59329-d272-468f-800d-56c2ae34fe50" (UID: "93c59329-d272-468f-800d-56c2ae34fe50"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:24:41 crc kubenswrapper[4890]: I1125 15:24:41.806358 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/93c59329-d272-468f-800d-56c2ae34fe50-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:42 crc kubenswrapper[4890]: I1125 15:24:42.290765 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"93c59329-d272-468f-800d-56c2ae34fe50","Type":"ContainerDied","Data":"72a4545161a5da128d5849fb9ad619db50a353ebad5027eab796250c02559186"} Nov 25 15:24:42 crc kubenswrapper[4890]: I1125 15:24:42.291077 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="72a4545161a5da128d5849fb9ad619db50a353ebad5027eab796250c02559186" Nov 25 15:24:42 crc kubenswrapper[4890]: I1125 15:24:42.290956 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 15:24:42 crc kubenswrapper[4890]: I1125 15:24:42.871259 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 25 15:24:42 crc kubenswrapper[4890]: E1125 15:24:42.871673 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93c59329-d272-468f-800d-56c2ae34fe50" containerName="pruner" Nov 25 15:24:42 crc kubenswrapper[4890]: I1125 15:24:42.871685 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="93c59329-d272-468f-800d-56c2ae34fe50" containerName="pruner" Nov 25 15:24:42 crc kubenswrapper[4890]: I1125 15:24:42.871864 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="93c59329-d272-468f-800d-56c2ae34fe50" containerName="pruner" Nov 25 15:24:42 crc kubenswrapper[4890]: I1125 15:24:42.872497 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 25 15:24:42 crc kubenswrapper[4890]: I1125 15:24:42.877539 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 25 15:24:42 crc kubenswrapper[4890]: I1125 15:24:42.877669 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 25 15:24:42 crc kubenswrapper[4890]: I1125 15:24:42.878926 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 25 15:24:42 crc kubenswrapper[4890]: I1125 15:24:42.926511 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/118dcc72-87d4-4aa3-9397-bee2c90eac63-var-lock\") pod \"installer-9-crc\" (UID: \"118dcc72-87d4-4aa3-9397-bee2c90eac63\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 15:24:42 crc kubenswrapper[4890]: I1125 15:24:42.926727 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/118dcc72-87d4-4aa3-9397-bee2c90eac63-kube-api-access\") pod \"installer-9-crc\" (UID: \"118dcc72-87d4-4aa3-9397-bee2c90eac63\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 15:24:42 crc kubenswrapper[4890]: I1125 15:24:42.926779 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/118dcc72-87d4-4aa3-9397-bee2c90eac63-kubelet-dir\") pod \"installer-9-crc\" (UID: \"118dcc72-87d4-4aa3-9397-bee2c90eac63\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 15:24:43 crc kubenswrapper[4890]: I1125 15:24:43.027947 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/118dcc72-87d4-4aa3-9397-bee2c90eac63-var-lock\") pod \"installer-9-crc\" (UID: \"118dcc72-87d4-4aa3-9397-bee2c90eac63\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 15:24:43 crc kubenswrapper[4890]: I1125 15:24:43.028266 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/118dcc72-87d4-4aa3-9397-bee2c90eac63-kube-api-access\") pod \"installer-9-crc\" (UID: \"118dcc72-87d4-4aa3-9397-bee2c90eac63\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 15:24:43 crc kubenswrapper[4890]: I1125 15:24:43.028367 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/118dcc72-87d4-4aa3-9397-bee2c90eac63-kubelet-dir\") pod \"installer-9-crc\" (UID: \"118dcc72-87d4-4aa3-9397-bee2c90eac63\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 15:24:43 crc kubenswrapper[4890]: I1125 15:24:43.028417 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/118dcc72-87d4-4aa3-9397-bee2c90eac63-kubelet-dir\") pod \"installer-9-crc\" (UID: \"118dcc72-87d4-4aa3-9397-bee2c90eac63\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 15:24:43 crc kubenswrapper[4890]: I1125 15:24:43.028133 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/118dcc72-87d4-4aa3-9397-bee2c90eac63-var-lock\") pod \"installer-9-crc\" (UID: \"118dcc72-87d4-4aa3-9397-bee2c90eac63\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 15:24:43 crc kubenswrapper[4890]: I1125 15:24:43.046937 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/118dcc72-87d4-4aa3-9397-bee2c90eac63-kube-api-access\") pod \"installer-9-crc\" (UID: \"118dcc72-87d4-4aa3-9397-bee2c90eac63\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 15:24:43 crc kubenswrapper[4890]: I1125 15:24:43.190882 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 25 15:24:43 crc kubenswrapper[4890]: I1125 15:24:43.624552 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 25 15:24:43 crc kubenswrapper[4890]: W1125 15:24:43.630122 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod118dcc72_87d4_4aa3_9397_bee2c90eac63.slice/crio-8a69b7be4b9d7dd54ecf1d02bfd1eb59fa73661fd7b8c3260e8ea8fce79a0121 WatchSource:0}: Error finding container 8a69b7be4b9d7dd54ecf1d02bfd1eb59fa73661fd7b8c3260e8ea8fce79a0121: Status 404 returned error can't find the container with id 8a69b7be4b9d7dd54ecf1d02bfd1eb59fa73661fd7b8c3260e8ea8fce79a0121 Nov 25 15:24:44 crc kubenswrapper[4890]: I1125 15:24:44.311469 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"118dcc72-87d4-4aa3-9397-bee2c90eac63","Type":"ContainerStarted","Data":"8a69b7be4b9d7dd54ecf1d02bfd1eb59fa73661fd7b8c3260e8ea8fce79a0121"} Nov 25 15:24:45 crc kubenswrapper[4890]: I1125 15:24:45.322779 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"118dcc72-87d4-4aa3-9397-bee2c90eac63","Type":"ContainerStarted","Data":"b8ffb3bdae25d8f8daa4f0a2463144167f9c688376c66531c3818c5e17fe1bcd"} Nov 25 15:24:45 crc kubenswrapper[4890]: I1125 15:24:45.338212 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=3.338190836 podStartE2EDuration="3.338190836s" podCreationTimestamp="2025-11-25 15:24:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:24:45.337626881 +0000 UTC m=+1343.780089511" watchObservedRunningTime="2025-11-25 15:24:45.338190836 +0000 UTC m=+1343.780653456" Nov 25 15:24:47 crc kubenswrapper[4890]: I1125 15:24:47.158084 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 15:24:47 crc kubenswrapper[4890]: I1125 15:24:47.923638 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 15:24:51 crc kubenswrapper[4890]: I1125 15:24:51.430653 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="ae7f09bf-3fc2-4da7-9b41-9f4564dd7562" containerName="rabbitmq" containerID="cri-o://4d7577d6f8303eed0d77102268bf3e3dd24a0be05f90274612e2dc2f3373322d" gracePeriod=604796 Nov 25 15:24:51 crc kubenswrapper[4890]: I1125 15:24:51.770237 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="c14ce73e-1a95-4f41-b6b6-33478079806f" containerName="rabbitmq" containerID="cri-o://f5a82450da67557d7cb5285c4f393e18175e4939d9428d0b24e1f390918f2959" gracePeriod=604797 Nov 25 15:24:55 crc kubenswrapper[4890]: I1125 15:24:55.492893 4890 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="ae7f09bf-3fc2-4da7-9b41-9f4564dd7562" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.101:5671: connect: connection refused" Nov 25 15:24:55 crc kubenswrapper[4890]: I1125 15:24:55.786473 4890 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="c14ce73e-1a95-4f41-b6b6-33478079806f" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.102:5671: connect: connection refused" Nov 25 15:24:56 crc kubenswrapper[4890]: I1125 15:24:56.447862 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 15:24:56 crc kubenswrapper[4890]: I1125 15:24:56.448249 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 15:24:56 crc kubenswrapper[4890]: I1125 15:24:56.448302 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" Nov 25 15:24:56 crc kubenswrapper[4890]: I1125 15:24:56.449141 4890 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"325b4a907d6b6c6f59e5653baa1b3273925ff29fed8f729e95af24a885f5e532"} pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 15:24:56 crc kubenswrapper[4890]: I1125 15:24:56.449302 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" containerID="cri-o://325b4a907d6b6c6f59e5653baa1b3273925ff29fed8f729e95af24a885f5e532" gracePeriod=600 Nov 25 15:24:57 crc kubenswrapper[4890]: I1125 15:24:57.427039 4890 generic.go:334] "Generic (PLEG): container finished" podID="4e4f849d-f239-4727-a73e-18327856929a" containerID="325b4a907d6b6c6f59e5653baa1b3273925ff29fed8f729e95af24a885f5e532" exitCode=0 Nov 25 15:24:57 crc kubenswrapper[4890]: I1125 15:24:57.427104 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" event={"ID":"4e4f849d-f239-4727-a73e-18327856929a","Type":"ContainerDied","Data":"325b4a907d6b6c6f59e5653baa1b3273925ff29fed8f729e95af24a885f5e532"} Nov 25 15:24:57 crc kubenswrapper[4890]: I1125 15:24:57.427437 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" event={"ID":"4e4f849d-f239-4727-a73e-18327856929a","Type":"ContainerStarted","Data":"9467aa8edc4d8bfe461dc02826750091f343018e37e7f269d305c77d21e832da"} Nov 25 15:24:57 crc kubenswrapper[4890]: I1125 15:24:57.427466 4890 scope.go:117] "RemoveContainer" containerID="53708c37db2dded329d6f7de3f6f52a10c460a5fc2cf14a1b264e8539a50f494" Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.451749 4890 generic.go:334] "Generic (PLEG): container finished" podID="c14ce73e-1a95-4f41-b6b6-33478079806f" containerID="f5a82450da67557d7cb5285c4f393e18175e4939d9428d0b24e1f390918f2959" exitCode=0 Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.452092 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c14ce73e-1a95-4f41-b6b6-33478079806f","Type":"ContainerDied","Data":"f5a82450da67557d7cb5285c4f393e18175e4939d9428d0b24e1f390918f2959"} Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.458240 4890 generic.go:334] "Generic (PLEG): container finished" podID="ae7f09bf-3fc2-4da7-9b41-9f4564dd7562" containerID="4d7577d6f8303eed0d77102268bf3e3dd24a0be05f90274612e2dc2f3373322d" exitCode=0 Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.458285 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562","Type":"ContainerDied","Data":"4d7577d6f8303eed0d77102268bf3e3dd24a0be05f90274612e2dc2f3373322d"} Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.680131 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.686997 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.856239 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-plugins-conf\") pod \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\" (UID: \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\") " Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.856311 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c14ce73e-1a95-4f41-b6b6-33478079806f-config-data\") pod \"c14ce73e-1a95-4f41-b6b6-33478079806f\" (UID: \"c14ce73e-1a95-4f41-b6b6-33478079806f\") " Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.856341 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c14ce73e-1a95-4f41-b6b6-33478079806f-pod-info\") pod \"c14ce73e-1a95-4f41-b6b6-33478079806f\" (UID: \"c14ce73e-1a95-4f41-b6b6-33478079806f\") " Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.856370 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-rabbitmq-tls\") pod \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\" (UID: \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\") " Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.856404 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-rabbitmq-erlang-cookie\") pod \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\" (UID: \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\") " Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.856465 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wbr45\" (UniqueName: \"kubernetes.io/projected/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-kube-api-access-wbr45\") pod \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\" (UID: \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\") " Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.856970 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "ae7f09bf-3fc2-4da7-9b41-9f4564dd7562" (UID: "ae7f09bf-3fc2-4da7-9b41-9f4564dd7562"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.857131 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "ae7f09bf-3fc2-4da7-9b41-9f4564dd7562" (UID: "ae7f09bf-3fc2-4da7-9b41-9f4564dd7562"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.856570 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-rabbitmq-confd\") pod \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\" (UID: \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\") " Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.857450 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c14ce73e-1a95-4f41-b6b6-33478079806f-rabbitmq-tls\") pod \"c14ce73e-1a95-4f41-b6b6-33478079806f\" (UID: \"c14ce73e-1a95-4f41-b6b6-33478079806f\") " Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.857512 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c14ce73e-1a95-4f41-b6b6-33478079806f-server-conf\") pod \"c14ce73e-1a95-4f41-b6b6-33478079806f\" (UID: \"c14ce73e-1a95-4f41-b6b6-33478079806f\") " Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.857540 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-server-conf\") pod \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\" (UID: \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\") " Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.857574 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c14ce73e-1a95-4f41-b6b6-33478079806f-rabbitmq-erlang-cookie\") pod \"c14ce73e-1a95-4f41-b6b6-33478079806f\" (UID: \"c14ce73e-1a95-4f41-b6b6-33478079806f\") " Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.857598 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"c14ce73e-1a95-4f41-b6b6-33478079806f\" (UID: \"c14ce73e-1a95-4f41-b6b6-33478079806f\") " Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.857625 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-rabbitmq-plugins\") pod \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\" (UID: \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\") " Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.857662 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c14ce73e-1a95-4f41-b6b6-33478079806f-rabbitmq-confd\") pod \"c14ce73e-1a95-4f41-b6b6-33478079806f\" (UID: \"c14ce73e-1a95-4f41-b6b6-33478079806f\") " Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.857687 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c14ce73e-1a95-4f41-b6b6-33478079806f-rabbitmq-plugins\") pod \"c14ce73e-1a95-4f41-b6b6-33478079806f\" (UID: \"c14ce73e-1a95-4f41-b6b6-33478079806f\") " Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.857727 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-config-data\") pod \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\" (UID: \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\") " Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.857768 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-erlang-cookie-secret\") pod \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\" (UID: \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\") " Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.857792 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c14ce73e-1a95-4f41-b6b6-33478079806f-erlang-cookie-secret\") pod \"c14ce73e-1a95-4f41-b6b6-33478079806f\" (UID: \"c14ce73e-1a95-4f41-b6b6-33478079806f\") " Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.857815 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\" (UID: \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\") " Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.857838 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tcxvg\" (UniqueName: \"kubernetes.io/projected/c14ce73e-1a95-4f41-b6b6-33478079806f-kube-api-access-tcxvg\") pod \"c14ce73e-1a95-4f41-b6b6-33478079806f\" (UID: \"c14ce73e-1a95-4f41-b6b6-33478079806f\") " Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.857882 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c14ce73e-1a95-4f41-b6b6-33478079806f-plugins-conf\") pod \"c14ce73e-1a95-4f41-b6b6-33478079806f\" (UID: \"c14ce73e-1a95-4f41-b6b6-33478079806f\") " Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.857913 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-pod-info\") pod \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\" (UID: \"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562\") " Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.858821 4890 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.858848 4890 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.869458 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c14ce73e-1a95-4f41-b6b6-33478079806f-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "c14ce73e-1a95-4f41-b6b6-33478079806f" (UID: "c14ce73e-1a95-4f41-b6b6-33478079806f"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.869730 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c14ce73e-1a95-4f41-b6b6-33478079806f-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "c14ce73e-1a95-4f41-b6b6-33478079806f" (UID: "c14ce73e-1a95-4f41-b6b6-33478079806f"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.871482 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-kube-api-access-wbr45" (OuterVolumeSpecName: "kube-api-access-wbr45") pod "ae7f09bf-3fc2-4da7-9b41-9f4564dd7562" (UID: "ae7f09bf-3fc2-4da7-9b41-9f4564dd7562"). InnerVolumeSpecName "kube-api-access-wbr45". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.872013 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c14ce73e-1a95-4f41-b6b6-33478079806f-kube-api-access-tcxvg" (OuterVolumeSpecName: "kube-api-access-tcxvg") pod "c14ce73e-1a95-4f41-b6b6-33478079806f" (UID: "c14ce73e-1a95-4f41-b6b6-33478079806f"). InnerVolumeSpecName "kube-api-access-tcxvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.874805 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c14ce73e-1a95-4f41-b6b6-33478079806f-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "c14ce73e-1a95-4f41-b6b6-33478079806f" (UID: "c14ce73e-1a95-4f41-b6b6-33478079806f"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.874892 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-pod-info" (OuterVolumeSpecName: "pod-info") pod "ae7f09bf-3fc2-4da7-9b41-9f4564dd7562" (UID: "ae7f09bf-3fc2-4da7-9b41-9f4564dd7562"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.879926 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "ae7f09bf-3fc2-4da7-9b41-9f4564dd7562" (UID: "ae7f09bf-3fc2-4da7-9b41-9f4564dd7562"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.882703 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "ae7f09bf-3fc2-4da7-9b41-9f4564dd7562" (UID: "ae7f09bf-3fc2-4da7-9b41-9f4564dd7562"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.882890 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/c14ce73e-1a95-4f41-b6b6-33478079806f-pod-info" (OuterVolumeSpecName: "pod-info") pod "c14ce73e-1a95-4f41-b6b6-33478079806f" (UID: "c14ce73e-1a95-4f41-b6b6-33478079806f"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.896387 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c14ce73e-1a95-4f41-b6b6-33478079806f-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "c14ce73e-1a95-4f41-b6b6-33478079806f" (UID: "c14ce73e-1a95-4f41-b6b6-33478079806f"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.896537 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "persistence") pod "ae7f09bf-3fc2-4da7-9b41-9f4564dd7562" (UID: "ae7f09bf-3fc2-4da7-9b41-9f4564dd7562"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.896677 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "ae7f09bf-3fc2-4da7-9b41-9f4564dd7562" (UID: "ae7f09bf-3fc2-4da7-9b41-9f4564dd7562"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.901473 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c14ce73e-1a95-4f41-b6b6-33478079806f-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "c14ce73e-1a95-4f41-b6b6-33478079806f" (UID: "c14ce73e-1a95-4f41-b6b6-33478079806f"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.914295 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "persistence") pod "c14ce73e-1a95-4f41-b6b6-33478079806f" (UID: "c14ce73e-1a95-4f41-b6b6-33478079806f"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.960401 4890 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c14ce73e-1a95-4f41-b6b6-33478079806f-pod-info\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.960435 4890 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.960449 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wbr45\" (UniqueName: \"kubernetes.io/projected/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-kube-api-access-wbr45\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.960461 4890 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c14ce73e-1a95-4f41-b6b6-33478079806f-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.960475 4890 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c14ce73e-1a95-4f41-b6b6-33478079806f-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.960506 4890 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.960520 4890 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.960532 4890 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c14ce73e-1a95-4f41-b6b6-33478079806f-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.960543 4890 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.960552 4890 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c14ce73e-1a95-4f41-b6b6-33478079806f-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.960568 4890 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.960577 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tcxvg\" (UniqueName: \"kubernetes.io/projected/c14ce73e-1a95-4f41-b6b6-33478079806f-kube-api-access-tcxvg\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.960588 4890 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c14ce73e-1a95-4f41-b6b6-33478079806f-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:58 crc kubenswrapper[4890]: I1125 15:24:58.960597 4890 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-pod-info\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.025589 4890 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.066701 4890 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.067900 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-config-data" (OuterVolumeSpecName: "config-data") pod "ae7f09bf-3fc2-4da7-9b41-9f4564dd7562" (UID: "ae7f09bf-3fc2-4da7-9b41-9f4564dd7562"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.078750 4890 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.080137 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c14ce73e-1a95-4f41-b6b6-33478079806f-config-data" (OuterVolumeSpecName: "config-data") pod "c14ce73e-1a95-4f41-b6b6-33478079806f" (UID: "c14ce73e-1a95-4f41-b6b6-33478079806f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.119817 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c14ce73e-1a95-4f41-b6b6-33478079806f-server-conf" (OuterVolumeSpecName: "server-conf") pod "c14ce73e-1a95-4f41-b6b6-33478079806f" (UID: "c14ce73e-1a95-4f41-b6b6-33478079806f"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.130892 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-server-conf" (OuterVolumeSpecName: "server-conf") pod "ae7f09bf-3fc2-4da7-9b41-9f4564dd7562" (UID: "ae7f09bf-3fc2-4da7-9b41-9f4564dd7562"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.169454 4890 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.169486 4890 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.169495 4890 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c14ce73e-1a95-4f41-b6b6-33478079806f-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.169503 4890 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c14ce73e-1a95-4f41-b6b6-33478079806f-server-conf\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.169513 4890 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-server-conf\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.181345 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "ae7f09bf-3fc2-4da7-9b41-9f4564dd7562" (UID: "ae7f09bf-3fc2-4da7-9b41-9f4564dd7562"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.243777 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c14ce73e-1a95-4f41-b6b6-33478079806f-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "c14ce73e-1a95-4f41-b6b6-33478079806f" (UID: "c14ce73e-1a95-4f41-b6b6-33478079806f"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.271529 4890 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.271573 4890 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c14ce73e-1a95-4f41-b6b6-33478079806f-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.469818 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.469883 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ae7f09bf-3fc2-4da7-9b41-9f4564dd7562","Type":"ContainerDied","Data":"c4b27d4bef264cb091befd2683b8d338325300986c622b7abf1025f011730555"} Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.469985 4890 scope.go:117] "RemoveContainer" containerID="4d7577d6f8303eed0d77102268bf3e3dd24a0be05f90274612e2dc2f3373322d" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.471940 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c14ce73e-1a95-4f41-b6b6-33478079806f","Type":"ContainerDied","Data":"251d05656390cb8a8d176016a4b9a2e1bdd2cb455eac2988bd187eb2e89c8336"} Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.472029 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.514739 4890 scope.go:117] "RemoveContainer" containerID="3c3e2348928c66cad8da81580a88530958cd3237ca6204733fd5e8d2343fdb71" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.546994 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.559241 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.587307 4890 scope.go:117] "RemoveContainer" containerID="f5a82450da67557d7cb5285c4f393e18175e4939d9428d0b24e1f390918f2959" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.625358 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.639590 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.644384 4890 scope.go:117] "RemoveContainer" containerID="7b0b47083ae2bc3b6521792dbfc2606a74d4a262bfbeca4040d61b4873018ae8" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.655478 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 15:24:59 crc kubenswrapper[4890]: E1125 15:24:59.655948 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c14ce73e-1a95-4f41-b6b6-33478079806f" containerName="rabbitmq" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.655968 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="c14ce73e-1a95-4f41-b6b6-33478079806f" containerName="rabbitmq" Nov 25 15:24:59 crc kubenswrapper[4890]: E1125 15:24:59.656015 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae7f09bf-3fc2-4da7-9b41-9f4564dd7562" containerName="rabbitmq" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.656023 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae7f09bf-3fc2-4da7-9b41-9f4564dd7562" containerName="rabbitmq" Nov 25 15:24:59 crc kubenswrapper[4890]: E1125 15:24:59.656033 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c14ce73e-1a95-4f41-b6b6-33478079806f" containerName="setup-container" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.656042 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="c14ce73e-1a95-4f41-b6b6-33478079806f" containerName="setup-container" Nov 25 15:24:59 crc kubenswrapper[4890]: E1125 15:24:59.656057 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae7f09bf-3fc2-4da7-9b41-9f4564dd7562" containerName="setup-container" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.656064 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae7f09bf-3fc2-4da7-9b41-9f4564dd7562" containerName="setup-container" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.656292 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="c14ce73e-1a95-4f41-b6b6-33478079806f" containerName="rabbitmq" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.656326 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae7f09bf-3fc2-4da7-9b41-9f4564dd7562" containerName="rabbitmq" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.657613 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.661899 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.662338 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.662580 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-2k6hj" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.662877 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.663114 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.663425 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.665236 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.665345 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.683211 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.685098 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.688504 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.688711 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.688820 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.688921 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.689022 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-9cgrd" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.689517 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.692061 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.692747 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.798407 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9bab37b3-681b-4cd5-9fee-503afa0d6d22-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"9bab37b3-681b-4cd5-9fee-503afa0d6d22\") " pod="openstack/rabbitmq-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.798460 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4x5v\" (UniqueName: \"kubernetes.io/projected/d08ea973-c3f2-4192-8bfc-b025f0c5391b-kube-api-access-v4x5v\") pod \"rabbitmq-cell1-server-0\" (UID: \"d08ea973-c3f2-4192-8bfc-b025f0c5391b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.798487 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9bab37b3-681b-4cd5-9fee-503afa0d6d22-config-data\") pod \"rabbitmq-server-0\" (UID: \"9bab37b3-681b-4cd5-9fee-503afa0d6d22\") " pod="openstack/rabbitmq-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.798509 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"9bab37b3-681b-4cd5-9fee-503afa0d6d22\") " pod="openstack/rabbitmq-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.798616 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d08ea973-c3f2-4192-8bfc-b025f0c5391b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"d08ea973-c3f2-4192-8bfc-b025f0c5391b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.798713 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpqrr\" (UniqueName: \"kubernetes.io/projected/9bab37b3-681b-4cd5-9fee-503afa0d6d22-kube-api-access-tpqrr\") pod \"rabbitmq-server-0\" (UID: \"9bab37b3-681b-4cd5-9fee-503afa0d6d22\") " pod="openstack/rabbitmq-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.798792 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9bab37b3-681b-4cd5-9fee-503afa0d6d22-server-conf\") pod \"rabbitmq-server-0\" (UID: \"9bab37b3-681b-4cd5-9fee-503afa0d6d22\") " pod="openstack/rabbitmq-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.798817 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d08ea973-c3f2-4192-8bfc-b025f0c5391b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d08ea973-c3f2-4192-8bfc-b025f0c5391b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.798833 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d08ea973-c3f2-4192-8bfc-b025f0c5391b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"d08ea973-c3f2-4192-8bfc-b025f0c5391b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.798986 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d08ea973-c3f2-4192-8bfc-b025f0c5391b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d08ea973-c3f2-4192-8bfc-b025f0c5391b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.799044 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d08ea973-c3f2-4192-8bfc-b025f0c5391b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"d08ea973-c3f2-4192-8bfc-b025f0c5391b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.799147 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9bab37b3-681b-4cd5-9fee-503afa0d6d22-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"9bab37b3-681b-4cd5-9fee-503afa0d6d22\") " pod="openstack/rabbitmq-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.799238 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9bab37b3-681b-4cd5-9fee-503afa0d6d22-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"9bab37b3-681b-4cd5-9fee-503afa0d6d22\") " pod="openstack/rabbitmq-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.799262 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9bab37b3-681b-4cd5-9fee-503afa0d6d22-pod-info\") pod \"rabbitmq-server-0\" (UID: \"9bab37b3-681b-4cd5-9fee-503afa0d6d22\") " pod="openstack/rabbitmq-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.799301 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9bab37b3-681b-4cd5-9fee-503afa0d6d22-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"9bab37b3-681b-4cd5-9fee-503afa0d6d22\") " pod="openstack/rabbitmq-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.799362 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d08ea973-c3f2-4192-8bfc-b025f0c5391b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"d08ea973-c3f2-4192-8bfc-b025f0c5391b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.799430 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9bab37b3-681b-4cd5-9fee-503afa0d6d22-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"9bab37b3-681b-4cd5-9fee-503afa0d6d22\") " pod="openstack/rabbitmq-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.799481 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d08ea973-c3f2-4192-8bfc-b025f0c5391b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.799514 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9bab37b3-681b-4cd5-9fee-503afa0d6d22-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"9bab37b3-681b-4cd5-9fee-503afa0d6d22\") " pod="openstack/rabbitmq-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.799585 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d08ea973-c3f2-4192-8bfc-b025f0c5391b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"d08ea973-c3f2-4192-8bfc-b025f0c5391b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.799618 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d08ea973-c3f2-4192-8bfc-b025f0c5391b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"d08ea973-c3f2-4192-8bfc-b025f0c5391b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.799719 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d08ea973-c3f2-4192-8bfc-b025f0c5391b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"d08ea973-c3f2-4192-8bfc-b025f0c5391b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.901925 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9bab37b3-681b-4cd5-9fee-503afa0d6d22-server-conf\") pod \"rabbitmq-server-0\" (UID: \"9bab37b3-681b-4cd5-9fee-503afa0d6d22\") " pod="openstack/rabbitmq-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.901982 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d08ea973-c3f2-4192-8bfc-b025f0c5391b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d08ea973-c3f2-4192-8bfc-b025f0c5391b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.902009 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d08ea973-c3f2-4192-8bfc-b025f0c5391b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"d08ea973-c3f2-4192-8bfc-b025f0c5391b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.902043 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d08ea973-c3f2-4192-8bfc-b025f0c5391b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d08ea973-c3f2-4192-8bfc-b025f0c5391b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.902065 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d08ea973-c3f2-4192-8bfc-b025f0c5391b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"d08ea973-c3f2-4192-8bfc-b025f0c5391b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.902101 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9bab37b3-681b-4cd5-9fee-503afa0d6d22-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"9bab37b3-681b-4cd5-9fee-503afa0d6d22\") " pod="openstack/rabbitmq-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.902127 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9bab37b3-681b-4cd5-9fee-503afa0d6d22-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"9bab37b3-681b-4cd5-9fee-503afa0d6d22\") " pod="openstack/rabbitmq-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.902143 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9bab37b3-681b-4cd5-9fee-503afa0d6d22-pod-info\") pod \"rabbitmq-server-0\" (UID: \"9bab37b3-681b-4cd5-9fee-503afa0d6d22\") " pod="openstack/rabbitmq-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.902218 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9bab37b3-681b-4cd5-9fee-503afa0d6d22-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"9bab37b3-681b-4cd5-9fee-503afa0d6d22\") " pod="openstack/rabbitmq-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.902256 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d08ea973-c3f2-4192-8bfc-b025f0c5391b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"d08ea973-c3f2-4192-8bfc-b025f0c5391b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.902286 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9bab37b3-681b-4cd5-9fee-503afa0d6d22-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"9bab37b3-681b-4cd5-9fee-503afa0d6d22\") " pod="openstack/rabbitmq-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.902319 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d08ea973-c3f2-4192-8bfc-b025f0c5391b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.902341 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9bab37b3-681b-4cd5-9fee-503afa0d6d22-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"9bab37b3-681b-4cd5-9fee-503afa0d6d22\") " pod="openstack/rabbitmq-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.902392 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d08ea973-c3f2-4192-8bfc-b025f0c5391b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"d08ea973-c3f2-4192-8bfc-b025f0c5391b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.902417 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d08ea973-c3f2-4192-8bfc-b025f0c5391b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"d08ea973-c3f2-4192-8bfc-b025f0c5391b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.902449 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d08ea973-c3f2-4192-8bfc-b025f0c5391b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"d08ea973-c3f2-4192-8bfc-b025f0c5391b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.902493 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9bab37b3-681b-4cd5-9fee-503afa0d6d22-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"9bab37b3-681b-4cd5-9fee-503afa0d6d22\") " pod="openstack/rabbitmq-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.902517 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4x5v\" (UniqueName: \"kubernetes.io/projected/d08ea973-c3f2-4192-8bfc-b025f0c5391b-kube-api-access-v4x5v\") pod \"rabbitmq-cell1-server-0\" (UID: \"d08ea973-c3f2-4192-8bfc-b025f0c5391b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.902539 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9bab37b3-681b-4cd5-9fee-503afa0d6d22-config-data\") pod \"rabbitmq-server-0\" (UID: \"9bab37b3-681b-4cd5-9fee-503afa0d6d22\") " pod="openstack/rabbitmq-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.902560 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"9bab37b3-681b-4cd5-9fee-503afa0d6d22\") " pod="openstack/rabbitmq-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.902596 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d08ea973-c3f2-4192-8bfc-b025f0c5391b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"d08ea973-c3f2-4192-8bfc-b025f0c5391b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.902640 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpqrr\" (UniqueName: \"kubernetes.io/projected/9bab37b3-681b-4cd5-9fee-503afa0d6d22-kube-api-access-tpqrr\") pod \"rabbitmq-server-0\" (UID: \"9bab37b3-681b-4cd5-9fee-503afa0d6d22\") " pod="openstack/rabbitmq-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.902693 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9bab37b3-681b-4cd5-9fee-503afa0d6d22-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"9bab37b3-681b-4cd5-9fee-503afa0d6d22\") " pod="openstack/rabbitmq-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.903144 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d08ea973-c3f2-4192-8bfc-b025f0c5391b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d08ea973-c3f2-4192-8bfc-b025f0c5391b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.903299 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9bab37b3-681b-4cd5-9fee-503afa0d6d22-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"9bab37b3-681b-4cd5-9fee-503afa0d6d22\") " pod="openstack/rabbitmq-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.903504 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d08ea973-c3f2-4192-8bfc-b025f0c5391b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"d08ea973-c3f2-4192-8bfc-b025f0c5391b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.903529 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9bab37b3-681b-4cd5-9fee-503afa0d6d22-server-conf\") pod \"rabbitmq-server-0\" (UID: \"9bab37b3-681b-4cd5-9fee-503afa0d6d22\") " pod="openstack/rabbitmq-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.903580 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d08ea973-c3f2-4192-8bfc-b025f0c5391b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"d08ea973-c3f2-4192-8bfc-b025f0c5391b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.903802 4890 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"9bab37b3-681b-4cd5-9fee-503afa0d6d22\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.903929 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9bab37b3-681b-4cd5-9fee-503afa0d6d22-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"9bab37b3-681b-4cd5-9fee-503afa0d6d22\") " pod="openstack/rabbitmq-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.904050 4890 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d08ea973-c3f2-4192-8bfc-b025f0c5391b\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.904156 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9bab37b3-681b-4cd5-9fee-503afa0d6d22-config-data\") pod \"rabbitmq-server-0\" (UID: \"9bab37b3-681b-4cd5-9fee-503afa0d6d22\") " pod="openstack/rabbitmq-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.904496 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d08ea973-c3f2-4192-8bfc-b025f0c5391b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"d08ea973-c3f2-4192-8bfc-b025f0c5391b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.904651 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d08ea973-c3f2-4192-8bfc-b025f0c5391b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d08ea973-c3f2-4192-8bfc-b025f0c5391b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.910443 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d08ea973-c3f2-4192-8bfc-b025f0c5391b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"d08ea973-c3f2-4192-8bfc-b025f0c5391b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.910443 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9bab37b3-681b-4cd5-9fee-503afa0d6d22-pod-info\") pod \"rabbitmq-server-0\" (UID: \"9bab37b3-681b-4cd5-9fee-503afa0d6d22\") " pod="openstack/rabbitmq-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.913422 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d08ea973-c3f2-4192-8bfc-b025f0c5391b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"d08ea973-c3f2-4192-8bfc-b025f0c5391b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.913449 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d08ea973-c3f2-4192-8bfc-b025f0c5391b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"d08ea973-c3f2-4192-8bfc-b025f0c5391b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.916863 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9bab37b3-681b-4cd5-9fee-503afa0d6d22-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"9bab37b3-681b-4cd5-9fee-503afa0d6d22\") " pod="openstack/rabbitmq-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.917235 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d08ea973-c3f2-4192-8bfc-b025f0c5391b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"d08ea973-c3f2-4192-8bfc-b025f0c5391b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.923701 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4x5v\" (UniqueName: \"kubernetes.io/projected/d08ea973-c3f2-4192-8bfc-b025f0c5391b-kube-api-access-v4x5v\") pod \"rabbitmq-cell1-server-0\" (UID: \"d08ea973-c3f2-4192-8bfc-b025f0c5391b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.923794 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9bab37b3-681b-4cd5-9fee-503afa0d6d22-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"9bab37b3-681b-4cd5-9fee-503afa0d6d22\") " pod="openstack/rabbitmq-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.924091 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9bab37b3-681b-4cd5-9fee-503afa0d6d22-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"9bab37b3-681b-4cd5-9fee-503afa0d6d22\") " pod="openstack/rabbitmq-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.932250 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpqrr\" (UniqueName: \"kubernetes.io/projected/9bab37b3-681b-4cd5-9fee-503afa0d6d22-kube-api-access-tpqrr\") pod \"rabbitmq-server-0\" (UID: \"9bab37b3-681b-4cd5-9fee-503afa0d6d22\") " pod="openstack/rabbitmq-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.949066 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"9bab37b3-681b-4cd5-9fee-503afa0d6d22\") " pod="openstack/rabbitmq-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.951775 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d08ea973-c3f2-4192-8bfc-b025f0c5391b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:24:59 crc kubenswrapper[4890]: I1125 15:24:59.990817 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:25:00 crc kubenswrapper[4890]: I1125 15:25:00.018828 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 25 15:25:00 crc kubenswrapper[4890]: I1125 15:25:00.193111 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae7f09bf-3fc2-4da7-9b41-9f4564dd7562" path="/var/lib/kubelet/pods/ae7f09bf-3fc2-4da7-9b41-9f4564dd7562/volumes" Nov 25 15:25:00 crc kubenswrapper[4890]: I1125 15:25:00.193929 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c14ce73e-1a95-4f41-b6b6-33478079806f" path="/var/lib/kubelet/pods/c14ce73e-1a95-4f41-b6b6-33478079806f/volumes" Nov 25 15:25:00 crc kubenswrapper[4890]: W1125 15:25:00.518469 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd08ea973_c3f2_4192_8bfc_b025f0c5391b.slice/crio-1cdd1fc2bd9dc10ab4e6ba5dc49d1bc3c15323210be138d6900f47606e5a2b15 WatchSource:0}: Error finding container 1cdd1fc2bd9dc10ab4e6ba5dc49d1bc3c15323210be138d6900f47606e5a2b15: Status 404 returned error can't find the container with id 1cdd1fc2bd9dc10ab4e6ba5dc49d1bc3c15323210be138d6900f47606e5a2b15 Nov 25 15:25:00 crc kubenswrapper[4890]: I1125 15:25:00.529072 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 15:25:00 crc kubenswrapper[4890]: I1125 15:25:00.541666 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 15:25:01 crc kubenswrapper[4890]: I1125 15:25:01.496032 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d08ea973-c3f2-4192-8bfc-b025f0c5391b","Type":"ContainerStarted","Data":"1cdd1fc2bd9dc10ab4e6ba5dc49d1bc3c15323210be138d6900f47606e5a2b15"} Nov 25 15:25:01 crc kubenswrapper[4890]: I1125 15:25:01.499814 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9bab37b3-681b-4cd5-9fee-503afa0d6d22","Type":"ContainerStarted","Data":"c1869f87c61a19513332dc63fb34ea8a846f548b70aea40720352fe8cba67fac"} Nov 25 15:25:02 crc kubenswrapper[4890]: I1125 15:25:02.522570 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9bab37b3-681b-4cd5-9fee-503afa0d6d22","Type":"ContainerStarted","Data":"b4e01c87ce56f03bf70dea18143d788a675a2a28fda3721303f87885ade5d9df"} Nov 25 15:25:02 crc kubenswrapper[4890]: I1125 15:25:02.559272 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d08ea973-c3f2-4192-8bfc-b025f0c5391b","Type":"ContainerStarted","Data":"8bb3adcfe9f7b378153065acb08cd345e9a3a2e085d8241903165dc9dd5a4297"} Nov 25 15:25:02 crc kubenswrapper[4890]: I1125 15:25:02.723575 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-jq72w"] Nov 25 15:25:02 crc kubenswrapper[4890]: I1125 15:25:02.725915 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-jq72w" Nov 25 15:25:02 crc kubenswrapper[4890]: I1125 15:25:02.732789 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Nov 25 15:25:02 crc kubenswrapper[4890]: I1125 15:25:02.751237 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-jq72w"] Nov 25 15:25:02 crc kubenswrapper[4890]: I1125 15:25:02.788150 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a6d1d8ff-4e18-4625-b712-20cb460839a0-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-jq72w\" (UID: \"a6d1d8ff-4e18-4625-b712-20cb460839a0\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-jq72w" Nov 25 15:25:02 crc kubenswrapper[4890]: I1125 15:25:02.791132 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a6d1d8ff-4e18-4625-b712-20cb460839a0-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-jq72w\" (UID: \"a6d1d8ff-4e18-4625-b712-20cb460839a0\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-jq72w" Nov 25 15:25:02 crc kubenswrapper[4890]: I1125 15:25:02.791613 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a6d1d8ff-4e18-4625-b712-20cb460839a0-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-jq72w\" (UID: \"a6d1d8ff-4e18-4625-b712-20cb460839a0\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-jq72w" Nov 25 15:25:02 crc kubenswrapper[4890]: I1125 15:25:02.791929 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a6d1d8ff-4e18-4625-b712-20cb460839a0-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-jq72w\" (UID: \"a6d1d8ff-4e18-4625-b712-20cb460839a0\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-jq72w" Nov 25 15:25:02 crc kubenswrapper[4890]: I1125 15:25:02.792136 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6d1d8ff-4e18-4625-b712-20cb460839a0-config\") pod \"dnsmasq-dns-79bd4cc8c9-jq72w\" (UID: \"a6d1d8ff-4e18-4625-b712-20cb460839a0\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-jq72w" Nov 25 15:25:02 crc kubenswrapper[4890]: I1125 15:25:02.792329 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gt2b4\" (UniqueName: \"kubernetes.io/projected/a6d1d8ff-4e18-4625-b712-20cb460839a0-kube-api-access-gt2b4\") pod \"dnsmasq-dns-79bd4cc8c9-jq72w\" (UID: \"a6d1d8ff-4e18-4625-b712-20cb460839a0\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-jq72w" Nov 25 15:25:02 crc kubenswrapper[4890]: I1125 15:25:02.794322 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a6d1d8ff-4e18-4625-b712-20cb460839a0-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-jq72w\" (UID: \"a6d1d8ff-4e18-4625-b712-20cb460839a0\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-jq72w" Nov 25 15:25:02 crc kubenswrapper[4890]: I1125 15:25:02.897870 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gt2b4\" (UniqueName: \"kubernetes.io/projected/a6d1d8ff-4e18-4625-b712-20cb460839a0-kube-api-access-gt2b4\") pod \"dnsmasq-dns-79bd4cc8c9-jq72w\" (UID: \"a6d1d8ff-4e18-4625-b712-20cb460839a0\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-jq72w" Nov 25 15:25:02 crc kubenswrapper[4890]: I1125 15:25:02.898745 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a6d1d8ff-4e18-4625-b712-20cb460839a0-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-jq72w\" (UID: \"a6d1d8ff-4e18-4625-b712-20cb460839a0\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-jq72w" Nov 25 15:25:02 crc kubenswrapper[4890]: I1125 15:25:02.899201 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a6d1d8ff-4e18-4625-b712-20cb460839a0-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-jq72w\" (UID: \"a6d1d8ff-4e18-4625-b712-20cb460839a0\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-jq72w" Nov 25 15:25:02 crc kubenswrapper[4890]: I1125 15:25:02.899738 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a6d1d8ff-4e18-4625-b712-20cb460839a0-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-jq72w\" (UID: \"a6d1d8ff-4e18-4625-b712-20cb460839a0\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-jq72w" Nov 25 15:25:02 crc kubenswrapper[4890]: I1125 15:25:02.900423 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a6d1d8ff-4e18-4625-b712-20cb460839a0-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-jq72w\" (UID: \"a6d1d8ff-4e18-4625-b712-20cb460839a0\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-jq72w" Nov 25 15:25:02 crc kubenswrapper[4890]: I1125 15:25:02.901103 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a6d1d8ff-4e18-4625-b712-20cb460839a0-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-jq72w\" (UID: \"a6d1d8ff-4e18-4625-b712-20cb460839a0\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-jq72w" Nov 25 15:25:02 crc kubenswrapper[4890]: I1125 15:25:02.901748 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6d1d8ff-4e18-4625-b712-20cb460839a0-config\") pod \"dnsmasq-dns-79bd4cc8c9-jq72w\" (UID: \"a6d1d8ff-4e18-4625-b712-20cb460839a0\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-jq72w" Nov 25 15:25:02 crc kubenswrapper[4890]: I1125 15:25:02.901839 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a6d1d8ff-4e18-4625-b712-20cb460839a0-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-jq72w\" (UID: \"a6d1d8ff-4e18-4625-b712-20cb460839a0\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-jq72w" Nov 25 15:25:02 crc kubenswrapper[4890]: I1125 15:25:02.900347 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a6d1d8ff-4e18-4625-b712-20cb460839a0-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-jq72w\" (UID: \"a6d1d8ff-4e18-4625-b712-20cb460839a0\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-jq72w" Nov 25 15:25:02 crc kubenswrapper[4890]: I1125 15:25:02.901575 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a6d1d8ff-4e18-4625-b712-20cb460839a0-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-jq72w\" (UID: \"a6d1d8ff-4e18-4625-b712-20cb460839a0\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-jq72w" Nov 25 15:25:02 crc kubenswrapper[4890]: I1125 15:25:02.900234 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a6d1d8ff-4e18-4625-b712-20cb460839a0-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-jq72w\" (UID: \"a6d1d8ff-4e18-4625-b712-20cb460839a0\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-jq72w" Nov 25 15:25:02 crc kubenswrapper[4890]: I1125 15:25:02.901010 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a6d1d8ff-4e18-4625-b712-20cb460839a0-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-jq72w\" (UID: \"a6d1d8ff-4e18-4625-b712-20cb460839a0\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-jq72w" Nov 25 15:25:02 crc kubenswrapper[4890]: I1125 15:25:02.902398 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6d1d8ff-4e18-4625-b712-20cb460839a0-config\") pod \"dnsmasq-dns-79bd4cc8c9-jq72w\" (UID: \"a6d1d8ff-4e18-4625-b712-20cb460839a0\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-jq72w" Nov 25 15:25:02 crc kubenswrapper[4890]: I1125 15:25:02.929440 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gt2b4\" (UniqueName: \"kubernetes.io/projected/a6d1d8ff-4e18-4625-b712-20cb460839a0-kube-api-access-gt2b4\") pod \"dnsmasq-dns-79bd4cc8c9-jq72w\" (UID: \"a6d1d8ff-4e18-4625-b712-20cb460839a0\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-jq72w" Nov 25 15:25:03 crc kubenswrapper[4890]: I1125 15:25:03.000418 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-jq72w"] Nov 25 15:25:03 crc kubenswrapper[4890]: I1125 15:25:03.001799 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-jq72w" Nov 25 15:25:03 crc kubenswrapper[4890]: I1125 15:25:03.039656 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55478c4467-j8b46"] Nov 25 15:25:03 crc kubenswrapper[4890]: I1125 15:25:03.041769 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55478c4467-j8b46" Nov 25 15:25:03 crc kubenswrapper[4890]: I1125 15:25:03.058470 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55478c4467-j8b46"] Nov 25 15:25:03 crc kubenswrapper[4890]: I1125 15:25:03.111764 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d2eef50-1148-4cd4-8b76-67dc29bf8846-dns-svc\") pod \"dnsmasq-dns-55478c4467-j8b46\" (UID: \"2d2eef50-1148-4cd4-8b76-67dc29bf8846\") " pod="openstack/dnsmasq-dns-55478c4467-j8b46" Nov 25 15:25:03 crc kubenswrapper[4890]: I1125 15:25:03.111882 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2d2eef50-1148-4cd4-8b76-67dc29bf8846-ovsdbserver-sb\") pod \"dnsmasq-dns-55478c4467-j8b46\" (UID: \"2d2eef50-1148-4cd4-8b76-67dc29bf8846\") " pod="openstack/dnsmasq-dns-55478c4467-j8b46" Nov 25 15:25:03 crc kubenswrapper[4890]: I1125 15:25:03.111928 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/2d2eef50-1148-4cd4-8b76-67dc29bf8846-openstack-edpm-ipam\") pod \"dnsmasq-dns-55478c4467-j8b46\" (UID: \"2d2eef50-1148-4cd4-8b76-67dc29bf8846\") " pod="openstack/dnsmasq-dns-55478c4467-j8b46" Nov 25 15:25:03 crc kubenswrapper[4890]: I1125 15:25:03.111959 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86sgg\" (UniqueName: \"kubernetes.io/projected/2d2eef50-1148-4cd4-8b76-67dc29bf8846-kube-api-access-86sgg\") pod \"dnsmasq-dns-55478c4467-j8b46\" (UID: \"2d2eef50-1148-4cd4-8b76-67dc29bf8846\") " pod="openstack/dnsmasq-dns-55478c4467-j8b46" Nov 25 15:25:03 crc kubenswrapper[4890]: I1125 15:25:03.112011 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d2eef50-1148-4cd4-8b76-67dc29bf8846-config\") pod \"dnsmasq-dns-55478c4467-j8b46\" (UID: \"2d2eef50-1148-4cd4-8b76-67dc29bf8846\") " pod="openstack/dnsmasq-dns-55478c4467-j8b46" Nov 25 15:25:03 crc kubenswrapper[4890]: I1125 15:25:03.112066 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2d2eef50-1148-4cd4-8b76-67dc29bf8846-dns-swift-storage-0\") pod \"dnsmasq-dns-55478c4467-j8b46\" (UID: \"2d2eef50-1148-4cd4-8b76-67dc29bf8846\") " pod="openstack/dnsmasq-dns-55478c4467-j8b46" Nov 25 15:25:03 crc kubenswrapper[4890]: I1125 15:25:03.112118 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2d2eef50-1148-4cd4-8b76-67dc29bf8846-ovsdbserver-nb\") pod \"dnsmasq-dns-55478c4467-j8b46\" (UID: \"2d2eef50-1148-4cd4-8b76-67dc29bf8846\") " pod="openstack/dnsmasq-dns-55478c4467-j8b46" Nov 25 15:25:03 crc kubenswrapper[4890]: I1125 15:25:03.214632 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d2eef50-1148-4cd4-8b76-67dc29bf8846-dns-svc\") pod \"dnsmasq-dns-55478c4467-j8b46\" (UID: \"2d2eef50-1148-4cd4-8b76-67dc29bf8846\") " pod="openstack/dnsmasq-dns-55478c4467-j8b46" Nov 25 15:25:03 crc kubenswrapper[4890]: I1125 15:25:03.215904 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d2eef50-1148-4cd4-8b76-67dc29bf8846-dns-svc\") pod \"dnsmasq-dns-55478c4467-j8b46\" (UID: \"2d2eef50-1148-4cd4-8b76-67dc29bf8846\") " pod="openstack/dnsmasq-dns-55478c4467-j8b46" Nov 25 15:25:03 crc kubenswrapper[4890]: I1125 15:25:03.216313 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2d2eef50-1148-4cd4-8b76-67dc29bf8846-ovsdbserver-sb\") pod \"dnsmasq-dns-55478c4467-j8b46\" (UID: \"2d2eef50-1148-4cd4-8b76-67dc29bf8846\") " pod="openstack/dnsmasq-dns-55478c4467-j8b46" Nov 25 15:25:03 crc kubenswrapper[4890]: I1125 15:25:03.229178 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2d2eef50-1148-4cd4-8b76-67dc29bf8846-ovsdbserver-sb\") pod \"dnsmasq-dns-55478c4467-j8b46\" (UID: \"2d2eef50-1148-4cd4-8b76-67dc29bf8846\") " pod="openstack/dnsmasq-dns-55478c4467-j8b46" Nov 25 15:25:03 crc kubenswrapper[4890]: I1125 15:25:03.229296 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/2d2eef50-1148-4cd4-8b76-67dc29bf8846-openstack-edpm-ipam\") pod \"dnsmasq-dns-55478c4467-j8b46\" (UID: \"2d2eef50-1148-4cd4-8b76-67dc29bf8846\") " pod="openstack/dnsmasq-dns-55478c4467-j8b46" Nov 25 15:25:03 crc kubenswrapper[4890]: I1125 15:25:03.229337 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86sgg\" (UniqueName: \"kubernetes.io/projected/2d2eef50-1148-4cd4-8b76-67dc29bf8846-kube-api-access-86sgg\") pod \"dnsmasq-dns-55478c4467-j8b46\" (UID: \"2d2eef50-1148-4cd4-8b76-67dc29bf8846\") " pod="openstack/dnsmasq-dns-55478c4467-j8b46" Nov 25 15:25:03 crc kubenswrapper[4890]: I1125 15:25:03.229392 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d2eef50-1148-4cd4-8b76-67dc29bf8846-config\") pod \"dnsmasq-dns-55478c4467-j8b46\" (UID: \"2d2eef50-1148-4cd4-8b76-67dc29bf8846\") " pod="openstack/dnsmasq-dns-55478c4467-j8b46" Nov 25 15:25:03 crc kubenswrapper[4890]: I1125 15:25:03.229473 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2d2eef50-1148-4cd4-8b76-67dc29bf8846-dns-swift-storage-0\") pod \"dnsmasq-dns-55478c4467-j8b46\" (UID: \"2d2eef50-1148-4cd4-8b76-67dc29bf8846\") " pod="openstack/dnsmasq-dns-55478c4467-j8b46" Nov 25 15:25:03 crc kubenswrapper[4890]: I1125 15:25:03.229553 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2d2eef50-1148-4cd4-8b76-67dc29bf8846-ovsdbserver-nb\") pod \"dnsmasq-dns-55478c4467-j8b46\" (UID: \"2d2eef50-1148-4cd4-8b76-67dc29bf8846\") " pod="openstack/dnsmasq-dns-55478c4467-j8b46" Nov 25 15:25:03 crc kubenswrapper[4890]: I1125 15:25:03.230341 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2d2eef50-1148-4cd4-8b76-67dc29bf8846-ovsdbserver-nb\") pod \"dnsmasq-dns-55478c4467-j8b46\" (UID: \"2d2eef50-1148-4cd4-8b76-67dc29bf8846\") " pod="openstack/dnsmasq-dns-55478c4467-j8b46" Nov 25 15:25:03 crc kubenswrapper[4890]: I1125 15:25:03.231822 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/2d2eef50-1148-4cd4-8b76-67dc29bf8846-openstack-edpm-ipam\") pod \"dnsmasq-dns-55478c4467-j8b46\" (UID: \"2d2eef50-1148-4cd4-8b76-67dc29bf8846\") " pod="openstack/dnsmasq-dns-55478c4467-j8b46" Nov 25 15:25:03 crc kubenswrapper[4890]: I1125 15:25:03.232036 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2d2eef50-1148-4cd4-8b76-67dc29bf8846-dns-swift-storage-0\") pod \"dnsmasq-dns-55478c4467-j8b46\" (UID: \"2d2eef50-1148-4cd4-8b76-67dc29bf8846\") " pod="openstack/dnsmasq-dns-55478c4467-j8b46" Nov 25 15:25:03 crc kubenswrapper[4890]: I1125 15:25:03.235241 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d2eef50-1148-4cd4-8b76-67dc29bf8846-config\") pod \"dnsmasq-dns-55478c4467-j8b46\" (UID: \"2d2eef50-1148-4cd4-8b76-67dc29bf8846\") " pod="openstack/dnsmasq-dns-55478c4467-j8b46" Nov 25 15:25:03 crc kubenswrapper[4890]: I1125 15:25:03.249867 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86sgg\" (UniqueName: \"kubernetes.io/projected/2d2eef50-1148-4cd4-8b76-67dc29bf8846-kube-api-access-86sgg\") pod \"dnsmasq-dns-55478c4467-j8b46\" (UID: \"2d2eef50-1148-4cd4-8b76-67dc29bf8846\") " pod="openstack/dnsmasq-dns-55478c4467-j8b46" Nov 25 15:25:03 crc kubenswrapper[4890]: I1125 15:25:03.465845 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55478c4467-j8b46" Nov 25 15:25:03 crc kubenswrapper[4890]: I1125 15:25:03.533601 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-jq72w"] Nov 25 15:25:03 crc kubenswrapper[4890]: W1125 15:25:03.539395 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6d1d8ff_4e18_4625_b712_20cb460839a0.slice/crio-cf5e5b001776857ee38b097ff13b8fd5ff4c84bffd407e52efdc8354f75d84d0 WatchSource:0}: Error finding container cf5e5b001776857ee38b097ff13b8fd5ff4c84bffd407e52efdc8354f75d84d0: Status 404 returned error can't find the container with id cf5e5b001776857ee38b097ff13b8fd5ff4c84bffd407e52efdc8354f75d84d0 Nov 25 15:25:03 crc kubenswrapper[4890]: I1125 15:25:03.572576 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-jq72w" event={"ID":"a6d1d8ff-4e18-4625-b712-20cb460839a0","Type":"ContainerStarted","Data":"cf5e5b001776857ee38b097ff13b8fd5ff4c84bffd407e52efdc8354f75d84d0"} Nov 25 15:25:04 crc kubenswrapper[4890]: I1125 15:25:04.015013 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55478c4467-j8b46"] Nov 25 15:25:04 crc kubenswrapper[4890]: I1125 15:25:04.582086 4890 generic.go:334] "Generic (PLEG): container finished" podID="a6d1d8ff-4e18-4625-b712-20cb460839a0" containerID="6cc52c6733b3eaf4423055d2b7bbf1585ef1118c2034f65833bfc1000185d14a" exitCode=0 Nov 25 15:25:04 crc kubenswrapper[4890]: I1125 15:25:04.582177 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-jq72w" event={"ID":"a6d1d8ff-4e18-4625-b712-20cb460839a0","Type":"ContainerDied","Data":"6cc52c6733b3eaf4423055d2b7bbf1585ef1118c2034f65833bfc1000185d14a"} Nov 25 15:25:04 crc kubenswrapper[4890]: I1125 15:25:04.586986 4890 generic.go:334] "Generic (PLEG): container finished" podID="2d2eef50-1148-4cd4-8b76-67dc29bf8846" containerID="93ad76ad487bed7fb447a68b06c0f6146e7d5fe52aac8b402b897a5769e5384e" exitCode=0 Nov 25 15:25:04 crc kubenswrapper[4890]: I1125 15:25:04.587045 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55478c4467-j8b46" event={"ID":"2d2eef50-1148-4cd4-8b76-67dc29bf8846","Type":"ContainerDied","Data":"93ad76ad487bed7fb447a68b06c0f6146e7d5fe52aac8b402b897a5769e5384e"} Nov 25 15:25:04 crc kubenswrapper[4890]: I1125 15:25:04.587074 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55478c4467-j8b46" event={"ID":"2d2eef50-1148-4cd4-8b76-67dc29bf8846","Type":"ContainerStarted","Data":"38e1a093edda3248bd5a34baf95e89ab52b142d0380d5cd0ce9c838143c52a26"} Nov 25 15:25:04 crc kubenswrapper[4890]: I1125 15:25:04.886062 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-jq72w" Nov 25 15:25:04 crc kubenswrapper[4890]: I1125 15:25:04.967407 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gt2b4\" (UniqueName: \"kubernetes.io/projected/a6d1d8ff-4e18-4625-b712-20cb460839a0-kube-api-access-gt2b4\") pod \"a6d1d8ff-4e18-4625-b712-20cb460839a0\" (UID: \"a6d1d8ff-4e18-4625-b712-20cb460839a0\") " Nov 25 15:25:04 crc kubenswrapper[4890]: I1125 15:25:04.967545 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a6d1d8ff-4e18-4625-b712-20cb460839a0-dns-swift-storage-0\") pod \"a6d1d8ff-4e18-4625-b712-20cb460839a0\" (UID: \"a6d1d8ff-4e18-4625-b712-20cb460839a0\") " Nov 25 15:25:04 crc kubenswrapper[4890]: I1125 15:25:04.967581 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6d1d8ff-4e18-4625-b712-20cb460839a0-config\") pod \"a6d1d8ff-4e18-4625-b712-20cb460839a0\" (UID: \"a6d1d8ff-4e18-4625-b712-20cb460839a0\") " Nov 25 15:25:04 crc kubenswrapper[4890]: I1125 15:25:04.967617 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a6d1d8ff-4e18-4625-b712-20cb460839a0-dns-svc\") pod \"a6d1d8ff-4e18-4625-b712-20cb460839a0\" (UID: \"a6d1d8ff-4e18-4625-b712-20cb460839a0\") " Nov 25 15:25:04 crc kubenswrapper[4890]: I1125 15:25:04.967661 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a6d1d8ff-4e18-4625-b712-20cb460839a0-ovsdbserver-nb\") pod \"a6d1d8ff-4e18-4625-b712-20cb460839a0\" (UID: \"a6d1d8ff-4e18-4625-b712-20cb460839a0\") " Nov 25 15:25:04 crc kubenswrapper[4890]: I1125 15:25:04.967710 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a6d1d8ff-4e18-4625-b712-20cb460839a0-ovsdbserver-sb\") pod \"a6d1d8ff-4e18-4625-b712-20cb460839a0\" (UID: \"a6d1d8ff-4e18-4625-b712-20cb460839a0\") " Nov 25 15:25:04 crc kubenswrapper[4890]: I1125 15:25:04.967790 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a6d1d8ff-4e18-4625-b712-20cb460839a0-openstack-edpm-ipam\") pod \"a6d1d8ff-4e18-4625-b712-20cb460839a0\" (UID: \"a6d1d8ff-4e18-4625-b712-20cb460839a0\") " Nov 25 15:25:04 crc kubenswrapper[4890]: I1125 15:25:04.977938 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6d1d8ff-4e18-4625-b712-20cb460839a0-kube-api-access-gt2b4" (OuterVolumeSpecName: "kube-api-access-gt2b4") pod "a6d1d8ff-4e18-4625-b712-20cb460839a0" (UID: "a6d1d8ff-4e18-4625-b712-20cb460839a0"). InnerVolumeSpecName "kube-api-access-gt2b4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:25:04 crc kubenswrapper[4890]: I1125 15:25:04.998543 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6d1d8ff-4e18-4625-b712-20cb460839a0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a6d1d8ff-4e18-4625-b712-20cb460839a0" (UID: "a6d1d8ff-4e18-4625-b712-20cb460839a0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:25:05 crc kubenswrapper[4890]: I1125 15:25:05.001127 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6d1d8ff-4e18-4625-b712-20cb460839a0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a6d1d8ff-4e18-4625-b712-20cb460839a0" (UID: "a6d1d8ff-4e18-4625-b712-20cb460839a0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:25:05 crc kubenswrapper[4890]: I1125 15:25:05.002077 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6d1d8ff-4e18-4625-b712-20cb460839a0-config" (OuterVolumeSpecName: "config") pod "a6d1d8ff-4e18-4625-b712-20cb460839a0" (UID: "a6d1d8ff-4e18-4625-b712-20cb460839a0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:25:05 crc kubenswrapper[4890]: I1125 15:25:05.010354 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6d1d8ff-4e18-4625-b712-20cb460839a0-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a6d1d8ff-4e18-4625-b712-20cb460839a0" (UID: "a6d1d8ff-4e18-4625-b712-20cb460839a0"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:25:05 crc kubenswrapper[4890]: I1125 15:25:05.015424 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6d1d8ff-4e18-4625-b712-20cb460839a0-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "a6d1d8ff-4e18-4625-b712-20cb460839a0" (UID: "a6d1d8ff-4e18-4625-b712-20cb460839a0"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:25:05 crc kubenswrapper[4890]: I1125 15:25:05.017831 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6d1d8ff-4e18-4625-b712-20cb460839a0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a6d1d8ff-4e18-4625-b712-20cb460839a0" (UID: "a6d1d8ff-4e18-4625-b712-20cb460839a0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:25:05 crc kubenswrapper[4890]: I1125 15:25:05.070741 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gt2b4\" (UniqueName: \"kubernetes.io/projected/a6d1d8ff-4e18-4625-b712-20cb460839a0-kube-api-access-gt2b4\") on node \"crc\" DevicePath \"\"" Nov 25 15:25:05 crc kubenswrapper[4890]: I1125 15:25:05.070999 4890 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a6d1d8ff-4e18-4625-b712-20cb460839a0-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 25 15:25:05 crc kubenswrapper[4890]: I1125 15:25:05.071077 4890 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6d1d8ff-4e18-4625-b712-20cb460839a0-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:25:05 crc kubenswrapper[4890]: I1125 15:25:05.071148 4890 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a6d1d8ff-4e18-4625-b712-20cb460839a0-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 15:25:05 crc kubenswrapper[4890]: I1125 15:25:05.071247 4890 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a6d1d8ff-4e18-4625-b712-20cb460839a0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 15:25:05 crc kubenswrapper[4890]: I1125 15:25:05.071317 4890 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a6d1d8ff-4e18-4625-b712-20cb460839a0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 15:25:05 crc kubenswrapper[4890]: I1125 15:25:05.071382 4890 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a6d1d8ff-4e18-4625-b712-20cb460839a0-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 25 15:25:05 crc kubenswrapper[4890]: I1125 15:25:05.597726 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55478c4467-j8b46" event={"ID":"2d2eef50-1148-4cd4-8b76-67dc29bf8846","Type":"ContainerStarted","Data":"50cdc1524f9386e9437b12e56fd0771c3f0de2cc51f0eb94e0c35a1d613cfc22"} Nov 25 15:25:05 crc kubenswrapper[4890]: I1125 15:25:05.598111 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55478c4467-j8b46" Nov 25 15:25:05 crc kubenswrapper[4890]: I1125 15:25:05.599670 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-jq72w" event={"ID":"a6d1d8ff-4e18-4625-b712-20cb460839a0","Type":"ContainerDied","Data":"cf5e5b001776857ee38b097ff13b8fd5ff4c84bffd407e52efdc8354f75d84d0"} Nov 25 15:25:05 crc kubenswrapper[4890]: I1125 15:25:05.599718 4890 scope.go:117] "RemoveContainer" containerID="6cc52c6733b3eaf4423055d2b7bbf1585ef1118c2034f65833bfc1000185d14a" Nov 25 15:25:05 crc kubenswrapper[4890]: I1125 15:25:05.599691 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-jq72w" Nov 25 15:25:05 crc kubenswrapper[4890]: I1125 15:25:05.640882 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55478c4467-j8b46" podStartSLOduration=2.640860005 podStartE2EDuration="2.640860005s" podCreationTimestamp="2025-11-25 15:25:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:25:05.632838603 +0000 UTC m=+1364.075301213" watchObservedRunningTime="2025-11-25 15:25:05.640860005 +0000 UTC m=+1364.083322635" Nov 25 15:25:05 crc kubenswrapper[4890]: I1125 15:25:05.716859 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-jq72w"] Nov 25 15:25:05 crc kubenswrapper[4890]: I1125 15:25:05.726418 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-jq72w"] Nov 25 15:25:06 crc kubenswrapper[4890]: I1125 15:25:06.183780 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6d1d8ff-4e18-4625-b712-20cb460839a0" path="/var/lib/kubelet/pods/a6d1d8ff-4e18-4625-b712-20cb460839a0/volumes" Nov 25 15:25:13 crc kubenswrapper[4890]: I1125 15:25:13.473303 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55478c4467-j8b46" Nov 25 15:25:13 crc kubenswrapper[4890]: I1125 15:25:13.551667 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-4fhbm"] Nov 25 15:25:13 crc kubenswrapper[4890]: I1125 15:25:13.672447 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-89c5cd4d5-4fhbm" podUID="96c4c8fb-37d7-4415-aef1-6da42e54fbc2" containerName="dnsmasq-dns" containerID="cri-o://60206e73ae858d7c0dc08e3962ac6d824a7cb1a06b0480372710c505233f663d" gracePeriod=10 Nov 25 15:25:13 crc kubenswrapper[4890]: I1125 15:25:13.844196 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kqz9d"] Nov 25 15:25:13 crc kubenswrapper[4890]: E1125 15:25:13.844888 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6d1d8ff-4e18-4625-b712-20cb460839a0" containerName="init" Nov 25 15:25:13 crc kubenswrapper[4890]: I1125 15:25:13.844904 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6d1d8ff-4e18-4625-b712-20cb460839a0" containerName="init" Nov 25 15:25:13 crc kubenswrapper[4890]: I1125 15:25:13.845131 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6d1d8ff-4e18-4625-b712-20cb460839a0" containerName="init" Nov 25 15:25:13 crc kubenswrapper[4890]: I1125 15:25:13.847021 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kqz9d" Nov 25 15:25:13 crc kubenswrapper[4890]: I1125 15:25:13.863231 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kqz9d"] Nov 25 15:25:13 crc kubenswrapper[4890]: I1125 15:25:13.944519 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgt4c\" (UniqueName: \"kubernetes.io/projected/56f8430f-3421-47db-b225-5c40fd14c8bd-kube-api-access-dgt4c\") pod \"certified-operators-kqz9d\" (UID: \"56f8430f-3421-47db-b225-5c40fd14c8bd\") " pod="openshift-marketplace/certified-operators-kqz9d" Nov 25 15:25:13 crc kubenswrapper[4890]: I1125 15:25:13.944572 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56f8430f-3421-47db-b225-5c40fd14c8bd-catalog-content\") pod \"certified-operators-kqz9d\" (UID: \"56f8430f-3421-47db-b225-5c40fd14c8bd\") " pod="openshift-marketplace/certified-operators-kqz9d" Nov 25 15:25:13 crc kubenswrapper[4890]: I1125 15:25:13.944598 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56f8430f-3421-47db-b225-5c40fd14c8bd-utilities\") pod \"certified-operators-kqz9d\" (UID: \"56f8430f-3421-47db-b225-5c40fd14c8bd\") " pod="openshift-marketplace/certified-operators-kqz9d" Nov 25 15:25:14 crc kubenswrapper[4890]: I1125 15:25:14.046196 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgt4c\" (UniqueName: \"kubernetes.io/projected/56f8430f-3421-47db-b225-5c40fd14c8bd-kube-api-access-dgt4c\") pod \"certified-operators-kqz9d\" (UID: \"56f8430f-3421-47db-b225-5c40fd14c8bd\") " pod="openshift-marketplace/certified-operators-kqz9d" Nov 25 15:25:14 crc kubenswrapper[4890]: I1125 15:25:14.046274 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56f8430f-3421-47db-b225-5c40fd14c8bd-catalog-content\") pod \"certified-operators-kqz9d\" (UID: \"56f8430f-3421-47db-b225-5c40fd14c8bd\") " pod="openshift-marketplace/certified-operators-kqz9d" Nov 25 15:25:14 crc kubenswrapper[4890]: I1125 15:25:14.046302 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56f8430f-3421-47db-b225-5c40fd14c8bd-utilities\") pod \"certified-operators-kqz9d\" (UID: \"56f8430f-3421-47db-b225-5c40fd14c8bd\") " pod="openshift-marketplace/certified-operators-kqz9d" Nov 25 15:25:14 crc kubenswrapper[4890]: I1125 15:25:14.046835 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56f8430f-3421-47db-b225-5c40fd14c8bd-utilities\") pod \"certified-operators-kqz9d\" (UID: \"56f8430f-3421-47db-b225-5c40fd14c8bd\") " pod="openshift-marketplace/certified-operators-kqz9d" Nov 25 15:25:14 crc kubenswrapper[4890]: I1125 15:25:14.046925 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56f8430f-3421-47db-b225-5c40fd14c8bd-catalog-content\") pod \"certified-operators-kqz9d\" (UID: \"56f8430f-3421-47db-b225-5c40fd14c8bd\") " pod="openshift-marketplace/certified-operators-kqz9d" Nov 25 15:25:14 crc kubenswrapper[4890]: I1125 15:25:14.073404 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgt4c\" (UniqueName: \"kubernetes.io/projected/56f8430f-3421-47db-b225-5c40fd14c8bd-kube-api-access-dgt4c\") pod \"certified-operators-kqz9d\" (UID: \"56f8430f-3421-47db-b225-5c40fd14c8bd\") " pod="openshift-marketplace/certified-operators-kqz9d" Nov 25 15:25:14 crc kubenswrapper[4890]: I1125 15:25:14.273608 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kqz9d" Nov 25 15:25:14 crc kubenswrapper[4890]: I1125 15:25:14.688781 4890 generic.go:334] "Generic (PLEG): container finished" podID="96c4c8fb-37d7-4415-aef1-6da42e54fbc2" containerID="60206e73ae858d7c0dc08e3962ac6d824a7cb1a06b0480372710c505233f663d" exitCode=0 Nov 25 15:25:14 crc kubenswrapper[4890]: I1125 15:25:14.689110 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-4fhbm" event={"ID":"96c4c8fb-37d7-4415-aef1-6da42e54fbc2","Type":"ContainerDied","Data":"60206e73ae858d7c0dc08e3962ac6d824a7cb1a06b0480372710c505233f663d"} Nov 25 15:25:14 crc kubenswrapper[4890]: I1125 15:25:14.831804 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-4fhbm" Nov 25 15:25:14 crc kubenswrapper[4890]: I1125 15:25:14.951022 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kqz9d"] Nov 25 15:25:14 crc kubenswrapper[4890]: W1125 15:25:14.966311 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod56f8430f_3421_47db_b225_5c40fd14c8bd.slice/crio-9cd842b7acaffe38a41935af190486c9f26e28ac77eecb45b683e2602bd9dd90 WatchSource:0}: Error finding container 9cd842b7acaffe38a41935af190486c9f26e28ac77eecb45b683e2602bd9dd90: Status 404 returned error can't find the container with id 9cd842b7acaffe38a41935af190486c9f26e28ac77eecb45b683e2602bd9dd90 Nov 25 15:25:14 crc kubenswrapper[4890]: I1125 15:25:14.973917 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/96c4c8fb-37d7-4415-aef1-6da42e54fbc2-ovsdbserver-nb\") pod \"96c4c8fb-37d7-4415-aef1-6da42e54fbc2\" (UID: \"96c4c8fb-37d7-4415-aef1-6da42e54fbc2\") " Nov 25 15:25:14 crc kubenswrapper[4890]: I1125 15:25:14.974014 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/96c4c8fb-37d7-4415-aef1-6da42e54fbc2-dns-svc\") pod \"96c4c8fb-37d7-4415-aef1-6da42e54fbc2\" (UID: \"96c4c8fb-37d7-4415-aef1-6da42e54fbc2\") " Nov 25 15:25:14 crc kubenswrapper[4890]: I1125 15:25:14.974131 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/96c4c8fb-37d7-4415-aef1-6da42e54fbc2-ovsdbserver-sb\") pod \"96c4c8fb-37d7-4415-aef1-6da42e54fbc2\" (UID: \"96c4c8fb-37d7-4415-aef1-6da42e54fbc2\") " Nov 25 15:25:14 crc kubenswrapper[4890]: I1125 15:25:14.974200 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/96c4c8fb-37d7-4415-aef1-6da42e54fbc2-dns-swift-storage-0\") pod \"96c4c8fb-37d7-4415-aef1-6da42e54fbc2\" (UID: \"96c4c8fb-37d7-4415-aef1-6da42e54fbc2\") " Nov 25 15:25:14 crc kubenswrapper[4890]: I1125 15:25:14.974229 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jf74x\" (UniqueName: \"kubernetes.io/projected/96c4c8fb-37d7-4415-aef1-6da42e54fbc2-kube-api-access-jf74x\") pod \"96c4c8fb-37d7-4415-aef1-6da42e54fbc2\" (UID: \"96c4c8fb-37d7-4415-aef1-6da42e54fbc2\") " Nov 25 15:25:14 crc kubenswrapper[4890]: I1125 15:25:14.974310 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96c4c8fb-37d7-4415-aef1-6da42e54fbc2-config\") pod \"96c4c8fb-37d7-4415-aef1-6da42e54fbc2\" (UID: \"96c4c8fb-37d7-4415-aef1-6da42e54fbc2\") " Nov 25 15:25:14 crc kubenswrapper[4890]: I1125 15:25:14.980838 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96c4c8fb-37d7-4415-aef1-6da42e54fbc2-kube-api-access-jf74x" (OuterVolumeSpecName: "kube-api-access-jf74x") pod "96c4c8fb-37d7-4415-aef1-6da42e54fbc2" (UID: "96c4c8fb-37d7-4415-aef1-6da42e54fbc2"). InnerVolumeSpecName "kube-api-access-jf74x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:25:15 crc kubenswrapper[4890]: I1125 15:25:15.027130 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96c4c8fb-37d7-4415-aef1-6da42e54fbc2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "96c4c8fb-37d7-4415-aef1-6da42e54fbc2" (UID: "96c4c8fb-37d7-4415-aef1-6da42e54fbc2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:25:15 crc kubenswrapper[4890]: I1125 15:25:15.033311 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96c4c8fb-37d7-4415-aef1-6da42e54fbc2-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "96c4c8fb-37d7-4415-aef1-6da42e54fbc2" (UID: "96c4c8fb-37d7-4415-aef1-6da42e54fbc2"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:25:15 crc kubenswrapper[4890]: I1125 15:25:15.036970 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96c4c8fb-37d7-4415-aef1-6da42e54fbc2-config" (OuterVolumeSpecName: "config") pod "96c4c8fb-37d7-4415-aef1-6da42e54fbc2" (UID: "96c4c8fb-37d7-4415-aef1-6da42e54fbc2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:25:15 crc kubenswrapper[4890]: I1125 15:25:15.037246 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96c4c8fb-37d7-4415-aef1-6da42e54fbc2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "96c4c8fb-37d7-4415-aef1-6da42e54fbc2" (UID: "96c4c8fb-37d7-4415-aef1-6da42e54fbc2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:25:15 crc kubenswrapper[4890]: I1125 15:25:15.041392 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96c4c8fb-37d7-4415-aef1-6da42e54fbc2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "96c4c8fb-37d7-4415-aef1-6da42e54fbc2" (UID: "96c4c8fb-37d7-4415-aef1-6da42e54fbc2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:25:15 crc kubenswrapper[4890]: I1125 15:25:15.076983 4890 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96c4c8fb-37d7-4415-aef1-6da42e54fbc2-config\") on node \"crc\" DevicePath \"\"" Nov 25 15:25:15 crc kubenswrapper[4890]: I1125 15:25:15.077017 4890 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/96c4c8fb-37d7-4415-aef1-6da42e54fbc2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 15:25:15 crc kubenswrapper[4890]: I1125 15:25:15.077028 4890 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/96c4c8fb-37d7-4415-aef1-6da42e54fbc2-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 15:25:15 crc kubenswrapper[4890]: I1125 15:25:15.077037 4890 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/96c4c8fb-37d7-4415-aef1-6da42e54fbc2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 15:25:15 crc kubenswrapper[4890]: I1125 15:25:15.077046 4890 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/96c4c8fb-37d7-4415-aef1-6da42e54fbc2-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 25 15:25:15 crc kubenswrapper[4890]: I1125 15:25:15.077055 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jf74x\" (UniqueName: \"kubernetes.io/projected/96c4c8fb-37d7-4415-aef1-6da42e54fbc2-kube-api-access-jf74x\") on node \"crc\" DevicePath \"\"" Nov 25 15:25:15 crc kubenswrapper[4890]: I1125 15:25:15.701940 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-4fhbm" event={"ID":"96c4c8fb-37d7-4415-aef1-6da42e54fbc2","Type":"ContainerDied","Data":"17c99d83e22cd428c3c04a271604974a773fd54a98646ae74486c93321491708"} Nov 25 15:25:15 crc kubenswrapper[4890]: I1125 15:25:15.702338 4890 scope.go:117] "RemoveContainer" containerID="60206e73ae858d7c0dc08e3962ac6d824a7cb1a06b0480372710c505233f663d" Nov 25 15:25:15 crc kubenswrapper[4890]: I1125 15:25:15.702000 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-4fhbm" Nov 25 15:25:15 crc kubenswrapper[4890]: I1125 15:25:15.706148 4890 generic.go:334] "Generic (PLEG): container finished" podID="56f8430f-3421-47db-b225-5c40fd14c8bd" containerID="7d1f31d85be0bb337c001fcc38d4c57c070b912c36cb0b5146698612c8a597d7" exitCode=0 Nov 25 15:25:15 crc kubenswrapper[4890]: I1125 15:25:15.706211 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kqz9d" event={"ID":"56f8430f-3421-47db-b225-5c40fd14c8bd","Type":"ContainerDied","Data":"7d1f31d85be0bb337c001fcc38d4c57c070b912c36cb0b5146698612c8a597d7"} Nov 25 15:25:15 crc kubenswrapper[4890]: I1125 15:25:15.706241 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kqz9d" event={"ID":"56f8430f-3421-47db-b225-5c40fd14c8bd","Type":"ContainerStarted","Data":"9cd842b7acaffe38a41935af190486c9f26e28ac77eecb45b683e2602bd9dd90"} Nov 25 15:25:15 crc kubenswrapper[4890]: I1125 15:25:15.746470 4890 scope.go:117] "RemoveContainer" containerID="ec93458d0b0b9fe23fe22cd76764503b41d8aa51f583e59ce8d0dfede25477eb" Nov 25 15:25:15 crc kubenswrapper[4890]: I1125 15:25:15.762501 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-4fhbm"] Nov 25 15:25:15 crc kubenswrapper[4890]: I1125 15:25:15.772834 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-4fhbm"] Nov 25 15:25:16 crc kubenswrapper[4890]: I1125 15:25:16.185275 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96c4c8fb-37d7-4415-aef1-6da42e54fbc2" path="/var/lib/kubelet/pods/96c4c8fb-37d7-4415-aef1-6da42e54fbc2/volumes" Nov 25 15:25:18 crc kubenswrapper[4890]: I1125 15:25:18.739323 4890 generic.go:334] "Generic (PLEG): container finished" podID="56f8430f-3421-47db-b225-5c40fd14c8bd" containerID="bd2aecab4c72b6d266fd711876be5a43e59971cc71bfadb5b3dee019f5c715c7" exitCode=0 Nov 25 15:25:18 crc kubenswrapper[4890]: I1125 15:25:18.739495 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kqz9d" event={"ID":"56f8430f-3421-47db-b225-5c40fd14c8bd","Type":"ContainerDied","Data":"bd2aecab4c72b6d266fd711876be5a43e59971cc71bfadb5b3dee019f5c715c7"} Nov 25 15:25:22 crc kubenswrapper[4890]: I1125 15:25:22.417010 4890 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 25 15:25:22 crc kubenswrapper[4890]: E1125 15:25:22.418250 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96c4c8fb-37d7-4415-aef1-6da42e54fbc2" containerName="dnsmasq-dns" Nov 25 15:25:22 crc kubenswrapper[4890]: I1125 15:25:22.418273 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="96c4c8fb-37d7-4415-aef1-6da42e54fbc2" containerName="dnsmasq-dns" Nov 25 15:25:22 crc kubenswrapper[4890]: E1125 15:25:22.418296 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96c4c8fb-37d7-4415-aef1-6da42e54fbc2" containerName="init" Nov 25 15:25:22 crc kubenswrapper[4890]: I1125 15:25:22.418304 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="96c4c8fb-37d7-4415-aef1-6da42e54fbc2" containerName="init" Nov 25 15:25:22 crc kubenswrapper[4890]: I1125 15:25:22.418554 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="96c4c8fb-37d7-4415-aef1-6da42e54fbc2" containerName="dnsmasq-dns" Nov 25 15:25:22 crc kubenswrapper[4890]: I1125 15:25:22.419273 4890 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 25 15:25:22 crc kubenswrapper[4890]: I1125 15:25:22.419471 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 15:25:22 crc kubenswrapper[4890]: I1125 15:25:22.419589 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://76d277b4001db9d63b0af88b98afcc9019321101429c9be47b3022d3575d4e06" gracePeriod=15 Nov 25 15:25:22 crc kubenswrapper[4890]: I1125 15:25:22.419741 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://081a6d623cd043c0919d04e2ca93314b2b85fe33048634aa2637297d54b52cee" gracePeriod=15 Nov 25 15:25:22 crc kubenswrapper[4890]: I1125 15:25:22.419794 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://f506ec649ce24f4df6fd33e4aeafd13e42167dd71323524a5a3164517659fcc9" gracePeriod=15 Nov 25 15:25:22 crc kubenswrapper[4890]: I1125 15:25:22.419838 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://63e3a186b14ea199d1aeba5751290c266ec918cb055120cfc5c8a2ae03b1469e" gracePeriod=15 Nov 25 15:25:22 crc kubenswrapper[4890]: I1125 15:25:22.419880 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://6a860a998c40aee3e94e4cfd748f54391b3096fd3ed75c84c7ea2a5a7cefa33a" gracePeriod=15 Nov 25 15:25:22 crc kubenswrapper[4890]: I1125 15:25:22.419946 4890 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 25 15:25:22 crc kubenswrapper[4890]: E1125 15:25:22.421138 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 25 15:25:22 crc kubenswrapper[4890]: I1125 15:25:22.421180 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 25 15:25:22 crc kubenswrapper[4890]: E1125 15:25:22.421191 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 25 15:25:22 crc kubenswrapper[4890]: I1125 15:25:22.421198 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 25 15:25:22 crc kubenswrapper[4890]: E1125 15:25:22.421209 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Nov 25 15:25:22 crc kubenswrapper[4890]: I1125 15:25:22.421215 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Nov 25 15:25:22 crc kubenswrapper[4890]: E1125 15:25:22.421230 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 25 15:25:22 crc kubenswrapper[4890]: I1125 15:25:22.421236 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 25 15:25:22 crc kubenswrapper[4890]: E1125 15:25:22.421249 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 15:25:22 crc kubenswrapper[4890]: I1125 15:25:22.421255 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 15:25:22 crc kubenswrapper[4890]: E1125 15:25:22.421278 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 25 15:25:22 crc kubenswrapper[4890]: I1125 15:25:22.421284 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 25 15:25:22 crc kubenswrapper[4890]: I1125 15:25:22.421447 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 15:25:22 crc kubenswrapper[4890]: I1125 15:25:22.421469 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 25 15:25:22 crc kubenswrapper[4890]: I1125 15:25:22.421486 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 25 15:25:22 crc kubenswrapper[4890]: I1125 15:25:22.421497 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 25 15:25:22 crc kubenswrapper[4890]: I1125 15:25:22.421506 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 25 15:25:22 crc kubenswrapper[4890]: I1125 15:25:22.428204 4890 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Nov 25 15:25:22 crc kubenswrapper[4890]: I1125 15:25:22.537585 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 15:25:22 crc kubenswrapper[4890]: I1125 15:25:22.538128 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 15:25:22 crc kubenswrapper[4890]: I1125 15:25:22.538259 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 15:25:22 crc kubenswrapper[4890]: I1125 15:25:22.538334 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 15:25:22 crc kubenswrapper[4890]: I1125 15:25:22.538381 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 15:25:22 crc kubenswrapper[4890]: I1125 15:25:22.538498 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 15:25:22 crc kubenswrapper[4890]: I1125 15:25:22.538607 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 15:25:22 crc kubenswrapper[4890]: I1125 15:25:22.538639 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 15:25:22 crc kubenswrapper[4890]: I1125 15:25:22.641044 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 15:25:22 crc kubenswrapper[4890]: I1125 15:25:22.641115 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 15:25:22 crc kubenswrapper[4890]: I1125 15:25:22.641150 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 15:25:22 crc kubenswrapper[4890]: I1125 15:25:22.641201 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 15:25:22 crc kubenswrapper[4890]: I1125 15:25:22.641208 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 15:25:22 crc kubenswrapper[4890]: I1125 15:25:22.641226 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 15:25:22 crc kubenswrapper[4890]: I1125 15:25:22.641244 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 15:25:22 crc kubenswrapper[4890]: I1125 15:25:22.641261 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 15:25:22 crc kubenswrapper[4890]: I1125 15:25:22.641213 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 15:25:22 crc kubenswrapper[4890]: I1125 15:25:22.641268 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 15:25:22 crc kubenswrapper[4890]: I1125 15:25:22.641364 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 15:25:22 crc kubenswrapper[4890]: I1125 15:25:22.641431 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 15:25:22 crc kubenswrapper[4890]: I1125 15:25:22.641339 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 15:25:22 crc kubenswrapper[4890]: I1125 15:25:22.641377 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 15:25:22 crc kubenswrapper[4890]: I1125 15:25:22.641531 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 15:25:22 crc kubenswrapper[4890]: I1125 15:25:22.641630 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 15:25:22 crc kubenswrapper[4890]: I1125 15:25:22.779714 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 25 15:25:22 crc kubenswrapper[4890]: I1125 15:25:22.780667 4890 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="6a860a998c40aee3e94e4cfd748f54391b3096fd3ed75c84c7ea2a5a7cefa33a" exitCode=2 Nov 25 15:25:23 crc kubenswrapper[4890]: E1125 15:25:23.096384 4890 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:23 crc kubenswrapper[4890]: E1125 15:25:23.096642 4890 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:23 crc kubenswrapper[4890]: E1125 15:25:23.096857 4890 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:23 crc kubenswrapper[4890]: E1125 15:25:23.097031 4890 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:23 crc kubenswrapper[4890]: E1125 15:25:23.097207 4890 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:23 crc kubenswrapper[4890]: I1125 15:25:23.097228 4890 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Nov 25 15:25:23 crc kubenswrapper[4890]: E1125 15:25:23.097390 4890 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" interval="200ms" Nov 25 15:25:23 crc kubenswrapper[4890]: E1125 15:25:23.298528 4890 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" interval="400ms" Nov 25 15:25:23 crc kubenswrapper[4890]: E1125 15:25:23.699729 4890 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" interval="800ms" Nov 25 15:25:23 crc kubenswrapper[4890]: I1125 15:25:23.794830 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 25 15:25:23 crc kubenswrapper[4890]: I1125 15:25:23.795777 4890 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="081a6d623cd043c0919d04e2ca93314b2b85fe33048634aa2637297d54b52cee" exitCode=0 Nov 25 15:25:23 crc kubenswrapper[4890]: I1125 15:25:23.795801 4890 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f506ec649ce24f4df6fd33e4aeafd13e42167dd71323524a5a3164517659fcc9" exitCode=0 Nov 25 15:25:23 crc kubenswrapper[4890]: I1125 15:25:23.795810 4890 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="63e3a186b14ea199d1aeba5751290c266ec918cb055120cfc5c8a2ae03b1469e" exitCode=0 Nov 25 15:25:23 crc kubenswrapper[4890]: I1125 15:25:23.802129 4890 generic.go:334] "Generic (PLEG): container finished" podID="118dcc72-87d4-4aa3-9397-bee2c90eac63" containerID="b8ffb3bdae25d8f8daa4f0a2463144167f9c688376c66531c3818c5e17fe1bcd" exitCode=0 Nov 25 15:25:23 crc kubenswrapper[4890]: I1125 15:25:23.802210 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"118dcc72-87d4-4aa3-9397-bee2c90eac63","Type":"ContainerDied","Data":"b8ffb3bdae25d8f8daa4f0a2463144167f9c688376c66531c3818c5e17fe1bcd"} Nov 25 15:25:23 crc kubenswrapper[4890]: I1125 15:25:23.803225 4890 status_manager.go:851] "Failed to get status for pod" podUID="118dcc72-87d4-4aa3-9397-bee2c90eac63" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:24 crc kubenswrapper[4890]: E1125 15:25:24.123652 4890 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.102.83.12:6443: connect: connection refused" event="&Event{ObjectMeta:{certified-operators-kqz9d.187b4958a52f5506 openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:certified-operators-kqz9d,UID:56f8430f-3421-47db-b225-5c40fd14c8bd,APIVersion:v1,ResourceVersion:46935,FieldPath:spec.containers{registry-server},},Reason:Pulled,Message:Successfully pulled image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\" in 5.381s (5.381s including waiting). Image size: 907837715 bytes.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-25 15:25:24.122973446 +0000 UTC m=+1382.565436056,LastTimestamp:2025-11-25 15:25:24.122973446 +0000 UTC m=+1382.565436056,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 25 15:25:24 crc kubenswrapper[4890]: E1125 15:25:24.501129 4890 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" interval="1.6s" Nov 25 15:25:24 crc kubenswrapper[4890]: I1125 15:25:24.814396 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kqz9d" event={"ID":"56f8430f-3421-47db-b225-5c40fd14c8bd","Type":"ContainerStarted","Data":"41b14b144856e47ab31876e7cebe7e0578f5a19f3916527de1b9c742c0b771c1"} Nov 25 15:25:24 crc kubenswrapper[4890]: I1125 15:25:24.815971 4890 status_manager.go:851] "Failed to get status for pod" podUID="118dcc72-87d4-4aa3-9397-bee2c90eac63" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:24 crc kubenswrapper[4890]: I1125 15:25:24.816759 4890 status_manager.go:851] "Failed to get status for pod" podUID="56f8430f-3421-47db-b225-5c40fd14c8bd" pod="openshift-marketplace/certified-operators-kqz9d" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-kqz9d\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:25 crc kubenswrapper[4890]: I1125 15:25:25.199663 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 25 15:25:25 crc kubenswrapper[4890]: I1125 15:25:25.200668 4890 status_manager.go:851] "Failed to get status for pod" podUID="118dcc72-87d4-4aa3-9397-bee2c90eac63" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:25 crc kubenswrapper[4890]: I1125 15:25:25.200979 4890 status_manager.go:851] "Failed to get status for pod" podUID="56f8430f-3421-47db-b225-5c40fd14c8bd" pod="openshift-marketplace/certified-operators-kqz9d" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-kqz9d\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:25 crc kubenswrapper[4890]: I1125 15:25:25.308829 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/118dcc72-87d4-4aa3-9397-bee2c90eac63-kubelet-dir\") pod \"118dcc72-87d4-4aa3-9397-bee2c90eac63\" (UID: \"118dcc72-87d4-4aa3-9397-bee2c90eac63\") " Nov 25 15:25:25 crc kubenswrapper[4890]: I1125 15:25:25.308933 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/118dcc72-87d4-4aa3-9397-bee2c90eac63-var-lock\") pod \"118dcc72-87d4-4aa3-9397-bee2c90eac63\" (UID: \"118dcc72-87d4-4aa3-9397-bee2c90eac63\") " Nov 25 15:25:25 crc kubenswrapper[4890]: I1125 15:25:25.308960 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/118dcc72-87d4-4aa3-9397-bee2c90eac63-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "118dcc72-87d4-4aa3-9397-bee2c90eac63" (UID: "118dcc72-87d4-4aa3-9397-bee2c90eac63"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 15:25:25 crc kubenswrapper[4890]: I1125 15:25:25.309114 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/118dcc72-87d4-4aa3-9397-bee2c90eac63-var-lock" (OuterVolumeSpecName: "var-lock") pod "118dcc72-87d4-4aa3-9397-bee2c90eac63" (UID: "118dcc72-87d4-4aa3-9397-bee2c90eac63"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 15:25:25 crc kubenswrapper[4890]: I1125 15:25:25.309144 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/118dcc72-87d4-4aa3-9397-bee2c90eac63-kube-api-access\") pod \"118dcc72-87d4-4aa3-9397-bee2c90eac63\" (UID: \"118dcc72-87d4-4aa3-9397-bee2c90eac63\") " Nov 25 15:25:25 crc kubenswrapper[4890]: I1125 15:25:25.309809 4890 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/118dcc72-87d4-4aa3-9397-bee2c90eac63-var-lock\") on node \"crc\" DevicePath \"\"" Nov 25 15:25:25 crc kubenswrapper[4890]: I1125 15:25:25.309829 4890 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/118dcc72-87d4-4aa3-9397-bee2c90eac63-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 25 15:25:25 crc kubenswrapper[4890]: I1125 15:25:25.326063 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/118dcc72-87d4-4aa3-9397-bee2c90eac63-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "118dcc72-87d4-4aa3-9397-bee2c90eac63" (UID: "118dcc72-87d4-4aa3-9397-bee2c90eac63"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:25:25 crc kubenswrapper[4890]: I1125 15:25:25.411628 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/118dcc72-87d4-4aa3-9397-bee2c90eac63-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 15:25:25 crc kubenswrapper[4890]: I1125 15:25:25.626318 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 25 15:25:25 crc kubenswrapper[4890]: I1125 15:25:25.627821 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 15:25:25 crc kubenswrapper[4890]: I1125 15:25:25.628653 4890 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:25 crc kubenswrapper[4890]: I1125 15:25:25.628900 4890 status_manager.go:851] "Failed to get status for pod" podUID="118dcc72-87d4-4aa3-9397-bee2c90eac63" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:25 crc kubenswrapper[4890]: I1125 15:25:25.629309 4890 status_manager.go:851] "Failed to get status for pod" podUID="56f8430f-3421-47db-b225-5c40fd14c8bd" pod="openshift-marketplace/certified-operators-kqz9d" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-kqz9d\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:25 crc kubenswrapper[4890]: I1125 15:25:25.718106 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 25 15:25:25 crc kubenswrapper[4890]: I1125 15:25:25.718179 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 25 15:25:25 crc kubenswrapper[4890]: I1125 15:25:25.718224 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 25 15:25:25 crc kubenswrapper[4890]: I1125 15:25:25.718753 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 15:25:25 crc kubenswrapper[4890]: I1125 15:25:25.718786 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 15:25:25 crc kubenswrapper[4890]: I1125 15:25:25.718802 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 15:25:25 crc kubenswrapper[4890]: I1125 15:25:25.746173 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/kube-state-metrics-0" podUID="a47203e9-a88a-42f7-86e1-164c6ad532b5" containerName="kube-state-metrics" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 25 15:25:25 crc kubenswrapper[4890]: I1125 15:25:25.821017 4890 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Nov 25 15:25:25 crc kubenswrapper[4890]: I1125 15:25:25.821058 4890 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 25 15:25:25 crc kubenswrapper[4890]: I1125 15:25:25.821072 4890 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 25 15:25:25 crc kubenswrapper[4890]: I1125 15:25:25.826676 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 25 15:25:25 crc kubenswrapper[4890]: I1125 15:25:25.827702 4890 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="76d277b4001db9d63b0af88b98afcc9019321101429c9be47b3022d3575d4e06" exitCode=0 Nov 25 15:25:25 crc kubenswrapper[4890]: I1125 15:25:25.827848 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 15:25:25 crc kubenswrapper[4890]: I1125 15:25:25.827862 4890 scope.go:117] "RemoveContainer" containerID="081a6d623cd043c0919d04e2ca93314b2b85fe33048634aa2637297d54b52cee" Nov 25 15:25:25 crc kubenswrapper[4890]: I1125 15:25:25.831200 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 25 15:25:25 crc kubenswrapper[4890]: I1125 15:25:25.831254 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"118dcc72-87d4-4aa3-9397-bee2c90eac63","Type":"ContainerDied","Data":"8a69b7be4b9d7dd54ecf1d02bfd1eb59fa73661fd7b8c3260e8ea8fce79a0121"} Nov 25 15:25:25 crc kubenswrapper[4890]: I1125 15:25:25.831300 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8a69b7be4b9d7dd54ecf1d02bfd1eb59fa73661fd7b8c3260e8ea8fce79a0121" Nov 25 15:25:25 crc kubenswrapper[4890]: I1125 15:25:25.851739 4890 status_manager.go:851] "Failed to get status for pod" podUID="118dcc72-87d4-4aa3-9397-bee2c90eac63" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:25 crc kubenswrapper[4890]: I1125 15:25:25.852140 4890 status_manager.go:851] "Failed to get status for pod" podUID="56f8430f-3421-47db-b225-5c40fd14c8bd" pod="openshift-marketplace/certified-operators-kqz9d" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-kqz9d\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:25 crc kubenswrapper[4890]: I1125 15:25:25.852614 4890 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:25 crc kubenswrapper[4890]: I1125 15:25:25.853373 4890 status_manager.go:851] "Failed to get status for pod" podUID="56f8430f-3421-47db-b225-5c40fd14c8bd" pod="openshift-marketplace/certified-operators-kqz9d" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-kqz9d\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:25 crc kubenswrapper[4890]: I1125 15:25:25.853584 4890 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:25 crc kubenswrapper[4890]: I1125 15:25:25.853749 4890 status_manager.go:851] "Failed to get status for pod" podUID="118dcc72-87d4-4aa3-9397-bee2c90eac63" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:25 crc kubenswrapper[4890]: I1125 15:25:25.863529 4890 scope.go:117] "RemoveContainer" containerID="f506ec649ce24f4df6fd33e4aeafd13e42167dd71323524a5a3164517659fcc9" Nov 25 15:25:25 crc kubenswrapper[4890]: I1125 15:25:25.897801 4890 scope.go:117] "RemoveContainer" containerID="63e3a186b14ea199d1aeba5751290c266ec918cb055120cfc5c8a2ae03b1469e" Nov 25 15:25:25 crc kubenswrapper[4890]: I1125 15:25:25.961105 4890 scope.go:117] "RemoveContainer" containerID="6a860a998c40aee3e94e4cfd748f54391b3096fd3ed75c84c7ea2a5a7cefa33a" Nov 25 15:25:25 crc kubenswrapper[4890]: I1125 15:25:25.983000 4890 scope.go:117] "RemoveContainer" containerID="76d277b4001db9d63b0af88b98afcc9019321101429c9be47b3022d3575d4e06" Nov 25 15:25:26 crc kubenswrapper[4890]: I1125 15:25:26.017587 4890 scope.go:117] "RemoveContainer" containerID="f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453" Nov 25 15:25:26 crc kubenswrapper[4890]: I1125 15:25:26.052959 4890 scope.go:117] "RemoveContainer" containerID="081a6d623cd043c0919d04e2ca93314b2b85fe33048634aa2637297d54b52cee" Nov 25 15:25:26 crc kubenswrapper[4890]: E1125 15:25:26.053841 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"081a6d623cd043c0919d04e2ca93314b2b85fe33048634aa2637297d54b52cee\": container with ID starting with 081a6d623cd043c0919d04e2ca93314b2b85fe33048634aa2637297d54b52cee not found: ID does not exist" containerID="081a6d623cd043c0919d04e2ca93314b2b85fe33048634aa2637297d54b52cee" Nov 25 15:25:26 crc kubenswrapper[4890]: I1125 15:25:26.053881 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"081a6d623cd043c0919d04e2ca93314b2b85fe33048634aa2637297d54b52cee"} err="failed to get container status \"081a6d623cd043c0919d04e2ca93314b2b85fe33048634aa2637297d54b52cee\": rpc error: code = NotFound desc = could not find container \"081a6d623cd043c0919d04e2ca93314b2b85fe33048634aa2637297d54b52cee\": container with ID starting with 081a6d623cd043c0919d04e2ca93314b2b85fe33048634aa2637297d54b52cee not found: ID does not exist" Nov 25 15:25:26 crc kubenswrapper[4890]: I1125 15:25:26.053912 4890 scope.go:117] "RemoveContainer" containerID="f506ec649ce24f4df6fd33e4aeafd13e42167dd71323524a5a3164517659fcc9" Nov 25 15:25:26 crc kubenswrapper[4890]: E1125 15:25:26.054483 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f506ec649ce24f4df6fd33e4aeafd13e42167dd71323524a5a3164517659fcc9\": container with ID starting with f506ec649ce24f4df6fd33e4aeafd13e42167dd71323524a5a3164517659fcc9 not found: ID does not exist" containerID="f506ec649ce24f4df6fd33e4aeafd13e42167dd71323524a5a3164517659fcc9" Nov 25 15:25:26 crc kubenswrapper[4890]: I1125 15:25:26.054505 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f506ec649ce24f4df6fd33e4aeafd13e42167dd71323524a5a3164517659fcc9"} err="failed to get container status \"f506ec649ce24f4df6fd33e4aeafd13e42167dd71323524a5a3164517659fcc9\": rpc error: code = NotFound desc = could not find container \"f506ec649ce24f4df6fd33e4aeafd13e42167dd71323524a5a3164517659fcc9\": container with ID starting with f506ec649ce24f4df6fd33e4aeafd13e42167dd71323524a5a3164517659fcc9 not found: ID does not exist" Nov 25 15:25:26 crc kubenswrapper[4890]: I1125 15:25:26.054521 4890 scope.go:117] "RemoveContainer" containerID="63e3a186b14ea199d1aeba5751290c266ec918cb055120cfc5c8a2ae03b1469e" Nov 25 15:25:26 crc kubenswrapper[4890]: E1125 15:25:26.054968 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63e3a186b14ea199d1aeba5751290c266ec918cb055120cfc5c8a2ae03b1469e\": container with ID starting with 63e3a186b14ea199d1aeba5751290c266ec918cb055120cfc5c8a2ae03b1469e not found: ID does not exist" containerID="63e3a186b14ea199d1aeba5751290c266ec918cb055120cfc5c8a2ae03b1469e" Nov 25 15:25:26 crc kubenswrapper[4890]: I1125 15:25:26.055005 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63e3a186b14ea199d1aeba5751290c266ec918cb055120cfc5c8a2ae03b1469e"} err="failed to get container status \"63e3a186b14ea199d1aeba5751290c266ec918cb055120cfc5c8a2ae03b1469e\": rpc error: code = NotFound desc = could not find container \"63e3a186b14ea199d1aeba5751290c266ec918cb055120cfc5c8a2ae03b1469e\": container with ID starting with 63e3a186b14ea199d1aeba5751290c266ec918cb055120cfc5c8a2ae03b1469e not found: ID does not exist" Nov 25 15:25:26 crc kubenswrapper[4890]: I1125 15:25:26.055032 4890 scope.go:117] "RemoveContainer" containerID="6a860a998c40aee3e94e4cfd748f54391b3096fd3ed75c84c7ea2a5a7cefa33a" Nov 25 15:25:26 crc kubenswrapper[4890]: E1125 15:25:26.055365 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a860a998c40aee3e94e4cfd748f54391b3096fd3ed75c84c7ea2a5a7cefa33a\": container with ID starting with 6a860a998c40aee3e94e4cfd748f54391b3096fd3ed75c84c7ea2a5a7cefa33a not found: ID does not exist" containerID="6a860a998c40aee3e94e4cfd748f54391b3096fd3ed75c84c7ea2a5a7cefa33a" Nov 25 15:25:26 crc kubenswrapper[4890]: I1125 15:25:26.055391 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a860a998c40aee3e94e4cfd748f54391b3096fd3ed75c84c7ea2a5a7cefa33a"} err="failed to get container status \"6a860a998c40aee3e94e4cfd748f54391b3096fd3ed75c84c7ea2a5a7cefa33a\": rpc error: code = NotFound desc = could not find container \"6a860a998c40aee3e94e4cfd748f54391b3096fd3ed75c84c7ea2a5a7cefa33a\": container with ID starting with 6a860a998c40aee3e94e4cfd748f54391b3096fd3ed75c84c7ea2a5a7cefa33a not found: ID does not exist" Nov 25 15:25:26 crc kubenswrapper[4890]: I1125 15:25:26.055407 4890 scope.go:117] "RemoveContainer" containerID="76d277b4001db9d63b0af88b98afcc9019321101429c9be47b3022d3575d4e06" Nov 25 15:25:26 crc kubenswrapper[4890]: E1125 15:25:26.055702 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76d277b4001db9d63b0af88b98afcc9019321101429c9be47b3022d3575d4e06\": container with ID starting with 76d277b4001db9d63b0af88b98afcc9019321101429c9be47b3022d3575d4e06 not found: ID does not exist" containerID="76d277b4001db9d63b0af88b98afcc9019321101429c9be47b3022d3575d4e06" Nov 25 15:25:26 crc kubenswrapper[4890]: I1125 15:25:26.055727 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76d277b4001db9d63b0af88b98afcc9019321101429c9be47b3022d3575d4e06"} err="failed to get container status \"76d277b4001db9d63b0af88b98afcc9019321101429c9be47b3022d3575d4e06\": rpc error: code = NotFound desc = could not find container \"76d277b4001db9d63b0af88b98afcc9019321101429c9be47b3022d3575d4e06\": container with ID starting with 76d277b4001db9d63b0af88b98afcc9019321101429c9be47b3022d3575d4e06 not found: ID does not exist" Nov 25 15:25:26 crc kubenswrapper[4890]: I1125 15:25:26.055743 4890 scope.go:117] "RemoveContainer" containerID="f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453" Nov 25 15:25:26 crc kubenswrapper[4890]: E1125 15:25:26.055925 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\": container with ID starting with f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453 not found: ID does not exist" containerID="f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453" Nov 25 15:25:26 crc kubenswrapper[4890]: I1125 15:25:26.055964 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453"} err="failed to get container status \"f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\": rpc error: code = NotFound desc = could not find container \"f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453\": container with ID starting with f2ef1cfdf4f7bcf3665208f27e3cde7d8b1c950c6dd5ae6142ac8f5ee7060453 not found: ID does not exist" Nov 25 15:25:26 crc kubenswrapper[4890]: E1125 15:25:26.102859 4890 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" interval="3.2s" Nov 25 15:25:26 crc kubenswrapper[4890]: I1125 15:25:26.183024 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Nov 25 15:25:27 crc kubenswrapper[4890]: E1125 15:25:27.457655 4890 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.12:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 15:25:27 crc kubenswrapper[4890]: I1125 15:25:27.458792 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 15:25:27 crc kubenswrapper[4890]: E1125 15:25:27.636349 4890 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.102.83.12:6443: connect: connection refused" event="&Event{ObjectMeta:{certified-operators-kqz9d.187b4958a52f5506 openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:certified-operators-kqz9d,UID:56f8430f-3421-47db-b225-5c40fd14c8bd,APIVersion:v1,ResourceVersion:46935,FieldPath:spec.containers{registry-server},},Reason:Pulled,Message:Successfully pulled image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\" in 5.381s (5.381s including waiting). Image size: 907837715 bytes.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-25 15:25:24.122973446 +0000 UTC m=+1382.565436056,LastTimestamp:2025-11-25 15:25:24.122973446 +0000 UTC m=+1382.565436056,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 25 15:25:27 crc kubenswrapper[4890]: I1125 15:25:27.851917 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"216dc27f3c7236a012847fb5349779c9c44f1b480be3dc3c4f0ae8086ab78f23"} Nov 25 15:25:28 crc kubenswrapper[4890]: I1125 15:25:28.863463 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"f8010cc46ab44d816173f84c66043f945180f00900ca208dc599fb56bcf061b2"} Nov 25 15:25:28 crc kubenswrapper[4890]: E1125 15:25:28.864144 4890 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.12:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 15:25:28 crc kubenswrapper[4890]: I1125 15:25:28.864284 4890 status_manager.go:851] "Failed to get status for pod" podUID="56f8430f-3421-47db-b225-5c40fd14c8bd" pod="openshift-marketplace/certified-operators-kqz9d" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-kqz9d\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:28 crc kubenswrapper[4890]: I1125 15:25:28.864762 4890 status_manager.go:851] "Failed to get status for pod" podUID="118dcc72-87d4-4aa3-9397-bee2c90eac63" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:29 crc kubenswrapper[4890]: E1125 15:25:29.304731 4890 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" interval="6.4s" Nov 25 15:25:29 crc kubenswrapper[4890]: E1125 15:25:29.872322 4890 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.12:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 15:25:32 crc kubenswrapper[4890]: I1125 15:25:32.185570 4890 status_manager.go:851] "Failed to get status for pod" podUID="56f8430f-3421-47db-b225-5c40fd14c8bd" pod="openshift-marketplace/certified-operators-kqz9d" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-kqz9d\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:32 crc kubenswrapper[4890]: I1125 15:25:32.186761 4890 status_manager.go:851] "Failed to get status for pod" podUID="118dcc72-87d4-4aa3-9397-bee2c90eac63" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:34 crc kubenswrapper[4890]: I1125 15:25:34.274547 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kqz9d" Nov 25 15:25:34 crc kubenswrapper[4890]: I1125 15:25:34.275597 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kqz9d" Nov 25 15:25:34 crc kubenswrapper[4890]: I1125 15:25:34.327276 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kqz9d" Nov 25 15:25:34 crc kubenswrapper[4890]: I1125 15:25:34.328011 4890 status_manager.go:851] "Failed to get status for pod" podUID="56f8430f-3421-47db-b225-5c40fd14c8bd" pod="openshift-marketplace/certified-operators-kqz9d" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-kqz9d\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:34 crc kubenswrapper[4890]: I1125 15:25:34.328281 4890 status_manager.go:851] "Failed to get status for pod" podUID="118dcc72-87d4-4aa3-9397-bee2c90eac63" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:34 crc kubenswrapper[4890]: I1125 15:25:34.924466 4890 generic.go:334] "Generic (PLEG): container finished" podID="d08ea973-c3f2-4192-8bfc-b025f0c5391b" containerID="8bb3adcfe9f7b378153065acb08cd345e9a3a2e085d8241903165dc9dd5a4297" exitCode=0 Nov 25 15:25:34 crc kubenswrapper[4890]: I1125 15:25:34.924578 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d08ea973-c3f2-4192-8bfc-b025f0c5391b","Type":"ContainerDied","Data":"8bb3adcfe9f7b378153065acb08cd345e9a3a2e085d8241903165dc9dd5a4297"} Nov 25 15:25:34 crc kubenswrapper[4890]: I1125 15:25:34.926391 4890 status_manager.go:851] "Failed to get status for pod" podUID="56f8430f-3421-47db-b225-5c40fd14c8bd" pod="openshift-marketplace/certified-operators-kqz9d" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-kqz9d\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:34 crc kubenswrapper[4890]: I1125 15:25:34.926558 4890 status_manager.go:851] "Failed to get status for pod" podUID="118dcc72-87d4-4aa3-9397-bee2c90eac63" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:34 crc kubenswrapper[4890]: I1125 15:25:34.926910 4890 status_manager.go:851] "Failed to get status for pod" podUID="d08ea973-c3f2-4192-8bfc-b025f0c5391b" pod="openstack/rabbitmq-cell1-server-0" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/rabbitmq-cell1-server-0\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:34 crc kubenswrapper[4890]: I1125 15:25:34.930516 4890 generic.go:334] "Generic (PLEG): container finished" podID="dad943fe-2209-40f9-a9bb-0be7ba80cca2" containerID="0fc718ab7f1457fe75eff145c26dc4f53ab7a75dd29150bb92b277bdff7d608f" exitCode=1 Nov 25 15:25:34 crc kubenswrapper[4890]: I1125 15:25:34.930599 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-66fc8c7fbf-lnjxm" event={"ID":"dad943fe-2209-40f9-a9bb-0be7ba80cca2","Type":"ContainerDied","Data":"0fc718ab7f1457fe75eff145c26dc4f53ab7a75dd29150bb92b277bdff7d608f"} Nov 25 15:25:34 crc kubenswrapper[4890]: I1125 15:25:34.931356 4890 scope.go:117] "RemoveContainer" containerID="0fc718ab7f1457fe75eff145c26dc4f53ab7a75dd29150bb92b277bdff7d608f" Nov 25 15:25:34 crc kubenswrapper[4890]: I1125 15:25:34.931511 4890 status_manager.go:851] "Failed to get status for pod" podUID="118dcc72-87d4-4aa3-9397-bee2c90eac63" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:34 crc kubenswrapper[4890]: I1125 15:25:34.931786 4890 status_manager.go:851] "Failed to get status for pod" podUID="d08ea973-c3f2-4192-8bfc-b025f0c5391b" pod="openstack/rabbitmq-cell1-server-0" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/rabbitmq-cell1-server-0\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:34 crc kubenswrapper[4890]: I1125 15:25:34.932142 4890 status_manager.go:851] "Failed to get status for pod" podUID="56f8430f-3421-47db-b225-5c40fd14c8bd" pod="openshift-marketplace/certified-operators-kqz9d" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-kqz9d\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:34 crc kubenswrapper[4890]: I1125 15:25:34.932768 4890 status_manager.go:851] "Failed to get status for pod" podUID="dad943fe-2209-40f9-a9bb-0be7ba80cca2" pod="metallb-system/metallb-operator-controller-manager-66fc8c7fbf-lnjxm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-66fc8c7fbf-lnjxm\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:34 crc kubenswrapper[4890]: I1125 15:25:34.933569 4890 generic.go:334] "Generic (PLEG): container finished" podID="9bab37b3-681b-4cd5-9fee-503afa0d6d22" containerID="b4e01c87ce56f03bf70dea18143d788a675a2a28fda3721303f87885ade5d9df" exitCode=0 Nov 25 15:25:34 crc kubenswrapper[4890]: I1125 15:25:34.933598 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9bab37b3-681b-4cd5-9fee-503afa0d6d22","Type":"ContainerDied","Data":"b4e01c87ce56f03bf70dea18143d788a675a2a28fda3721303f87885ade5d9df"} Nov 25 15:25:34 crc kubenswrapper[4890]: I1125 15:25:34.934710 4890 status_manager.go:851] "Failed to get status for pod" podUID="d08ea973-c3f2-4192-8bfc-b025f0c5391b" pod="openstack/rabbitmq-cell1-server-0" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/rabbitmq-cell1-server-0\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:34 crc kubenswrapper[4890]: I1125 15:25:34.935031 4890 status_manager.go:851] "Failed to get status for pod" podUID="9bab37b3-681b-4cd5-9fee-503afa0d6d22" pod="openstack/rabbitmq-server-0" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/rabbitmq-server-0\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:34 crc kubenswrapper[4890]: I1125 15:25:34.937582 4890 status_manager.go:851] "Failed to get status for pod" podUID="56f8430f-3421-47db-b225-5c40fd14c8bd" pod="openshift-marketplace/certified-operators-kqz9d" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-kqz9d\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:34 crc kubenswrapper[4890]: I1125 15:25:34.937993 4890 status_manager.go:851] "Failed to get status for pod" podUID="dad943fe-2209-40f9-a9bb-0be7ba80cca2" pod="metallb-system/metallb-operator-controller-manager-66fc8c7fbf-lnjxm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-66fc8c7fbf-lnjxm\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:34 crc kubenswrapper[4890]: I1125 15:25:34.938407 4890 status_manager.go:851] "Failed to get status for pod" podUID="118dcc72-87d4-4aa3-9397-bee2c90eac63" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:34 crc kubenswrapper[4890]: E1125 15:25:34.964675 4890 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openstack/persistence-rabbitmq-server-0: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/persistentvolumeclaims/persistence-rabbitmq-server-0\": dial tcp 38.102.83.12:6443: connect: connection refused" pod="openstack/rabbitmq-server-0" volumeName="persistence" Nov 25 15:25:34 crc kubenswrapper[4890]: E1125 15:25:34.965524 4890 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openstack/persistence-rabbitmq-cell1-server-0: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/persistentvolumeclaims/persistence-rabbitmq-cell1-server-0\": dial tcp 38.102.83.12:6443: connect: connection refused" pod="openstack/rabbitmq-cell1-server-0" volumeName="persistence" Nov 25 15:25:34 crc kubenswrapper[4890]: I1125 15:25:34.987701 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kqz9d" Nov 25 15:25:34 crc kubenswrapper[4890]: I1125 15:25:34.988481 4890 status_manager.go:851] "Failed to get status for pod" podUID="d08ea973-c3f2-4192-8bfc-b025f0c5391b" pod="openstack/rabbitmq-cell1-server-0" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/rabbitmq-cell1-server-0\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:34 crc kubenswrapper[4890]: I1125 15:25:34.988767 4890 status_manager.go:851] "Failed to get status for pod" podUID="9bab37b3-681b-4cd5-9fee-503afa0d6d22" pod="openstack/rabbitmq-server-0" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/rabbitmq-server-0\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:34 crc kubenswrapper[4890]: I1125 15:25:34.989045 4890 status_manager.go:851] "Failed to get status for pod" podUID="56f8430f-3421-47db-b225-5c40fd14c8bd" pod="openshift-marketplace/certified-operators-kqz9d" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-kqz9d\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:34 crc kubenswrapper[4890]: I1125 15:25:34.989343 4890 status_manager.go:851] "Failed to get status for pod" podUID="dad943fe-2209-40f9-a9bb-0be7ba80cca2" pod="metallb-system/metallb-operator-controller-manager-66fc8c7fbf-lnjxm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-66fc8c7fbf-lnjxm\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:34 crc kubenswrapper[4890]: I1125 15:25:34.990538 4890 status_manager.go:851] "Failed to get status for pod" podUID="118dcc72-87d4-4aa3-9397-bee2c90eac63" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:35 crc kubenswrapper[4890]: E1125 15:25:35.706671 4890 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" interval="7s" Nov 25 15:25:35 crc kubenswrapper[4890]: I1125 15:25:35.746342 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/kube-state-metrics-0" podUID="a47203e9-a88a-42f7-86e1-164c6ad532b5" containerName="kube-state-metrics" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 25 15:25:35 crc kubenswrapper[4890]: I1125 15:25:35.945200 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d08ea973-c3f2-4192-8bfc-b025f0c5391b","Type":"ContainerStarted","Data":"f0b3afeee6410bd63ac0a4cfafef343fa9f037f8f30d828779453689dbd871bd"} Nov 25 15:25:35 crc kubenswrapper[4890]: I1125 15:25:35.945462 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:25:35 crc kubenswrapper[4890]: I1125 15:25:35.946057 4890 status_manager.go:851] "Failed to get status for pod" podUID="dad943fe-2209-40f9-a9bb-0be7ba80cca2" pod="metallb-system/metallb-operator-controller-manager-66fc8c7fbf-lnjxm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-66fc8c7fbf-lnjxm\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:35 crc kubenswrapper[4890]: I1125 15:25:35.946315 4890 status_manager.go:851] "Failed to get status for pod" podUID="118dcc72-87d4-4aa3-9397-bee2c90eac63" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:35 crc kubenswrapper[4890]: I1125 15:25:35.946518 4890 status_manager.go:851] "Failed to get status for pod" podUID="d08ea973-c3f2-4192-8bfc-b025f0c5391b" pod="openstack/rabbitmq-cell1-server-0" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/rabbitmq-cell1-server-0\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:35 crc kubenswrapper[4890]: I1125 15:25:35.946758 4890 status_manager.go:851] "Failed to get status for pod" podUID="9bab37b3-681b-4cd5-9fee-503afa0d6d22" pod="openstack/rabbitmq-server-0" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/rabbitmq-server-0\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:35 crc kubenswrapper[4890]: I1125 15:25:35.947038 4890 status_manager.go:851] "Failed to get status for pod" podUID="56f8430f-3421-47db-b225-5c40fd14c8bd" pod="openshift-marketplace/certified-operators-kqz9d" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-kqz9d\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:35 crc kubenswrapper[4890]: I1125 15:25:35.947323 4890 generic.go:334] "Generic (PLEG): container finished" podID="dad943fe-2209-40f9-a9bb-0be7ba80cca2" containerID="be30ac4807bb1b8659327603f58cc562ca5c318c9246dbbc3a89d446e0498389" exitCode=1 Nov 25 15:25:35 crc kubenswrapper[4890]: I1125 15:25:35.947343 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-66fc8c7fbf-lnjxm" event={"ID":"dad943fe-2209-40f9-a9bb-0be7ba80cca2","Type":"ContainerDied","Data":"be30ac4807bb1b8659327603f58cc562ca5c318c9246dbbc3a89d446e0498389"} Nov 25 15:25:35 crc kubenswrapper[4890]: I1125 15:25:35.947384 4890 scope.go:117] "RemoveContainer" containerID="0fc718ab7f1457fe75eff145c26dc4f53ab7a75dd29150bb92b277bdff7d608f" Nov 25 15:25:35 crc kubenswrapper[4890]: I1125 15:25:35.947946 4890 scope.go:117] "RemoveContainer" containerID="be30ac4807bb1b8659327603f58cc562ca5c318c9246dbbc3a89d446e0498389" Nov 25 15:25:35 crc kubenswrapper[4890]: I1125 15:25:35.948109 4890 status_manager.go:851] "Failed to get status for pod" podUID="d08ea973-c3f2-4192-8bfc-b025f0c5391b" pod="openstack/rabbitmq-cell1-server-0" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/rabbitmq-cell1-server-0\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:35 crc kubenswrapper[4890]: E1125 15:25:35.948151 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=metallb-operator-controller-manager-66fc8c7fbf-lnjxm_metallb-system(dad943fe-2209-40f9-a9bb-0be7ba80cca2)\"" pod="metallb-system/metallb-operator-controller-manager-66fc8c7fbf-lnjxm" podUID="dad943fe-2209-40f9-a9bb-0be7ba80cca2" Nov 25 15:25:35 crc kubenswrapper[4890]: I1125 15:25:35.948352 4890 status_manager.go:851] "Failed to get status for pod" podUID="9bab37b3-681b-4cd5-9fee-503afa0d6d22" pod="openstack/rabbitmq-server-0" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/rabbitmq-server-0\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:35 crc kubenswrapper[4890]: I1125 15:25:35.948594 4890 status_manager.go:851] "Failed to get status for pod" podUID="56f8430f-3421-47db-b225-5c40fd14c8bd" pod="openshift-marketplace/certified-operators-kqz9d" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-kqz9d\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:35 crc kubenswrapper[4890]: I1125 15:25:35.948849 4890 status_manager.go:851] "Failed to get status for pod" podUID="dad943fe-2209-40f9-a9bb-0be7ba80cca2" pod="metallb-system/metallb-operator-controller-manager-66fc8c7fbf-lnjxm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-66fc8c7fbf-lnjxm\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:35 crc kubenswrapper[4890]: I1125 15:25:35.949074 4890 status_manager.go:851] "Failed to get status for pod" podUID="118dcc72-87d4-4aa3-9397-bee2c90eac63" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:35 crc kubenswrapper[4890]: I1125 15:25:35.953850 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9bab37b3-681b-4cd5-9fee-503afa0d6d22","Type":"ContainerStarted","Data":"6bfb516e175eec7473f31226ee4f9ff525a35f39558ca88aeffbd781243e1b96"} Nov 25 15:25:35 crc kubenswrapper[4890]: I1125 15:25:35.955245 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 25 15:25:35 crc kubenswrapper[4890]: I1125 15:25:35.955380 4890 status_manager.go:851] "Failed to get status for pod" podUID="56f8430f-3421-47db-b225-5c40fd14c8bd" pod="openshift-marketplace/certified-operators-kqz9d" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-kqz9d\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:35 crc kubenswrapper[4890]: I1125 15:25:35.955659 4890 status_manager.go:851] "Failed to get status for pod" podUID="dad943fe-2209-40f9-a9bb-0be7ba80cca2" pod="metallb-system/metallb-operator-controller-manager-66fc8c7fbf-lnjxm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-66fc8c7fbf-lnjxm\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:35 crc kubenswrapper[4890]: I1125 15:25:35.956032 4890 status_manager.go:851] "Failed to get status for pod" podUID="118dcc72-87d4-4aa3-9397-bee2c90eac63" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:35 crc kubenswrapper[4890]: I1125 15:25:35.956307 4890 status_manager.go:851] "Failed to get status for pod" podUID="d08ea973-c3f2-4192-8bfc-b025f0c5391b" pod="openstack/rabbitmq-cell1-server-0" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/rabbitmq-cell1-server-0\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:35 crc kubenswrapper[4890]: I1125 15:25:35.956568 4890 status_manager.go:851] "Failed to get status for pod" podUID="9bab37b3-681b-4cd5-9fee-503afa0d6d22" pod="openstack/rabbitmq-server-0" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/rabbitmq-server-0\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:35 crc kubenswrapper[4890]: E1125 15:25:35.985525 4890 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openstack/persistence-rabbitmq-server-0: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/persistentvolumeclaims/persistence-rabbitmq-server-0\": dial tcp 38.102.83.12:6443: connect: connection refused" pod="openstack/rabbitmq-server-0" volumeName="persistence" Nov 25 15:25:35 crc kubenswrapper[4890]: E1125 15:25:35.986111 4890 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openstack/persistence-rabbitmq-cell1-server-0: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/persistentvolumeclaims/persistence-rabbitmq-cell1-server-0\": dial tcp 38.102.83.12:6443: connect: connection refused" pod="openstack/rabbitmq-cell1-server-0" volumeName="persistence" Nov 25 15:25:36 crc kubenswrapper[4890]: I1125 15:25:36.173560 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 15:25:36 crc kubenswrapper[4890]: I1125 15:25:36.174951 4890 status_manager.go:851] "Failed to get status for pod" podUID="dad943fe-2209-40f9-a9bb-0be7ba80cca2" pod="metallb-system/metallb-operator-controller-manager-66fc8c7fbf-lnjxm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-66fc8c7fbf-lnjxm\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:36 crc kubenswrapper[4890]: I1125 15:25:36.175469 4890 status_manager.go:851] "Failed to get status for pod" podUID="118dcc72-87d4-4aa3-9397-bee2c90eac63" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:36 crc kubenswrapper[4890]: I1125 15:25:36.176057 4890 status_manager.go:851] "Failed to get status for pod" podUID="d08ea973-c3f2-4192-8bfc-b025f0c5391b" pod="openstack/rabbitmq-cell1-server-0" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/rabbitmq-cell1-server-0\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:36 crc kubenswrapper[4890]: I1125 15:25:36.176716 4890 status_manager.go:851] "Failed to get status for pod" podUID="9bab37b3-681b-4cd5-9fee-503afa0d6d22" pod="openstack/rabbitmq-server-0" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/rabbitmq-server-0\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:36 crc kubenswrapper[4890]: I1125 15:25:36.177331 4890 status_manager.go:851] "Failed to get status for pod" podUID="56f8430f-3421-47db-b225-5c40fd14c8bd" pod="openshift-marketplace/certified-operators-kqz9d" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-kqz9d\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:36 crc kubenswrapper[4890]: I1125 15:25:36.197408 4890 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5bd5e648-6322-404d-bf1b-3b59a8092291" Nov 25 15:25:36 crc kubenswrapper[4890]: I1125 15:25:36.197469 4890 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5bd5e648-6322-404d-bf1b-3b59a8092291" Nov 25 15:25:36 crc kubenswrapper[4890]: E1125 15:25:36.198112 4890 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 15:25:36 crc kubenswrapper[4890]: I1125 15:25:36.199063 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 15:25:36 crc kubenswrapper[4890]: W1125 15:25:36.228993 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-5bc158ca0339b015b6a4cb7d330733e87c2adf4c50a32e9396faee3eef30899f WatchSource:0}: Error finding container 5bc158ca0339b015b6a4cb7d330733e87c2adf4c50a32e9396faee3eef30899f: Status 404 returned error can't find the container with id 5bc158ca0339b015b6a4cb7d330733e87c2adf4c50a32e9396faee3eef30899f Nov 25 15:25:36 crc kubenswrapper[4890]: I1125 15:25:36.963403 4890 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="08c363b69758c11ccf37d6a4409c0021d72621968149299a88021e8ae18115ae" exitCode=0 Nov 25 15:25:36 crc kubenswrapper[4890]: I1125 15:25:36.963473 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"08c363b69758c11ccf37d6a4409c0021d72621968149299a88021e8ae18115ae"} Nov 25 15:25:36 crc kubenswrapper[4890]: I1125 15:25:36.963500 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"5bc158ca0339b015b6a4cb7d330733e87c2adf4c50a32e9396faee3eef30899f"} Nov 25 15:25:36 crc kubenswrapper[4890]: I1125 15:25:36.963725 4890 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5bd5e648-6322-404d-bf1b-3b59a8092291" Nov 25 15:25:36 crc kubenswrapper[4890]: I1125 15:25:36.963738 4890 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5bd5e648-6322-404d-bf1b-3b59a8092291" Nov 25 15:25:36 crc kubenswrapper[4890]: E1125 15:25:36.964104 4890 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 15:25:36 crc kubenswrapper[4890]: I1125 15:25:36.964312 4890 status_manager.go:851] "Failed to get status for pod" podUID="d08ea973-c3f2-4192-8bfc-b025f0c5391b" pod="openstack/rabbitmq-cell1-server-0" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/rabbitmq-cell1-server-0\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:36 crc kubenswrapper[4890]: I1125 15:25:36.964489 4890 status_manager.go:851] "Failed to get status for pod" podUID="9bab37b3-681b-4cd5-9fee-503afa0d6d22" pod="openstack/rabbitmq-server-0" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/rabbitmq-server-0\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:36 crc kubenswrapper[4890]: I1125 15:25:36.964711 4890 status_manager.go:851] "Failed to get status for pod" podUID="56f8430f-3421-47db-b225-5c40fd14c8bd" pod="openshift-marketplace/certified-operators-kqz9d" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-kqz9d\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:36 crc kubenswrapper[4890]: I1125 15:25:36.964998 4890 status_manager.go:851] "Failed to get status for pod" podUID="dad943fe-2209-40f9-a9bb-0be7ba80cca2" pod="metallb-system/metallb-operator-controller-manager-66fc8c7fbf-lnjxm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-66fc8c7fbf-lnjxm\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:36 crc kubenswrapper[4890]: I1125 15:25:36.965439 4890 status_manager.go:851] "Failed to get status for pod" podUID="118dcc72-87d4-4aa3-9397-bee2c90eac63" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Nov 25 15:25:37 crc kubenswrapper[4890]: E1125 15:25:37.007220 4890 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openstack/persistence-rabbitmq-cell1-server-0: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/persistentvolumeclaims/persistence-rabbitmq-cell1-server-0\": dial tcp 38.102.83.12:6443: connect: connection refused" pod="openstack/rabbitmq-cell1-server-0" volumeName="persistence" Nov 25 15:25:37 crc kubenswrapper[4890]: E1125 15:25:37.008251 4890 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openstack/persistence-rabbitmq-server-0: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/persistentvolumeclaims/persistence-rabbitmq-server-0\": dial tcp 38.102.83.12:6443: connect: connection refused" pod="openstack/rabbitmq-server-0" volumeName="persistence" Nov 25 15:25:37 crc kubenswrapper[4890]: E1125 15:25:37.638137 4890 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.102.83.12:6443: connect: connection refused" event="&Event{ObjectMeta:{certified-operators-kqz9d.187b4958a52f5506 openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:certified-operators-kqz9d,UID:56f8430f-3421-47db-b225-5c40fd14c8bd,APIVersion:v1,ResourceVersion:46935,FieldPath:spec.containers{registry-server},},Reason:Pulled,Message:Successfully pulled image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\" in 5.381s (5.381s including waiting). Image size: 907837715 bytes.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-25 15:25:24.122973446 +0000 UTC m=+1382.565436056,LastTimestamp:2025-11-25 15:25:24.122973446 +0000 UTC m=+1382.565436056,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 25 15:25:37 crc kubenswrapper[4890]: I1125 15:25:37.980261 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"dde332dd711a1e4f248617285fcb0aeafe3896bb8f4fbf986ad84caa69183c7e"} Nov 25 15:25:37 crc kubenswrapper[4890]: I1125 15:25:37.985854 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 25 15:25:37 crc kubenswrapper[4890]: I1125 15:25:37.985905 4890 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="93325a6407389c4a7db0ba3c1a9c84a6980db3c3277cdf3860583a564b562bbe" exitCode=1 Nov 25 15:25:37 crc kubenswrapper[4890]: I1125 15:25:37.985939 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"93325a6407389c4a7db0ba3c1a9c84a6980db3c3277cdf3860583a564b562bbe"} Nov 25 15:25:37 crc kubenswrapper[4890]: I1125 15:25:37.986976 4890 scope.go:117] "RemoveContainer" containerID="93325a6407389c4a7db0ba3c1a9c84a6980db3c3277cdf3860583a564b562bbe" Nov 25 15:25:39 crc kubenswrapper[4890]: I1125 15:25:38.999551 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"ced820c6dcffd50dced1ffd183b566730e8b2f829258957d430d49998db15ecb"} Nov 25 15:25:39 crc kubenswrapper[4890]: I1125 15:25:39.000056 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"45dfffe7847fdd74ddf209ef2c7679d7cb94a3aabf8924763e3c95576de0a914"} Nov 25 15:25:39 crc kubenswrapper[4890]: I1125 15:25:39.003337 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 25 15:25:39 crc kubenswrapper[4890]: I1125 15:25:39.003401 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b68b7a22193a1042f7b2a06df452033ab04e929a3b8f18b668593749e140c576"} Nov 25 15:25:39 crc kubenswrapper[4890]: I1125 15:25:39.938437 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-66fc8c7fbf-lnjxm" Nov 25 15:25:39 crc kubenswrapper[4890]: I1125 15:25:39.939519 4890 scope.go:117] "RemoveContainer" containerID="be30ac4807bb1b8659327603f58cc562ca5c318c9246dbbc3a89d446e0498389" Nov 25 15:25:39 crc kubenswrapper[4890]: E1125 15:25:39.939790 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=metallb-operator-controller-manager-66fc8c7fbf-lnjxm_metallb-system(dad943fe-2209-40f9-a9bb-0be7ba80cca2)\"" pod="metallb-system/metallb-operator-controller-manager-66fc8c7fbf-lnjxm" podUID="dad943fe-2209-40f9-a9bb-0be7ba80cca2" Nov 25 15:25:40 crc kubenswrapper[4890]: I1125 15:25:40.022820 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"55cdd8f90323eea1a883b22d919c5a77bda9634388a99b8a96da50d10823dcf6"} Nov 25 15:25:40 crc kubenswrapper[4890]: I1125 15:25:40.022875 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"bf873727fd951bd9b34b04b0e456d1a776e08140ab70f036227e14e7993b8b7c"} Nov 25 15:25:41 crc kubenswrapper[4890]: I1125 15:25:41.031144 4890 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5bd5e648-6322-404d-bf1b-3b59a8092291" Nov 25 15:25:41 crc kubenswrapper[4890]: I1125 15:25:41.031501 4890 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5bd5e648-6322-404d-bf1b-3b59a8092291" Nov 25 15:25:41 crc kubenswrapper[4890]: I1125 15:25:41.031193 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 15:25:41 crc kubenswrapper[4890]: I1125 15:25:41.040983 4890 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 15:25:41 crc kubenswrapper[4890]: I1125 15:25:41.200040 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 15:25:41 crc kubenswrapper[4890]: I1125 15:25:41.200122 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 15:25:41 crc kubenswrapper[4890]: I1125 15:25:41.206239 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 15:25:42 crc kubenswrapper[4890]: I1125 15:25:42.040755 4890 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5bd5e648-6322-404d-bf1b-3b59a8092291" Nov 25 15:25:42 crc kubenswrapper[4890]: I1125 15:25:42.040798 4890 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5bd5e648-6322-404d-bf1b-3b59a8092291" Nov 25 15:25:42 crc kubenswrapper[4890]: I1125 15:25:42.045291 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 15:25:43 crc kubenswrapper[4890]: I1125 15:25:43.054804 4890 generic.go:334] "Generic (PLEG): container finished" podID="076c2a0e-7fdb-42e1-ab73-e296bba2b7a9" containerID="5d003110096be8e35d99eab0930c5aa6a3915752a25f20c2a6536a69af3d3afa" exitCode=1 Nov 25 15:25:43 crc kubenswrapper[4890]: I1125 15:25:43.055413 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-lfnwl" event={"ID":"076c2a0e-7fdb-42e1-ab73-e296bba2b7a9","Type":"ContainerDied","Data":"5d003110096be8e35d99eab0930c5aa6a3915752a25f20c2a6536a69af3d3afa"} Nov 25 15:25:43 crc kubenswrapper[4890]: I1125 15:25:43.055930 4890 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5bd5e648-6322-404d-bf1b-3b59a8092291" Nov 25 15:25:43 crc kubenswrapper[4890]: I1125 15:25:43.055949 4890 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5bd5e648-6322-404d-bf1b-3b59a8092291" Nov 25 15:25:43 crc kubenswrapper[4890]: I1125 15:25:43.056724 4890 scope.go:117] "RemoveContainer" containerID="5d003110096be8e35d99eab0930c5aa6a3915752a25f20c2a6536a69af3d3afa" Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.067114 4890 generic.go:334] "Generic (PLEG): container finished" podID="4364eec5-444a-41f6-af1c-fe16b24bd044" containerID="cc917ad88eaa4725e8cdac41f0da61db2690d8f49ca913c86a6c76f895cef181" exitCode=1 Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.067194 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-829nk" event={"ID":"4364eec5-444a-41f6-af1c-fe16b24bd044","Type":"ContainerDied","Data":"cc917ad88eaa4725e8cdac41f0da61db2690d8f49ca913c86a6c76f895cef181"} Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.068204 4890 scope.go:117] "RemoveContainer" containerID="cc917ad88eaa4725e8cdac41f0da61db2690d8f49ca913c86a6c76f895cef181" Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.069627 4890 generic.go:334] "Generic (PLEG): container finished" podID="679b3bf3-1ebc-4d19-9ad2-588bd41565dd" containerID="9cb6b91fbb6bd830583da0085dd152e38f130ea75d9d3d4475674031434bf81c" exitCode=1 Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.069693 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-hxzf4" event={"ID":"679b3bf3-1ebc-4d19-9ad2-588bd41565dd","Type":"ContainerDied","Data":"9cb6b91fbb6bd830583da0085dd152e38f130ea75d9d3d4475674031434bf81c"} Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.070511 4890 scope.go:117] "RemoveContainer" containerID="9cb6b91fbb6bd830583da0085dd152e38f130ea75d9d3d4475674031434bf81c" Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.071436 4890 generic.go:334] "Generic (PLEG): container finished" podID="72c26c5b-3370-4994-a5f0-82128c62c2df" containerID="544c8c93f950a78e83efe0178f366fc587812386594cb297859d064206d045e1" exitCode=1 Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.071470 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-774b86978c-xvfgd" event={"ID":"72c26c5b-3370-4994-a5f0-82128c62c2df","Type":"ContainerDied","Data":"544c8c93f950a78e83efe0178f366fc587812386594cb297859d064206d045e1"} Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.071915 4890 scope.go:117] "RemoveContainer" containerID="544c8c93f950a78e83efe0178f366fc587812386594cb297859d064206d045e1" Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.076237 4890 generic.go:334] "Generic (PLEG): container finished" podID="0bdbada3-e08c-4f8a-bd55-bac955d4370d" containerID="a3d62940c9f4d8d0e7c33babf91f1142db44aa39a25637bc44e7832b2e2c5c86" exitCode=1 Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.076316 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-jcw7x" event={"ID":"0bdbada3-e08c-4f8a-bd55-bac955d4370d","Type":"ContainerDied","Data":"a3d62940c9f4d8d0e7c33babf91f1142db44aa39a25637bc44e7832b2e2c5c86"} Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.077053 4890 scope.go:117] "RemoveContainer" containerID="a3d62940c9f4d8d0e7c33babf91f1142db44aa39a25637bc44e7832b2e2c5c86" Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.080475 4890 generic.go:334] "Generic (PLEG): container finished" podID="9f0e3cf9-dcc0-405f-a2de-9148844ec3ae" containerID="cc0132b2f896b99524175db61d9c1524df2e54f587019f94d706da2cdea210cc" exitCode=1 Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.080532 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-nshw7" event={"ID":"9f0e3cf9-dcc0-405f-a2de-9148844ec3ae","Type":"ContainerDied","Data":"cc0132b2f896b99524175db61d9c1524df2e54f587019f94d706da2cdea210cc"} Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.081102 4890 scope.go:117] "RemoveContainer" containerID="cc0132b2f896b99524175db61d9c1524df2e54f587019f94d706da2cdea210cc" Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.083687 4890 generic.go:334] "Generic (PLEG): container finished" podID="c528348f-0907-423a-a0c2-f15cf18ff99a" containerID="dbe71c9a5936928af6baf15ad268bc19748f775a26475ca9db52499631dbeef9" exitCode=1 Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.083748 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-j6pqs" event={"ID":"c528348f-0907-423a-a0c2-f15cf18ff99a","Type":"ContainerDied","Data":"dbe71c9a5936928af6baf15ad268bc19748f775a26475ca9db52499631dbeef9"} Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.084367 4890 scope.go:117] "RemoveContainer" containerID="dbe71c9a5936928af6baf15ad268bc19748f775a26475ca9db52499631dbeef9" Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.085190 4890 generic.go:334] "Generic (PLEG): container finished" podID="10534635-efe1-4e20-8c9b-9cbea73fbe87" containerID="eb7398300eba18c949ded5060fd4b31292e20d51a55d721c612ff62b864206f4" exitCode=1 Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.085234 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-5lr7r" event={"ID":"10534635-efe1-4e20-8c9b-9cbea73fbe87","Type":"ContainerDied","Data":"eb7398300eba18c949ded5060fd4b31292e20d51a55d721c612ff62b864206f4"} Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.085537 4890 scope.go:117] "RemoveContainer" containerID="eb7398300eba18c949ded5060fd4b31292e20d51a55d721c612ff62b864206f4" Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.090327 4890 generic.go:334] "Generic (PLEG): container finished" podID="a5d8a8ba-50c0-4fb6-823f-f28cf0d0f125" containerID="8756ce10612e1443f132d473e033dbcdfe66560ddde56b5f30ba360b8c363acb" exitCode=1 Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.090416 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-rs6hx" event={"ID":"a5d8a8ba-50c0-4fb6-823f-f28cf0d0f125","Type":"ContainerDied","Data":"8756ce10612e1443f132d473e033dbcdfe66560ddde56b5f30ba360b8c363acb"} Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.091200 4890 scope.go:117] "RemoveContainer" containerID="8756ce10612e1443f132d473e033dbcdfe66560ddde56b5f30ba360b8c363acb" Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.094547 4890 generic.go:334] "Generic (PLEG): container finished" podID="d175a35b-8b9b-49c0-8fc2-02ef62f131d9" containerID="1eda77f255a83e5efb319f627a4f7cd146d6e94ac18d2411cf044b47387b7986" exitCode=1 Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.094602 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-pmmf2" event={"ID":"d175a35b-8b9b-49c0-8fc2-02ef62f131d9","Type":"ContainerDied","Data":"1eda77f255a83e5efb319f627a4f7cd146d6e94ac18d2411cf044b47387b7986"} Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.095206 4890 scope.go:117] "RemoveContainer" containerID="1eda77f255a83e5efb319f627a4f7cd146d6e94ac18d2411cf044b47387b7986" Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.097888 4890 generic.go:334] "Generic (PLEG): container finished" podID="29f250db-2a96-41ad-bb8f-5b503a9288d8" containerID="75820db98fdfa144b9503a1d7e8750f8d454f7258242e932a8932d40b2582263" exitCode=1 Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.097930 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-vx75t" event={"ID":"29f250db-2a96-41ad-bb8f-5b503a9288d8","Type":"ContainerDied","Data":"75820db98fdfa144b9503a1d7e8750f8d454f7258242e932a8932d40b2582263"} Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.098366 4890 scope.go:117] "RemoveContainer" containerID="75820db98fdfa144b9503a1d7e8750f8d454f7258242e932a8932d40b2582263" Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.103581 4890 generic.go:334] "Generic (PLEG): container finished" podID="f272410f-230d-43f2-946e-e34e4d9c05ea" containerID="8e7f0def045216a15a744d26de933e25b9a2e25ac620c03d95240aa41cc03038" exitCode=1 Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.103667 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-slpt4" event={"ID":"f272410f-230d-43f2-946e-e34e4d9c05ea","Type":"ContainerDied","Data":"8e7f0def045216a15a744d26de933e25b9a2e25ac620c03d95240aa41cc03038"} Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.104395 4890 scope.go:117] "RemoveContainer" containerID="8e7f0def045216a15a744d26de933e25b9a2e25ac620c03d95240aa41cc03038" Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.107640 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-lfnwl" event={"ID":"076c2a0e-7fdb-42e1-ab73-e296bba2b7a9","Type":"ContainerStarted","Data":"57328d9eeb31abecef2987ab0e98f3bd5485e45dbe190b251c30dc54effefe13"} Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.108243 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-lfnwl" Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.112925 4890 generic.go:334] "Generic (PLEG): container finished" podID="9637fded-765d-4037-be8f-82e59ffdfc14" containerID="8fb4dbced7063fa7a276b9f33f0f1924462d950c851895f4c391b9a313b109b4" exitCode=1 Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.112990 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cb74df96-6mqt2" event={"ID":"9637fded-765d-4037-be8f-82e59ffdfc14","Type":"ContainerDied","Data":"8fb4dbced7063fa7a276b9f33f0f1924462d950c851895f4c391b9a313b109b4"} Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.113623 4890 scope.go:117] "RemoveContainer" containerID="8fb4dbced7063fa7a276b9f33f0f1924462d950c851895f4c391b9a313b109b4" Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.123919 4890 generic.go:334] "Generic (PLEG): container finished" podID="7e576602-19f5-4656-84dd-5ab80abca895" containerID="aa72b2ab7e7b9beabe0d4dc521c692a9eb7688f8f0273d14f7a29938b97646c9" exitCode=1 Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.124032 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-stflb" event={"ID":"7e576602-19f5-4656-84dd-5ab80abca895","Type":"ContainerDied","Data":"aa72b2ab7e7b9beabe0d4dc521c692a9eb7688f8f0273d14f7a29938b97646c9"} Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.124712 4890 scope.go:117] "RemoveContainer" containerID="aa72b2ab7e7b9beabe0d4dc521c692a9eb7688f8f0273d14f7a29938b97646c9" Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.130063 4890 generic.go:334] "Generic (PLEG): container finished" podID="77a5e8f9-92ab-48b5-be19-29a7a8e5df49" containerID="d9880e46a52dd26c73643cdd5427eeee12138ceffa656e9d825b0e88a70d2335" exitCode=1 Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.130134 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-stwjm" event={"ID":"77a5e8f9-92ab-48b5-be19-29a7a8e5df49","Type":"ContainerDied","Data":"d9880e46a52dd26c73643cdd5427eeee12138ceffa656e9d825b0e88a70d2335"} Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.130869 4890 scope.go:117] "RemoveContainer" containerID="d9880e46a52dd26c73643cdd5427eeee12138ceffa656e9d825b0e88a70d2335" Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.134151 4890 generic.go:334] "Generic (PLEG): container finished" podID="d50aec9d-1673-4b81-bb57-04cae418843e" containerID="b281ef61e39f8d5a43f4b56119cbfae16eb18c923bfe412c39c180018be2370e" exitCode=1 Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.134235 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-6zr4x" event={"ID":"d50aec9d-1673-4b81-bb57-04cae418843e","Type":"ContainerDied","Data":"b281ef61e39f8d5a43f4b56119cbfae16eb18c923bfe412c39c180018be2370e"} Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.138625 4890 scope.go:117] "RemoveContainer" containerID="b281ef61e39f8d5a43f4b56119cbfae16eb18c923bfe412c39c180018be2370e" Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.149292 4890 generic.go:334] "Generic (PLEG): container finished" podID="3e496c3d-9485-44bc-b3b1-dc1682a88434" containerID="25d97eb8b0662c62e96b5e278d3fba5c2fb55235f072e6116280714f791a4c15" exitCode=1 Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.149351 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-68587559f4-9b4cf" event={"ID":"3e496c3d-9485-44bc-b3b1-dc1682a88434","Type":"ContainerDied","Data":"25d97eb8b0662c62e96b5e278d3fba5c2fb55235f072e6116280714f791a4c15"} Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.150374 4890 scope.go:117] "RemoveContainer" containerID="25d97eb8b0662c62e96b5e278d3fba5c2fb55235f072e6116280714f791a4c15" Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.296574 4890 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="6101713b-0389-4ac1-adb1-ca660ca84eba" Nov 25 15:25:44 crc kubenswrapper[4890]: I1125 15:25:44.344407 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 15:25:45 crc kubenswrapper[4890]: I1125 15:25:45.228816 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-5lr7r" event={"ID":"10534635-efe1-4e20-8c9b-9cbea73fbe87","Type":"ContainerStarted","Data":"53ce7419c37ea35c3fa85593635742afdc4fe4db5d8a4c08b2330354ae59906e"} Nov 25 15:25:45 crc kubenswrapper[4890]: I1125 15:25:45.233458 4890 generic.go:334] "Generic (PLEG): container finished" podID="0769cde4-b49f-41c5-b720-e0dbc667eb4a" containerID="97e02800cd95b6b08806006fece844f452d79638de429148891a1ede18c31c8f" exitCode=1 Nov 25 15:25:45 crc kubenswrapper[4890]: I1125 15:25:45.233571 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6jq7q" event={"ID":"0769cde4-b49f-41c5-b720-e0dbc667eb4a","Type":"ContainerDied","Data":"97e02800cd95b6b08806006fece844f452d79638de429148891a1ede18c31c8f"} Nov 25 15:25:45 crc kubenswrapper[4890]: I1125 15:25:45.234471 4890 scope.go:117] "RemoveContainer" containerID="97e02800cd95b6b08806006fece844f452d79638de429148891a1ede18c31c8f" Nov 25 15:25:45 crc kubenswrapper[4890]: I1125 15:25:45.250012 4890 generic.go:334] "Generic (PLEG): container finished" podID="861317e1-222d-4f14-b931-e9e7d100ebd6" containerID="b8f781558438b476e77ca8c2495fc65050de7477ff54cb3185dc7dd498e07866" exitCode=1 Nov 25 15:25:45 crc kubenswrapper[4890]: I1125 15:25:45.250127 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-dzkgt" event={"ID":"861317e1-222d-4f14-b931-e9e7d100ebd6","Type":"ContainerDied","Data":"b8f781558438b476e77ca8c2495fc65050de7477ff54cb3185dc7dd498e07866"} Nov 25 15:25:45 crc kubenswrapper[4890]: I1125 15:25:45.251749 4890 scope.go:117] "RemoveContainer" containerID="b8f781558438b476e77ca8c2495fc65050de7477ff54cb3185dc7dd498e07866" Nov 25 15:25:45 crc kubenswrapper[4890]: I1125 15:25:45.272576 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-vx75t" event={"ID":"29f250db-2a96-41ad-bb8f-5b503a9288d8","Type":"ContainerStarted","Data":"0770c2f791722b326d3323baa96eca63ec6eac08347883ea0d8f92a8abb2ed7d"} Nov 25 15:25:45 crc kubenswrapper[4890]: I1125 15:25:45.280436 4890 generic.go:334] "Generic (PLEG): container finished" podID="076c2a0e-7fdb-42e1-ab73-e296bba2b7a9" containerID="57328d9eeb31abecef2987ab0e98f3bd5485e45dbe190b251c30dc54effefe13" exitCode=1 Nov 25 15:25:45 crc kubenswrapper[4890]: I1125 15:25:45.280600 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-lfnwl" event={"ID":"076c2a0e-7fdb-42e1-ab73-e296bba2b7a9","Type":"ContainerDied","Data":"57328d9eeb31abecef2987ab0e98f3bd5485e45dbe190b251c30dc54effefe13"} Nov 25 15:25:45 crc kubenswrapper[4890]: I1125 15:25:45.280668 4890 scope.go:117] "RemoveContainer" containerID="5d003110096be8e35d99eab0930c5aa6a3915752a25f20c2a6536a69af3d3afa" Nov 25 15:25:45 crc kubenswrapper[4890]: I1125 15:25:45.281853 4890 scope.go:117] "RemoveContainer" containerID="57328d9eeb31abecef2987ab0e98f3bd5485e45dbe190b251c30dc54effefe13" Nov 25 15:25:45 crc kubenswrapper[4890]: E1125 15:25:45.282203 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=glance-operator-controller-manager-68b95954c9-lfnwl_openstack-operators(076c2a0e-7fdb-42e1-ab73-e296bba2b7a9)\"" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-lfnwl" podUID="076c2a0e-7fdb-42e1-ab73-e296bba2b7a9" Nov 25 15:25:45 crc kubenswrapper[4890]: I1125 15:25:45.289885 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-hxzf4" event={"ID":"679b3bf3-1ebc-4d19-9ad2-588bd41565dd","Type":"ContainerStarted","Data":"6de55425fa91945cc6e8f62649df335dd495a944f0f7aed9121c87e80c8f3a96"} Nov 25 15:25:45 crc kubenswrapper[4890]: I1125 15:25:45.295453 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-j6pqs" event={"ID":"c528348f-0907-423a-a0c2-f15cf18ff99a","Type":"ContainerStarted","Data":"52f8db982762a517d8d96beb141da93c238d379ce39501ef137cb21b7fdcf14d"} Nov 25 15:25:45 crc kubenswrapper[4890]: I1125 15:25:45.301083 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-pmmf2" event={"ID":"d175a35b-8b9b-49c0-8fc2-02ef62f131d9","Type":"ContainerStarted","Data":"e0214f8d055e3e0e07dd810288a0139e23c1d5a5bcfdaa0457141e014e32d69d"} Nov 25 15:25:45 crc kubenswrapper[4890]: I1125 15:25:45.306106 4890 generic.go:334] "Generic (PLEG): container finished" podID="4364eec5-444a-41f6-af1c-fe16b24bd044" containerID="f516cdea0e4e13c5f35f9cf192df13cf34b38f3a05d2fb0db0fb898037f8319c" exitCode=1 Nov 25 15:25:45 crc kubenswrapper[4890]: I1125 15:25:45.306151 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-829nk" event={"ID":"4364eec5-444a-41f6-af1c-fe16b24bd044","Type":"ContainerDied","Data":"f516cdea0e4e13c5f35f9cf192df13cf34b38f3a05d2fb0db0fb898037f8319c"} Nov 25 15:25:45 crc kubenswrapper[4890]: I1125 15:25:45.308364 4890 scope.go:117] "RemoveContainer" containerID="f516cdea0e4e13c5f35f9cf192df13cf34b38f3a05d2fb0db0fb898037f8319c" Nov 25 15:25:45 crc kubenswrapper[4890]: E1125 15:25:45.308760 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=neutron-operator-controller-manager-7c57c8bbc4-829nk_openstack-operators(4364eec5-444a-41f6-af1c-fe16b24bd044)\"" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-829nk" podUID="4364eec5-444a-41f6-af1c-fe16b24bd044" Nov 25 15:25:45 crc kubenswrapper[4890]: I1125 15:25:45.743918 4890 scope.go:117] "RemoveContainer" containerID="cc917ad88eaa4725e8cdac41f0da61db2690d8f49ca913c86a6c76f895cef181" Nov 25 15:25:45 crc kubenswrapper[4890]: I1125 15:25:45.750825 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/kube-state-metrics-0" podUID="a47203e9-a88a-42f7-86e1-164c6ad532b5" containerName="kube-state-metrics" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 25 15:25:45 crc kubenswrapper[4890]: I1125 15:25:45.750893 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/kube-state-metrics-0" Nov 25 15:25:45 crc kubenswrapper[4890]: I1125 15:25:45.752048 4890 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-state-metrics" containerStatusID={"Type":"cri-o","ID":"f6c757f7b337424920e92dc7103283ff1d00267dabeb93ed84f34ff593609e9a"} pod="openstack/kube-state-metrics-0" containerMessage="Container kube-state-metrics failed liveness probe, will be restarted" Nov 25 15:25:45 crc kubenswrapper[4890]: I1125 15:25:45.752091 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="a47203e9-a88a-42f7-86e1-164c6ad532b5" containerName="kube-state-metrics" containerID="cri-o://f6c757f7b337424920e92dc7103283ff1d00267dabeb93ed84f34ff593609e9a" gracePeriod=30 Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.110336 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-pmmf2" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.111437 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-pmmf2" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.116962 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-jcw7x" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.117040 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-jcw7x" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.127804 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-stwjm" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.131018 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-stwjm" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.157657 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-lfnwl" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.207922 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-774b86978c-xvfgd" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.208331 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/heat-operator-controller-manager-774b86978c-xvfgd" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.223039 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-j6pqs" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.223094 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-j6pqs" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.301697 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-hxzf4" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.301748 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-hxzf4" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.315244 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-rs6hx" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.315327 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-rs6hx" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.320664 4890 generic.go:334] "Generic (PLEG): container finished" podID="f272410f-230d-43f2-946e-e34e4d9c05ea" containerID="6328b6da81615173e561e6807d29b480ebef9c78fbca0d829f62256dd629eac3" exitCode=1 Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.320753 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-slpt4" event={"ID":"f272410f-230d-43f2-946e-e34e4d9c05ea","Type":"ContainerDied","Data":"6328b6da81615173e561e6807d29b480ebef9c78fbca0d829f62256dd629eac3"} Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.321769 4890 scope.go:117] "RemoveContainer" containerID="6328b6da81615173e561e6807d29b480ebef9c78fbca0d829f62256dd629eac3" Nov 25 15:25:46 crc kubenswrapper[4890]: E1125 15:25:46.322059 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=manila-operator-controller-manager-58bb8d67cc-slpt4_openstack-operators(f272410f-230d-43f2-946e-e34e4d9c05ea)\"" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-slpt4" podUID="f272410f-230d-43f2-946e-e34e4d9c05ea" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.323937 4890 scope.go:117] "RemoveContainer" containerID="57328d9eeb31abecef2987ab0e98f3bd5485e45dbe190b251c30dc54effefe13" Nov 25 15:25:46 crc kubenswrapper[4890]: E1125 15:25:46.324194 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=glance-operator-controller-manager-68b95954c9-lfnwl_openstack-operators(076c2a0e-7fdb-42e1-ab73-e296bba2b7a9)\"" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-lfnwl" podUID="076c2a0e-7fdb-42e1-ab73-e296bba2b7a9" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.328818 4890 generic.go:334] "Generic (PLEG): container finished" podID="e26594ec-4bc2-4051-8a49-fc490bd3b6c9" containerID="61b9be5d90ce8a666acdd758f569aaa247bc6bf33719297ce9c65e886c773f90" exitCode=1 Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.328880 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-dmlx4" event={"ID":"e26594ec-4bc2-4051-8a49-fc490bd3b6c9","Type":"ContainerDied","Data":"61b9be5d90ce8a666acdd758f569aaa247bc6bf33719297ce9c65e886c773f90"} Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.329249 4890 scope.go:117] "RemoveContainer" containerID="61b9be5d90ce8a666acdd758f569aaa247bc6bf33719297ce9c65e886c773f90" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.333533 4890 generic.go:334] "Generic (PLEG): container finished" podID="0bdbada3-e08c-4f8a-bd55-bac955d4370d" containerID="cce83adafd4303a3c4644332384918e2da8852bc205228c4dadd7e6a87ff1cf8" exitCode=1 Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.333623 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-jcw7x" event={"ID":"0bdbada3-e08c-4f8a-bd55-bac955d4370d","Type":"ContainerDied","Data":"cce83adafd4303a3c4644332384918e2da8852bc205228c4dadd7e6a87ff1cf8"} Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.334283 4890 scope.go:117] "RemoveContainer" containerID="cce83adafd4303a3c4644332384918e2da8852bc205228c4dadd7e6a87ff1cf8" Nov 25 15:25:46 crc kubenswrapper[4890]: E1125 15:25:46.334534 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=cinder-operator-controller-manager-79856dc55c-jcw7x_openstack-operators(0bdbada3-e08c-4f8a-bd55-bac955d4370d)\"" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-jcw7x" podUID="0bdbada3-e08c-4f8a-bd55-bac955d4370d" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.338869 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-dzkgt" event={"ID":"861317e1-222d-4f14-b931-e9e7d100ebd6","Type":"ContainerStarted","Data":"7ff93d6707b3f0b8fb868414ecf039dfc3ca37a71aa91bcc35b9114d425a199f"} Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.344667 4890 generic.go:334] "Generic (PLEG): container finished" podID="679b3bf3-1ebc-4d19-9ad2-588bd41565dd" containerID="6de55425fa91945cc6e8f62649df335dd495a944f0f7aed9121c87e80c8f3a96" exitCode=1 Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.344744 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-hxzf4" event={"ID":"679b3bf3-1ebc-4d19-9ad2-588bd41565dd","Type":"ContainerDied","Data":"6de55425fa91945cc6e8f62649df335dd495a944f0f7aed9121c87e80c8f3a96"} Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.345445 4890 scope.go:117] "RemoveContainer" containerID="6de55425fa91945cc6e8f62649df335dd495a944f0f7aed9121c87e80c8f3a96" Nov 25 15:25:46 crc kubenswrapper[4890]: E1125 15:25:46.345704 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=ironic-operator-controller-manager-5bfcdc958c-hxzf4_openstack-operators(679b3bf3-1ebc-4d19-9ad2-588bd41565dd)\"" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-hxzf4" podUID="679b3bf3-1ebc-4d19-9ad2-588bd41565dd" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.346859 4890 generic.go:334] "Generic (PLEG): container finished" podID="a9efeabb-bf1c-4de2-8322-961b284e95cc" containerID="360f2cd67207fde2598cf2336d7ee654c1e558557dae8116644fbbad97f92bbe" exitCode=1 Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.346931 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-56bcr" event={"ID":"a9efeabb-bf1c-4de2-8322-961b284e95cc","Type":"ContainerDied","Data":"360f2cd67207fde2598cf2336d7ee654c1e558557dae8116644fbbad97f92bbe"} Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.347286 4890 scope.go:117] "RemoveContainer" containerID="360f2cd67207fde2598cf2336d7ee654c1e558557dae8116644fbbad97f92bbe" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.349701 4890 generic.go:334] "Generic (PLEG): container finished" podID="72c26c5b-3370-4994-a5f0-82128c62c2df" containerID="77f771985277fc92c381c09a36784838a56dce95cd158b58ae0dd79c46671845" exitCode=1 Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.349813 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-774b86978c-xvfgd" event={"ID":"72c26c5b-3370-4994-a5f0-82128c62c2df","Type":"ContainerDied","Data":"77f771985277fc92c381c09a36784838a56dce95cd158b58ae0dd79c46671845"} Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.350356 4890 scope.go:117] "RemoveContainer" containerID="77f771985277fc92c381c09a36784838a56dce95cd158b58ae0dd79c46671845" Nov 25 15:25:46 crc kubenswrapper[4890]: E1125 15:25:46.350590 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=heat-operator-controller-manager-774b86978c-xvfgd_openstack-operators(72c26c5b-3370-4994-a5f0-82128c62c2df)\"" pod="openstack-operators/heat-operator-controller-manager-774b86978c-xvfgd" podUID="72c26c5b-3370-4994-a5f0-82128c62c2df" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.355787 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-68587559f4-9b4cf" event={"ID":"3e496c3d-9485-44bc-b3b1-dc1682a88434","Type":"ContainerStarted","Data":"5bdc10c6af2349a7635d97fa9053a19229cec02d80cae8cde73fc72b252787eb"} Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.357150 4890 generic.go:334] "Generic (PLEG): container finished" podID="a47203e9-a88a-42f7-86e1-164c6ad532b5" containerID="f6c757f7b337424920e92dc7103283ff1d00267dabeb93ed84f34ff593609e9a" exitCode=2 Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.357227 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"a47203e9-a88a-42f7-86e1-164c6ad532b5","Type":"ContainerDied","Data":"f6c757f7b337424920e92dc7103283ff1d00267dabeb93ed84f34ff593609e9a"} Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.359009 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6jq7q" event={"ID":"0769cde4-b49f-41c5-b720-e0dbc667eb4a","Type":"ContainerStarted","Data":"70634ed74a9c04a3a724fdb83aff92615249c19958b69dd169263857756bf093"} Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.364886 4890 generic.go:334] "Generic (PLEG): container finished" podID="03353c8b-91f4-41ae-afc1-c522b1c73bb4" containerID="820303aa9ed50019579596b6edb70067e3cf2b1af539efec84d656ec8ffdf5f5" exitCode=1 Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.364929 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-wvjw7" event={"ID":"03353c8b-91f4-41ae-afc1-c522b1c73bb4","Type":"ContainerDied","Data":"820303aa9ed50019579596b6edb70067e3cf2b1af539efec84d656ec8ffdf5f5"} Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.365511 4890 scope.go:117] "RemoveContainer" containerID="820303aa9ed50019579596b6edb70067e3cf2b1af539efec84d656ec8ffdf5f5" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.374830 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-stwjm" event={"ID":"77a5e8f9-92ab-48b5-be19-29a7a8e5df49","Type":"ContainerStarted","Data":"e7b7fcdef297e60c89760ed3c86c2d93e8c7a6a0088d065d1090410732523bb5"} Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.385459 4890 generic.go:334] "Generic (PLEG): container finished" podID="29f250db-2a96-41ad-bb8f-5b503a9288d8" containerID="0770c2f791722b326d3323baa96eca63ec6eac08347883ea0d8f92a8abb2ed7d" exitCode=1 Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.385562 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-vx75t" event={"ID":"29f250db-2a96-41ad-bb8f-5b503a9288d8","Type":"ContainerDied","Data":"0770c2f791722b326d3323baa96eca63ec6eac08347883ea0d8f92a8abb2ed7d"} Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.386180 4890 scope.go:117] "RemoveContainer" containerID="0770c2f791722b326d3323baa96eca63ec6eac08347883ea0d8f92a8abb2ed7d" Nov 25 15:25:46 crc kubenswrapper[4890]: E1125 15:25:46.386484 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=ovn-operator-controller-manager-66cf5c67ff-vx75t_openstack-operators(29f250db-2a96-41ad-bb8f-5b503a9288d8)\"" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-vx75t" podUID="29f250db-2a96-41ad-bb8f-5b503a9288d8" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.390008 4890 generic.go:334] "Generic (PLEG): container finished" podID="c528348f-0907-423a-a0c2-f15cf18ff99a" containerID="52f8db982762a517d8d96beb141da93c238d379ce39501ef137cb21b7fdcf14d" exitCode=1 Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.390083 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-j6pqs" event={"ID":"c528348f-0907-423a-a0c2-f15cf18ff99a","Type":"ContainerDied","Data":"52f8db982762a517d8d96beb141da93c238d379ce39501ef137cb21b7fdcf14d"} Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.391130 4890 scope.go:117] "RemoveContainer" containerID="52f8db982762a517d8d96beb141da93c238d379ce39501ef137cb21b7fdcf14d" Nov 25 15:25:46 crc kubenswrapper[4890]: E1125 15:25:46.391414 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=horizon-operator-controller-manager-68c9694994-j6pqs_openstack-operators(c528348f-0907-423a-a0c2-f15cf18ff99a)\"" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-j6pqs" podUID="c528348f-0907-423a-a0c2-f15cf18ff99a" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.394509 4890 generic.go:334] "Generic (PLEG): container finished" podID="d175a35b-8b9b-49c0-8fc2-02ef62f131d9" containerID="e0214f8d055e3e0e07dd810288a0139e23c1d5a5bcfdaa0457141e014e32d69d" exitCode=1 Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.394581 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-pmmf2" event={"ID":"d175a35b-8b9b-49c0-8fc2-02ef62f131d9","Type":"ContainerDied","Data":"e0214f8d055e3e0e07dd810288a0139e23c1d5a5bcfdaa0457141e014e32d69d"} Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.395318 4890 scope.go:117] "RemoveContainer" containerID="e0214f8d055e3e0e07dd810288a0139e23c1d5a5bcfdaa0457141e014e32d69d" Nov 25 15:25:46 crc kubenswrapper[4890]: E1125 15:25:46.395578 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=barbican-operator-controller-manager-86dc4d89c8-pmmf2_openstack-operators(d175a35b-8b9b-49c0-8fc2-02ef62f131d9)\"" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-pmmf2" podUID="d175a35b-8b9b-49c0-8fc2-02ef62f131d9" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.399394 4890 generic.go:334] "Generic (PLEG): container finished" podID="7e576602-19f5-4656-84dd-5ab80abca895" containerID="01cc934a2ae0af3a25c372a4e138035e62b92af8b615443d315c5b0d7695f5c1" exitCode=1 Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.399485 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-stflb" event={"ID":"7e576602-19f5-4656-84dd-5ab80abca895","Type":"ContainerDied","Data":"01cc934a2ae0af3a25c372a4e138035e62b92af8b615443d315c5b0d7695f5c1"} Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.400375 4890 scope.go:117] "RemoveContainer" containerID="01cc934a2ae0af3a25c372a4e138035e62b92af8b615443d315c5b0d7695f5c1" Nov 25 15:25:46 crc kubenswrapper[4890]: E1125 15:25:46.400709 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=infra-operator-controller-manager-d5cc86f4b-stflb_openstack-operators(7e576602-19f5-4656-84dd-5ab80abca895)\"" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-stflb" podUID="7e576602-19f5-4656-84dd-5ab80abca895" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.404105 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-6zr4x" event={"ID":"d50aec9d-1673-4b81-bb57-04cae418843e","Type":"ContainerStarted","Data":"286d09d1c4a9ff164ba4164079699a9e512eed1984749d1fee1c487e3968209c"} Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.412319 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-nshw7" event={"ID":"9f0e3cf9-dcc0-405f-a2de-9148844ec3ae","Type":"ContainerStarted","Data":"60ab9e903ab7dd461c9fab060610f14563535916392274aabb88f8d7b90adbde"} Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.413204 4890 scope.go:117] "RemoveContainer" containerID="60ab9e903ab7dd461c9fab060610f14563535916392274aabb88f8d7b90adbde" Nov 25 15:25:46 crc kubenswrapper[4890]: E1125 15:25:46.413513 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=placement-operator-controller-manager-5db546f9d9-nshw7_openstack-operators(9f0e3cf9-dcc0-405f-a2de-9148844ec3ae)\"" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-nshw7" podUID="9f0e3cf9-dcc0-405f-a2de-9148844ec3ae" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.416016 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cb74df96-6mqt2" event={"ID":"9637fded-765d-4037-be8f-82e59ffdfc14","Type":"ContainerStarted","Data":"08c8cf792781bf06cacad23e81fcf1345b078c8a8ec6feed22f67831a5c42e64"} Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.416229 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5cb74df96-6mqt2" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.420780 4890 generic.go:334] "Generic (PLEG): container finished" podID="a5d8a8ba-50c0-4fb6-823f-f28cf0d0f125" containerID="38d0123e0cedad944df1c84a66fc4aba91c298dc840c5489a45e6ec2fd23cc66" exitCode=1 Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.421632 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-rs6hx" event={"ID":"a5d8a8ba-50c0-4fb6-823f-f28cf0d0f125","Type":"ContainerDied","Data":"38d0123e0cedad944df1c84a66fc4aba91c298dc840c5489a45e6ec2fd23cc66"} Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.421816 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-5lr7r" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.421836 4890 scope.go:117] "RemoveContainer" containerID="38d0123e0cedad944df1c84a66fc4aba91c298dc840c5489a45e6ec2fd23cc66" Nov 25 15:25:46 crc kubenswrapper[4890]: E1125 15:25:46.422607 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=keystone-operator-controller-manager-748dc6576f-rs6hx_openstack-operators(a5d8a8ba-50c0-4fb6-823f-f28cf0d0f125)\"" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-rs6hx" podUID="a5d8a8ba-50c0-4fb6-823f-f28cf0d0f125" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.436895 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-wvjw7" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.436959 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-wvjw7" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.515986 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-slpt4" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.516058 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-slpt4" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.566774 4890 scope.go:117] "RemoveContainer" containerID="8e7f0def045216a15a744d26de933e25b9a2e25ac620c03d95240aa41cc03038" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.567639 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-6zr4x" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.567675 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-6zr4x" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.623227 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6jq7q" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.623464 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6jq7q" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.648722 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-56bcr" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.648768 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-56bcr" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.718838 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-vx75t" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.718874 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-vx75t" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.756742 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-nshw7" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.757008 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-nshw7" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.773515 4890 scope.go:117] "RemoveContainer" containerID="a3d62940c9f4d8d0e7c33babf91f1142db44aa39a25637bc44e7832b2e2c5c86" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.774875 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-dzkgt" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.775001 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-dzkgt" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.790026 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-829nk" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.790251 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-829nk" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.790967 4890 scope.go:117] "RemoveContainer" containerID="f516cdea0e4e13c5f35f9cf192df13cf34b38f3a05d2fb0db0fb898037f8319c" Nov 25 15:25:46 crc kubenswrapper[4890]: E1125 15:25:46.791303 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=neutron-operator-controller-manager-7c57c8bbc4-829nk_openstack-operators(4364eec5-444a-41f6-af1c-fe16b24bd044)\"" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-829nk" podUID="4364eec5-444a-41f6-af1c-fe16b24bd044" Nov 25 15:25:46 crc kubenswrapper[4890]: I1125 15:25:46.891539 4890 scope.go:117] "RemoveContainer" containerID="9cb6b91fbb6bd830583da0085dd152e38f130ea75d9d3d4475674031434bf81c" Nov 25 15:25:47 crc kubenswrapper[4890]: I1125 15:25:47.251804 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 15:25:47 crc kubenswrapper[4890]: I1125 15:25:47.252096 4890 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Nov 25 15:25:47 crc kubenswrapper[4890]: I1125 15:25:47.252128 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Nov 25 15:25:47 crc kubenswrapper[4890]: I1125 15:25:47.472458 4890 generic.go:334] "Generic (PLEG): container finished" podID="9f0e3cf9-dcc0-405f-a2de-9148844ec3ae" containerID="60ab9e903ab7dd461c9fab060610f14563535916392274aabb88f8d7b90adbde" exitCode=1 Nov 25 15:25:47 crc kubenswrapper[4890]: I1125 15:25:47.473625 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-nshw7" event={"ID":"9f0e3cf9-dcc0-405f-a2de-9148844ec3ae","Type":"ContainerDied","Data":"60ab9e903ab7dd461c9fab060610f14563535916392274aabb88f8d7b90adbde"} Nov 25 15:25:47 crc kubenswrapper[4890]: I1125 15:25:47.474367 4890 scope.go:117] "RemoveContainer" containerID="60ab9e903ab7dd461c9fab060610f14563535916392274aabb88f8d7b90adbde" Nov 25 15:25:47 crc kubenswrapper[4890]: E1125 15:25:47.474805 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=placement-operator-controller-manager-5db546f9d9-nshw7_openstack-operators(9f0e3cf9-dcc0-405f-a2de-9148844ec3ae)\"" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-nshw7" podUID="9f0e3cf9-dcc0-405f-a2de-9148844ec3ae" Nov 25 15:25:47 crc kubenswrapper[4890]: E1125 15:25:47.492212 4890 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda9efeabb_bf1c_4de2_8322_961b284e95cc.slice/crio-conmon-09913842764d83958bb7259fb1a01e61d3da9512eb07f23da446059eb49a046a.scope\": RecentStats: unable to find data in memory cache]" Nov 25 15:25:47 crc kubenswrapper[4890]: I1125 15:25:47.497690 4890 generic.go:334] "Generic (PLEG): container finished" podID="77a5e8f9-92ab-48b5-be19-29a7a8e5df49" containerID="e7b7fcdef297e60c89760ed3c86c2d93e8c7a6a0088d065d1090410732523bb5" exitCode=1 Nov 25 15:25:47 crc kubenswrapper[4890]: I1125 15:25:47.497776 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-stwjm" event={"ID":"77a5e8f9-92ab-48b5-be19-29a7a8e5df49","Type":"ContainerDied","Data":"e7b7fcdef297e60c89760ed3c86c2d93e8c7a6a0088d065d1090410732523bb5"} Nov 25 15:25:47 crc kubenswrapper[4890]: I1125 15:25:47.498455 4890 scope.go:117] "RemoveContainer" containerID="e7b7fcdef297e60c89760ed3c86c2d93e8c7a6a0088d065d1090410732523bb5" Nov 25 15:25:47 crc kubenswrapper[4890]: E1125 15:25:47.498682 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=designate-operator-controller-manager-7d695c9b56-stwjm_openstack-operators(77a5e8f9-92ab-48b5-be19-29a7a8e5df49)\"" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-stwjm" podUID="77a5e8f9-92ab-48b5-be19-29a7a8e5df49" Nov 25 15:25:47 crc kubenswrapper[4890]: I1125 15:25:47.513715 4890 scope.go:117] "RemoveContainer" containerID="6328b6da81615173e561e6807d29b480ebef9c78fbca0d829f62256dd629eac3" Nov 25 15:25:47 crc kubenswrapper[4890]: E1125 15:25:47.514010 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=manila-operator-controller-manager-58bb8d67cc-slpt4_openstack-operators(f272410f-230d-43f2-946e-e34e4d9c05ea)\"" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-slpt4" podUID="f272410f-230d-43f2-946e-e34e4d9c05ea" Nov 25 15:25:47 crc kubenswrapper[4890]: I1125 15:25:47.517298 4890 generic.go:334] "Generic (PLEG): container finished" podID="bd60efda-e354-44d0-af56-c110651bdbaf" containerID="f6e34e9f7f991c3d5b3dc786e507a602a630e025bed28294ab4e01d68a0244e4" exitCode=1 Nov 25 15:25:47 crc kubenswrapper[4890]: I1125 15:25:47.517368 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6d88ccc4fc-9smnn" event={"ID":"bd60efda-e354-44d0-af56-c110651bdbaf","Type":"ContainerDied","Data":"f6e34e9f7f991c3d5b3dc786e507a602a630e025bed28294ab4e01d68a0244e4"} Nov 25 15:25:47 crc kubenswrapper[4890]: I1125 15:25:47.517991 4890 scope.go:117] "RemoveContainer" containerID="f6e34e9f7f991c3d5b3dc786e507a602a630e025bed28294ab4e01d68a0244e4" Nov 25 15:25:47 crc kubenswrapper[4890]: I1125 15:25:47.531339 4890 scope.go:117] "RemoveContainer" containerID="544c8c93f950a78e83efe0178f366fc587812386594cb297859d064206d045e1" Nov 25 15:25:47 crc kubenswrapper[4890]: I1125 15:25:47.534384 4890 generic.go:334] "Generic (PLEG): container finished" podID="3e496c3d-9485-44bc-b3b1-dc1682a88434" containerID="5bdc10c6af2349a7635d97fa9053a19229cec02d80cae8cde73fc72b252787eb" exitCode=1 Nov 25 15:25:47 crc kubenswrapper[4890]: I1125 15:25:47.534434 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-68587559f4-9b4cf" event={"ID":"3e496c3d-9485-44bc-b3b1-dc1682a88434","Type":"ContainerDied","Data":"5bdc10c6af2349a7635d97fa9053a19229cec02d80cae8cde73fc72b252787eb"} Nov 25 15:25:47 crc kubenswrapper[4890]: I1125 15:25:47.535057 4890 scope.go:117] "RemoveContainer" containerID="5bdc10c6af2349a7635d97fa9053a19229cec02d80cae8cde73fc72b252787eb" Nov 25 15:25:47 crc kubenswrapper[4890]: E1125 15:25:47.535290 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=openstack-operator-controller-manager-68587559f4-9b4cf_openstack-operators(3e496c3d-9485-44bc-b3b1-dc1682a88434)\"" pod="openstack-operators/openstack-operator-controller-manager-68587559f4-9b4cf" podUID="3e496c3d-9485-44bc-b3b1-dc1682a88434" Nov 25 15:25:47 crc kubenswrapper[4890]: I1125 15:25:47.547124 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-dmlx4" event={"ID":"e26594ec-4bc2-4051-8a49-fc490bd3b6c9","Type":"ContainerStarted","Data":"45be6c4aa1f57c239660be552320d18f875cde9c70ec5abb3fa04a6ed3e9139e"} Nov 25 15:25:47 crc kubenswrapper[4890]: I1125 15:25:47.551269 4890 scope.go:117] "RemoveContainer" containerID="cce83adafd4303a3c4644332384918e2da8852bc205228c4dadd7e6a87ff1cf8" Nov 25 15:25:47 crc kubenswrapper[4890]: E1125 15:25:47.551565 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=cinder-operator-controller-manager-79856dc55c-jcw7x_openstack-operators(0bdbada3-e08c-4f8a-bd55-bac955d4370d)\"" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-jcw7x" podUID="0bdbada3-e08c-4f8a-bd55-bac955d4370d" Nov 25 15:25:47 crc kubenswrapper[4890]: I1125 15:25:47.569069 4890 scope.go:117] "RemoveContainer" containerID="70634ed74a9c04a3a724fdb83aff92615249c19958b69dd169263857756bf093" Nov 25 15:25:47 crc kubenswrapper[4890]: E1125 15:25:47.569642 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=nova-operator-controller-manager-79556f57fc-6jq7q_openstack-operators(0769cde4-b49f-41c5-b720-e0dbc667eb4a)\"" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6jq7q" podUID="0769cde4-b49f-41c5-b720-e0dbc667eb4a" Nov 25 15:25:47 crc kubenswrapper[4890]: I1125 15:25:47.569759 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6jq7q" event={"ID":"0769cde4-b49f-41c5-b720-e0dbc667eb4a","Type":"ContainerDied","Data":"70634ed74a9c04a3a724fdb83aff92615249c19958b69dd169263857756bf093"} Nov 25 15:25:47 crc kubenswrapper[4890]: I1125 15:25:47.569846 4890 generic.go:334] "Generic (PLEG): container finished" podID="0769cde4-b49f-41c5-b720-e0dbc667eb4a" containerID="70634ed74a9c04a3a724fdb83aff92615249c19958b69dd169263857756bf093" exitCode=1 Nov 25 15:25:47 crc kubenswrapper[4890]: I1125 15:25:47.581052 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-wvjw7" event={"ID":"03353c8b-91f4-41ae-afc1-c522b1c73bb4","Type":"ContainerStarted","Data":"34aad7779779b8785f00f0293acf127026ceeeb1d0bf6285415e611bc6e3be23"} Nov 25 15:25:47 crc kubenswrapper[4890]: I1125 15:25:47.581931 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-wvjw7" Nov 25 15:25:47 crc kubenswrapper[4890]: I1125 15:25:47.585835 4890 generic.go:334] "Generic (PLEG): container finished" podID="861317e1-222d-4f14-b931-e9e7d100ebd6" containerID="7ff93d6707b3f0b8fb868414ecf039dfc3ca37a71aa91bcc35b9114d425a199f" exitCode=1 Nov 25 15:25:47 crc kubenswrapper[4890]: I1125 15:25:47.585894 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-dzkgt" event={"ID":"861317e1-222d-4f14-b931-e9e7d100ebd6","Type":"ContainerDied","Data":"7ff93d6707b3f0b8fb868414ecf039dfc3ca37a71aa91bcc35b9114d425a199f"} Nov 25 15:25:47 crc kubenswrapper[4890]: I1125 15:25:47.586632 4890 scope.go:117] "RemoveContainer" containerID="7ff93d6707b3f0b8fb868414ecf039dfc3ca37a71aa91bcc35b9114d425a199f" Nov 25 15:25:47 crc kubenswrapper[4890]: E1125 15:25:47.586945 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=swift-operator-controller-manager-6fdc4fcf86-dzkgt_openstack-operators(861317e1-222d-4f14-b931-e9e7d100ebd6)\"" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-dzkgt" podUID="861317e1-222d-4f14-b931-e9e7d100ebd6" Nov 25 15:25:47 crc kubenswrapper[4890]: I1125 15:25:47.588803 4890 generic.go:334] "Generic (PLEG): container finished" podID="a9efeabb-bf1c-4de2-8322-961b284e95cc" containerID="09913842764d83958bb7259fb1a01e61d3da9512eb07f23da446059eb49a046a" exitCode=1 Nov 25 15:25:47 crc kubenswrapper[4890]: I1125 15:25:47.589283 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-56bcr" event={"ID":"a9efeabb-bf1c-4de2-8322-961b284e95cc","Type":"ContainerDied","Data":"09913842764d83958bb7259fb1a01e61d3da9512eb07f23da446059eb49a046a"} Nov 25 15:25:47 crc kubenswrapper[4890]: I1125 15:25:47.589827 4890 scope.go:117] "RemoveContainer" containerID="09913842764d83958bb7259fb1a01e61d3da9512eb07f23da446059eb49a046a" Nov 25 15:25:47 crc kubenswrapper[4890]: E1125 15:25:47.590074 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=octavia-operator-controller-manager-fd75fd47d-56bcr_openstack-operators(a9efeabb-bf1c-4de2-8322-961b284e95cc)\"" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-56bcr" podUID="a9efeabb-bf1c-4de2-8322-961b284e95cc" Nov 25 15:25:47 crc kubenswrapper[4890]: I1125 15:25:47.595463 4890 generic.go:334] "Generic (PLEG): container finished" podID="d50aec9d-1673-4b81-bb57-04cae418843e" containerID="286d09d1c4a9ff164ba4164079699a9e512eed1984749d1fee1c487e3968209c" exitCode=1 Nov 25 15:25:47 crc kubenswrapper[4890]: I1125 15:25:47.595546 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-6zr4x" event={"ID":"d50aec9d-1673-4b81-bb57-04cae418843e","Type":"ContainerDied","Data":"286d09d1c4a9ff164ba4164079699a9e512eed1984749d1fee1c487e3968209c"} Nov 25 15:25:47 crc kubenswrapper[4890]: I1125 15:25:47.596369 4890 scope.go:117] "RemoveContainer" containerID="6de55425fa91945cc6e8f62649df335dd495a944f0f7aed9121c87e80c8f3a96" Nov 25 15:25:47 crc kubenswrapper[4890]: E1125 15:25:47.596661 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=ironic-operator-controller-manager-5bfcdc958c-hxzf4_openstack-operators(679b3bf3-1ebc-4d19-9ad2-588bd41565dd)\"" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-hxzf4" podUID="679b3bf3-1ebc-4d19-9ad2-588bd41565dd" Nov 25 15:25:47 crc kubenswrapper[4890]: I1125 15:25:47.597053 4890 scope.go:117] "RemoveContainer" containerID="0770c2f791722b326d3323baa96eca63ec6eac08347883ea0d8f92a8abb2ed7d" Nov 25 15:25:47 crc kubenswrapper[4890]: I1125 15:25:47.597290 4890 scope.go:117] "RemoveContainer" containerID="38d0123e0cedad944df1c84a66fc4aba91c298dc840c5489a45e6ec2fd23cc66" Nov 25 15:25:47 crc kubenswrapper[4890]: E1125 15:25:47.597320 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=ovn-operator-controller-manager-66cf5c67ff-vx75t_openstack-operators(29f250db-2a96-41ad-bb8f-5b503a9288d8)\"" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-vx75t" podUID="29f250db-2a96-41ad-bb8f-5b503a9288d8" Nov 25 15:25:47 crc kubenswrapper[4890]: I1125 15:25:47.597360 4890 scope.go:117] "RemoveContainer" containerID="286d09d1c4a9ff164ba4164079699a9e512eed1984749d1fee1c487e3968209c" Nov 25 15:25:47 crc kubenswrapper[4890]: E1125 15:25:47.597582 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=keystone-operator-controller-manager-748dc6576f-rs6hx_openstack-operators(a5d8a8ba-50c0-4fb6-823f-f28cf0d0f125)\"" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-rs6hx" podUID="a5d8a8ba-50c0-4fb6-823f-f28cf0d0f125" Nov 25 15:25:47 crc kubenswrapper[4890]: E1125 15:25:47.597693 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=telemetry-operator-controller-manager-567f98c9d-6zr4x_openstack-operators(d50aec9d-1673-4b81-bb57-04cae418843e)\"" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-6zr4x" podUID="d50aec9d-1673-4b81-bb57-04cae418843e" Nov 25 15:25:47 crc kubenswrapper[4890]: I1125 15:25:47.598068 4890 scope.go:117] "RemoveContainer" containerID="e0214f8d055e3e0e07dd810288a0139e23c1d5a5bcfdaa0457141e014e32d69d" Nov 25 15:25:47 crc kubenswrapper[4890]: E1125 15:25:47.598317 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=barbican-operator-controller-manager-86dc4d89c8-pmmf2_openstack-operators(d175a35b-8b9b-49c0-8fc2-02ef62f131d9)\"" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-pmmf2" podUID="d175a35b-8b9b-49c0-8fc2-02ef62f131d9" Nov 25 15:25:47 crc kubenswrapper[4890]: I1125 15:25:47.598620 4890 scope.go:117] "RemoveContainer" containerID="77f771985277fc92c381c09a36784838a56dce95cd158b58ae0dd79c46671845" Nov 25 15:25:47 crc kubenswrapper[4890]: E1125 15:25:47.598861 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=heat-operator-controller-manager-774b86978c-xvfgd_openstack-operators(72c26c5b-3370-4994-a5f0-82128c62c2df)\"" pod="openstack-operators/heat-operator-controller-manager-774b86978c-xvfgd" podUID="72c26c5b-3370-4994-a5f0-82128c62c2df" Nov 25 15:25:47 crc kubenswrapper[4890]: I1125 15:25:47.599653 4890 scope.go:117] "RemoveContainer" containerID="57328d9eeb31abecef2987ab0e98f3bd5485e45dbe190b251c30dc54effefe13" Nov 25 15:25:47 crc kubenswrapper[4890]: E1125 15:25:47.599916 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=glance-operator-controller-manager-68b95954c9-lfnwl_openstack-operators(076c2a0e-7fdb-42e1-ab73-e296bba2b7a9)\"" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-lfnwl" podUID="076c2a0e-7fdb-42e1-ab73-e296bba2b7a9" Nov 25 15:25:47 crc kubenswrapper[4890]: I1125 15:25:47.602426 4890 scope.go:117] "RemoveContainer" containerID="52f8db982762a517d8d96beb141da93c238d379ce39501ef137cb21b7fdcf14d" Nov 25 15:25:47 crc kubenswrapper[4890]: E1125 15:25:47.602793 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=horizon-operator-controller-manager-68c9694994-j6pqs_openstack-operators(c528348f-0907-423a-a0c2-f15cf18ff99a)\"" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-j6pqs" podUID="c528348f-0907-423a-a0c2-f15cf18ff99a" Nov 25 15:25:47 crc kubenswrapper[4890]: I1125 15:25:47.733393 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-stflb" Nov 25 15:25:47 crc kubenswrapper[4890]: I1125 15:25:47.733458 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-stflb" Nov 25 15:25:47 crc kubenswrapper[4890]: I1125 15:25:47.734263 4890 scope.go:117] "RemoveContainer" containerID="01cc934a2ae0af3a25c372a4e138035e62b92af8b615443d315c5b0d7695f5c1" Nov 25 15:25:47 crc kubenswrapper[4890]: E1125 15:25:47.734544 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=infra-operator-controller-manager-d5cc86f4b-stflb_openstack-operators(7e576602-19f5-4656-84dd-5ab80abca895)\"" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-stflb" podUID="7e576602-19f5-4656-84dd-5ab80abca895" Nov 25 15:25:47 crc kubenswrapper[4890]: I1125 15:25:47.802098 4890 scope.go:117] "RemoveContainer" containerID="75820db98fdfa144b9503a1d7e8750f8d454f7258242e932a8932d40b2582263" Nov 25 15:25:48 crc kubenswrapper[4890]: I1125 15:25:48.002263 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-6d88ccc4fc-9smnn" Nov 25 15:25:48 crc kubenswrapper[4890]: I1125 15:25:48.188784 4890 scope.go:117] "RemoveContainer" containerID="dbe71c9a5936928af6baf15ad268bc19748f775a26475ca9db52499631dbeef9" Nov 25 15:25:48 crc kubenswrapper[4890]: I1125 15:25:48.428317 4890 scope.go:117] "RemoveContainer" containerID="1eda77f255a83e5efb319f627a4f7cd146d6e94ac18d2411cf044b47387b7986" Nov 25 15:25:48 crc kubenswrapper[4890]: I1125 15:25:48.608123 4890 generic.go:334] "Generic (PLEG): container finished" podID="03353c8b-91f4-41ae-afc1-c522b1c73bb4" containerID="34aad7779779b8785f00f0293acf127026ceeeb1d0bf6285415e611bc6e3be23" exitCode=1 Nov 25 15:25:48 crc kubenswrapper[4890]: I1125 15:25:48.608244 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-wvjw7" event={"ID":"03353c8b-91f4-41ae-afc1-c522b1c73bb4","Type":"ContainerDied","Data":"34aad7779779b8785f00f0293acf127026ceeeb1d0bf6285415e611bc6e3be23"} Nov 25 15:25:48 crc kubenswrapper[4890]: I1125 15:25:48.609038 4890 scope.go:117] "RemoveContainer" containerID="34aad7779779b8785f00f0293acf127026ceeeb1d0bf6285415e611bc6e3be23" Nov 25 15:25:48 crc kubenswrapper[4890]: E1125 15:25:48.609422 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=mariadb-operator-controller-manager-cb6c4fdb7-wvjw7_openstack-operators(03353c8b-91f4-41ae-afc1-c522b1c73bb4)\"" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-wvjw7" podUID="03353c8b-91f4-41ae-afc1-c522b1c73bb4" Nov 25 15:25:48 crc kubenswrapper[4890]: I1125 15:25:48.612877 4890 scope.go:117] "RemoveContainer" containerID="0770c2f791722b326d3323baa96eca63ec6eac08347883ea0d8f92a8abb2ed7d" Nov 25 15:25:48 crc kubenswrapper[4890]: E1125 15:25:48.613144 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=ovn-operator-controller-manager-66cf5c67ff-vx75t_openstack-operators(29f250db-2a96-41ad-bb8f-5b503a9288d8)\"" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-vx75t" podUID="29f250db-2a96-41ad-bb8f-5b503a9288d8" Nov 25 15:25:48 crc kubenswrapper[4890]: I1125 15:25:48.616401 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6d88ccc4fc-9smnn" event={"ID":"bd60efda-e354-44d0-af56-c110651bdbaf","Type":"ContainerStarted","Data":"1ea348723c401f5343cffcf799c73a110084ee25b07cbf4585aa49024ac031e0"} Nov 25 15:25:48 crc kubenswrapper[4890]: I1125 15:25:48.617293 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-6d88ccc4fc-9smnn" Nov 25 15:25:48 crc kubenswrapper[4890]: I1125 15:25:48.619268 4890 generic.go:334] "Generic (PLEG): container finished" podID="a9a69828-3f0d-4d55-a840-10c2c791444c" containerID="76d8244364a59157c2b9cbffef85f40bfbb1b0424e7c06c66a9fdf3f856a8843" exitCode=1 Nov 25 15:25:48 crc kubenswrapper[4890]: I1125 15:25:48.619325 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-864885998-62ddz" event={"ID":"a9a69828-3f0d-4d55-a840-10c2c791444c","Type":"ContainerDied","Data":"76d8244364a59157c2b9cbffef85f40bfbb1b0424e7c06c66a9fdf3f856a8843"} Nov 25 15:25:48 crc kubenswrapper[4890]: I1125 15:25:48.619753 4890 scope.go:117] "RemoveContainer" containerID="76d8244364a59157c2b9cbffef85f40bfbb1b0424e7c06c66a9fdf3f856a8843" Nov 25 15:25:48 crc kubenswrapper[4890]: I1125 15:25:48.622977 4890 generic.go:334] "Generic (PLEG): container finished" podID="e26594ec-4bc2-4051-8a49-fc490bd3b6c9" containerID="45be6c4aa1f57c239660be552320d18f875cde9c70ec5abb3fa04a6ed3e9139e" exitCode=1 Nov 25 15:25:48 crc kubenswrapper[4890]: I1125 15:25:48.623103 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-dmlx4" event={"ID":"e26594ec-4bc2-4051-8a49-fc490bd3b6c9","Type":"ContainerDied","Data":"45be6c4aa1f57c239660be552320d18f875cde9c70ec5abb3fa04a6ed3e9139e"} Nov 25 15:25:48 crc kubenswrapper[4890]: I1125 15:25:48.623659 4890 scope.go:117] "RemoveContainer" containerID="45be6c4aa1f57c239660be552320d18f875cde9c70ec5abb3fa04a6ed3e9139e" Nov 25 15:25:48 crc kubenswrapper[4890]: I1125 15:25:48.623949 4890 scope.go:117] "RemoveContainer" containerID="6328b6da81615173e561e6807d29b480ebef9c78fbca0d829f62256dd629eac3" Nov 25 15:25:48 crc kubenswrapper[4890]: E1125 15:25:48.623971 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=operator pod=rabbitmq-cluster-operator-manager-668c99d594-dmlx4_openstack-operators(e26594ec-4bc2-4051-8a49-fc490bd3b6c9)\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-dmlx4" podUID="e26594ec-4bc2-4051-8a49-fc490bd3b6c9" Nov 25 15:25:48 crc kubenswrapper[4890]: E1125 15:25:48.624271 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=manila-operator-controller-manager-58bb8d67cc-slpt4_openstack-operators(f272410f-230d-43f2-946e-e34e4d9c05ea)\"" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-slpt4" podUID="f272410f-230d-43f2-946e-e34e4d9c05ea" Nov 25 15:25:48 crc kubenswrapper[4890]: I1125 15:25:48.624662 4890 scope.go:117] "RemoveContainer" containerID="70634ed74a9c04a3a724fdb83aff92615249c19958b69dd169263857756bf093" Nov 25 15:25:48 crc kubenswrapper[4890]: I1125 15:25:48.624690 4890 scope.go:117] "RemoveContainer" containerID="7ff93d6707b3f0b8fb868414ecf039dfc3ca37a71aa91bcc35b9114d425a199f" Nov 25 15:25:48 crc kubenswrapper[4890]: I1125 15:25:48.624745 4890 scope.go:117] "RemoveContainer" containerID="e7b7fcdef297e60c89760ed3c86c2d93e8c7a6a0088d065d1090410732523bb5" Nov 25 15:25:48 crc kubenswrapper[4890]: I1125 15:25:48.624765 4890 scope.go:117] "RemoveContainer" containerID="5bdc10c6af2349a7635d97fa9053a19229cec02d80cae8cde73fc72b252787eb" Nov 25 15:25:48 crc kubenswrapper[4890]: E1125 15:25:48.624884 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=nova-operator-controller-manager-79556f57fc-6jq7q_openstack-operators(0769cde4-b49f-41c5-b720-e0dbc667eb4a)\"" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6jq7q" podUID="0769cde4-b49f-41c5-b720-e0dbc667eb4a" Nov 25 15:25:48 crc kubenswrapper[4890]: E1125 15:25:48.624895 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=swift-operator-controller-manager-6fdc4fcf86-dzkgt_openstack-operators(861317e1-222d-4f14-b931-e9e7d100ebd6)\"" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-dzkgt" podUID="861317e1-222d-4f14-b931-e9e7d100ebd6" Nov 25 15:25:48 crc kubenswrapper[4890]: E1125 15:25:48.624947 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=openstack-operator-controller-manager-68587559f4-9b4cf_openstack-operators(3e496c3d-9485-44bc-b3b1-dc1682a88434)\"" pod="openstack-operators/openstack-operator-controller-manager-68587559f4-9b4cf" podUID="3e496c3d-9485-44bc-b3b1-dc1682a88434" Nov 25 15:25:48 crc kubenswrapper[4890]: I1125 15:25:48.625077 4890 scope.go:117] "RemoveContainer" containerID="286d09d1c4a9ff164ba4164079699a9e512eed1984749d1fee1c487e3968209c" Nov 25 15:25:48 crc kubenswrapper[4890]: E1125 15:25:48.625091 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=designate-operator-controller-manager-7d695c9b56-stwjm_openstack-operators(77a5e8f9-92ab-48b5-be19-29a7a8e5df49)\"" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-stwjm" podUID="77a5e8f9-92ab-48b5-be19-29a7a8e5df49" Nov 25 15:25:48 crc kubenswrapper[4890]: I1125 15:25:48.625338 4890 scope.go:117] "RemoveContainer" containerID="09913842764d83958bb7259fb1a01e61d3da9512eb07f23da446059eb49a046a" Nov 25 15:25:48 crc kubenswrapper[4890]: E1125 15:25:48.625398 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=telemetry-operator-controller-manager-567f98c9d-6zr4x_openstack-operators(d50aec9d-1673-4b81-bb57-04cae418843e)\"" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-6zr4x" podUID="d50aec9d-1673-4b81-bb57-04cae418843e" Nov 25 15:25:48 crc kubenswrapper[4890]: I1125 15:25:48.625441 4890 scope.go:117] "RemoveContainer" containerID="60ab9e903ab7dd461c9fab060610f14563535916392274aabb88f8d7b90adbde" Nov 25 15:25:48 crc kubenswrapper[4890]: E1125 15:25:48.625626 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=placement-operator-controller-manager-5db546f9d9-nshw7_openstack-operators(9f0e3cf9-dcc0-405f-a2de-9148844ec3ae)\"" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-nshw7" podUID="9f0e3cf9-dcc0-405f-a2de-9148844ec3ae" Nov 25 15:25:48 crc kubenswrapper[4890]: E1125 15:25:48.625668 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=octavia-operator-controller-manager-fd75fd47d-56bcr_openstack-operators(a9efeabb-bf1c-4de2-8322-961b284e95cc)\"" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-56bcr" podUID="a9efeabb-bf1c-4de2-8322-961b284e95cc" Nov 25 15:25:48 crc kubenswrapper[4890]: I1125 15:25:48.732986 4890 scope.go:117] "RemoveContainer" containerID="aa72b2ab7e7b9beabe0d4dc521c692a9eb7688f8f0273d14f7a29938b97646c9" Nov 25 15:25:49 crc kubenswrapper[4890]: I1125 15:25:49.083969 4890 scope.go:117] "RemoveContainer" containerID="8756ce10612e1443f132d473e033dbcdfe66560ddde56b5f30ba360b8c363acb" Nov 25 15:25:49 crc kubenswrapper[4890]: I1125 15:25:49.384389 4890 scope.go:117] "RemoveContainer" containerID="cc0132b2f896b99524175db61d9c1524df2e54f587019f94d706da2cdea210cc" Nov 25 15:25:49 crc kubenswrapper[4890]: I1125 15:25:49.569450 4890 scope.go:117] "RemoveContainer" containerID="d9880e46a52dd26c73643cdd5427eeee12138ceffa656e9d825b0e88a70d2335" Nov 25 15:25:49 crc kubenswrapper[4890]: I1125 15:25:49.642675 4890 scope.go:117] "RemoveContainer" containerID="45be6c4aa1f57c239660be552320d18f875cde9c70ec5abb3fa04a6ed3e9139e" Nov 25 15:25:49 crc kubenswrapper[4890]: E1125 15:25:49.642910 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=operator pod=rabbitmq-cluster-operator-manager-668c99d594-dmlx4_openstack-operators(e26594ec-4bc2-4051-8a49-fc490bd3b6c9)\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-dmlx4" podUID="e26594ec-4bc2-4051-8a49-fc490bd3b6c9" Nov 25 15:25:49 crc kubenswrapper[4890]: I1125 15:25:49.643184 4890 scope.go:117] "RemoveContainer" containerID="70634ed74a9c04a3a724fdb83aff92615249c19958b69dd169263857756bf093" Nov 25 15:25:49 crc kubenswrapper[4890]: E1125 15:25:49.643427 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=nova-operator-controller-manager-79556f57fc-6jq7q_openstack-operators(0769cde4-b49f-41c5-b720-e0dbc667eb4a)\"" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6jq7q" podUID="0769cde4-b49f-41c5-b720-e0dbc667eb4a" Nov 25 15:25:49 crc kubenswrapper[4890]: I1125 15:25:49.643625 4890 scope.go:117] "RemoveContainer" containerID="34aad7779779b8785f00f0293acf127026ceeeb1d0bf6285415e611bc6e3be23" Nov 25 15:25:49 crc kubenswrapper[4890]: E1125 15:25:49.643996 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=mariadb-operator-controller-manager-cb6c4fdb7-wvjw7_openstack-operators(03353c8b-91f4-41ae-afc1-c522b1c73bb4)\"" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-wvjw7" podUID="03353c8b-91f4-41ae-afc1-c522b1c73bb4" Nov 25 15:25:49 crc kubenswrapper[4890]: I1125 15:25:49.989414 4890 scope.go:117] "RemoveContainer" containerID="25d97eb8b0662c62e96b5e278d3fba5c2fb55235f072e6116280714f791a4c15" Nov 25 15:25:49 crc kubenswrapper[4890]: I1125 15:25:49.993250 4890 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="d08ea973-c3f2-4192-8bfc-b025f0c5391b" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.210:5671: connect: connection refused" Nov 25 15:25:50 crc kubenswrapper[4890]: I1125 15:25:50.022513 4890 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="9bab37b3-681b-4cd5-9fee-503afa0d6d22" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.211:5671: connect: connection refused" Nov 25 15:25:50 crc kubenswrapper[4890]: I1125 15:25:50.269416 4890 scope.go:117] "RemoveContainer" containerID="97e02800cd95b6b08806006fece844f452d79638de429148891a1ede18c31c8f" Nov 25 15:25:50 crc kubenswrapper[4890]: I1125 15:25:50.656028 4890 generic.go:334] "Generic (PLEG): container finished" podID="a9a69828-3f0d-4d55-a840-10c2c791444c" containerID="4dbafe9e16c939461a203c2af1cab51ba3a18f270889d1e522c1d51306fb473d" exitCode=1 Nov 25 15:25:50 crc kubenswrapper[4890]: I1125 15:25:50.656092 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-864885998-62ddz" event={"ID":"a9a69828-3f0d-4d55-a840-10c2c791444c","Type":"ContainerDied","Data":"4dbafe9e16c939461a203c2af1cab51ba3a18f270889d1e522c1d51306fb473d"} Nov 25 15:25:50 crc kubenswrapper[4890]: I1125 15:25:50.656837 4890 scope.go:117] "RemoveContainer" containerID="4dbafe9e16c939461a203c2af1cab51ba3a18f270889d1e522c1d51306fb473d" Nov 25 15:25:50 crc kubenswrapper[4890]: E1125 15:25:50.657089 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=watcher-operator-controller-manager-864885998-62ddz_openstack-operators(a9a69828-3f0d-4d55-a840-10c2c791444c)\"" pod="openstack-operators/watcher-operator-controller-manager-864885998-62ddz" podUID="a9a69828-3f0d-4d55-a840-10c2c791444c" Nov 25 15:25:50 crc kubenswrapper[4890]: I1125 15:25:50.661455 4890 generic.go:334] "Generic (PLEG): container finished" podID="a47203e9-a88a-42f7-86e1-164c6ad532b5" containerID="2e409da14a5c1cd59defa63e3a546ed3a94ed44d1a0de5a7202afe0f4832568b" exitCode=1 Nov 25 15:25:50 crc kubenswrapper[4890]: I1125 15:25:50.661521 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"a47203e9-a88a-42f7-86e1-164c6ad532b5","Type":"ContainerDied","Data":"2e409da14a5c1cd59defa63e3a546ed3a94ed44d1a0de5a7202afe0f4832568b"} Nov 25 15:25:50 crc kubenswrapper[4890]: I1125 15:25:50.662527 4890 scope.go:117] "RemoveContainer" containerID="2e409da14a5c1cd59defa63e3a546ed3a94ed44d1a0de5a7202afe0f4832568b" Nov 25 15:25:50 crc kubenswrapper[4890]: I1125 15:25:50.663143 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-68587559f4-9b4cf" Nov 25 15:25:50 crc kubenswrapper[4890]: I1125 15:25:50.663253 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/openstack-operator-controller-manager-68587559f4-9b4cf" Nov 25 15:25:50 crc kubenswrapper[4890]: I1125 15:25:50.663714 4890 scope.go:117] "RemoveContainer" containerID="5bdc10c6af2349a7635d97fa9053a19229cec02d80cae8cde73fc72b252787eb" Nov 25 15:25:50 crc kubenswrapper[4890]: E1125 15:25:50.663966 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=openstack-operator-controller-manager-68587559f4-9b4cf_openstack-operators(3e496c3d-9485-44bc-b3b1-dc1682a88434)\"" pod="openstack-operators/openstack-operator-controller-manager-68587559f4-9b4cf" podUID="3e496c3d-9485-44bc-b3b1-dc1682a88434" Nov 25 15:25:50 crc kubenswrapper[4890]: I1125 15:25:50.709036 4890 scope.go:117] "RemoveContainer" containerID="b8f781558438b476e77ca8c2495fc65050de7477ff54cb3185dc7dd498e07866" Nov 25 15:25:50 crc kubenswrapper[4890]: I1125 15:25:50.955938 4890 scope.go:117] "RemoveContainer" containerID="360f2cd67207fde2598cf2336d7ee654c1e558557dae8116644fbbad97f92bbe" Nov 25 15:25:51 crc kubenswrapper[4890]: I1125 15:25:51.136593 4890 scope.go:117] "RemoveContainer" containerID="b281ef61e39f8d5a43f4b56119cbfae16eb18c923bfe412c39c180018be2370e" Nov 25 15:25:51 crc kubenswrapper[4890]: I1125 15:25:51.570480 4890 scope.go:117] "RemoveContainer" containerID="820303aa9ed50019579596b6edb70067e3cf2b1af539efec84d656ec8ffdf5f5" Nov 25 15:25:51 crc kubenswrapper[4890]: I1125 15:25:51.778789 4890 scope.go:117] "RemoveContainer" containerID="61b9be5d90ce8a666acdd758f569aaa247bc6bf33719297ce9c65e886c773f90" Nov 25 15:25:52 crc kubenswrapper[4890]: I1125 15:25:52.481361 4890 scope.go:117] "RemoveContainer" containerID="76d8244364a59157c2b9cbffef85f40bfbb1b0424e7c06c66a9fdf3f856a8843" Nov 25 15:25:53 crc kubenswrapper[4890]: I1125 15:25:53.173229 4890 scope.go:117] "RemoveContainer" containerID="be30ac4807bb1b8659327603f58cc562ca5c318c9246dbbc3a89d446e0498389" Nov 25 15:25:53 crc kubenswrapper[4890]: I1125 15:25:53.811778 4890 scope.go:117] "RemoveContainer" containerID="f6c757f7b337424920e92dc7103283ff1d00267dabeb93ed84f34ff593609e9a" Nov 25 15:25:53 crc kubenswrapper[4890]: I1125 15:25:53.856015 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 25 15:25:54 crc kubenswrapper[4890]: I1125 15:25:54.154682 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 25 15:25:54 crc kubenswrapper[4890]: I1125 15:25:54.280849 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 25 15:25:55 crc kubenswrapper[4890]: I1125 15:25:55.195441 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-snhn6" Nov 25 15:25:55 crc kubenswrapper[4890]: I1125 15:25:55.314062 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 25 15:25:55 crc kubenswrapper[4890]: I1125 15:25:55.328447 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 25 15:25:55 crc kubenswrapper[4890]: I1125 15:25:55.431998 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 25 15:25:55 crc kubenswrapper[4890]: I1125 15:25:55.472812 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 25 15:25:55 crc kubenswrapper[4890]: I1125 15:25:55.571295 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 25 15:25:55 crc kubenswrapper[4890]: I1125 15:25:55.604775 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-2dmvv" Nov 25 15:25:55 crc kubenswrapper[4890]: I1125 15:25:55.615534 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Nov 25 15:25:55 crc kubenswrapper[4890]: I1125 15:25:55.738529 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Nov 25 15:25:55 crc kubenswrapper[4890]: I1125 15:25:55.738992 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 25 15:25:55 crc kubenswrapper[4890]: I1125 15:25:55.739036 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/kube-state-metrics-0" Nov 25 15:25:55 crc kubenswrapper[4890]: I1125 15:25:55.859446 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 25 15:25:55 crc kubenswrapper[4890]: I1125 15:25:55.927804 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 25 15:25:56 crc kubenswrapper[4890]: I1125 15:25:56.076787 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 25 15:25:56 crc kubenswrapper[4890]: I1125 15:25:56.096127 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-527z8" Nov 25 15:25:56 crc kubenswrapper[4890]: I1125 15:25:56.111045 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-pmmf2" Nov 25 15:25:56 crc kubenswrapper[4890]: I1125 15:25:56.111708 4890 scope.go:117] "RemoveContainer" containerID="e0214f8d055e3e0e07dd810288a0139e23c1d5a5bcfdaa0457141e014e32d69d" Nov 25 15:25:56 crc kubenswrapper[4890]: I1125 15:25:56.116834 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-jcw7x" Nov 25 15:25:56 crc kubenswrapper[4890]: I1125 15:25:56.117915 4890 scope.go:117] "RemoveContainer" containerID="cce83adafd4303a3c4644332384918e2da8852bc205228c4dadd7e6a87ff1cf8" Nov 25 15:25:56 crc kubenswrapper[4890]: I1125 15:25:56.128626 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-stwjm" Nov 25 15:25:56 crc kubenswrapper[4890]: I1125 15:25:56.129323 4890 scope.go:117] "RemoveContainer" containerID="e7b7fcdef297e60c89760ed3c86c2d93e8c7a6a0088d065d1090410732523bb5" Nov 25 15:25:56 crc kubenswrapper[4890]: E1125 15:25:56.129694 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=designate-operator-controller-manager-7d695c9b56-stwjm_openstack-operators(77a5e8f9-92ab-48b5-be19-29a7a8e5df49)\"" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-stwjm" podUID="77a5e8f9-92ab-48b5-be19-29a7a8e5df49" Nov 25 15:25:56 crc kubenswrapper[4890]: I1125 15:25:56.192662 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Nov 25 15:25:56 crc kubenswrapper[4890]: I1125 15:25:56.192730 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 25 15:25:56 crc kubenswrapper[4890]: I1125 15:25:56.195271 4890 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-k7z2s" Nov 25 15:25:56 crc kubenswrapper[4890]: I1125 15:25:56.208072 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-774b86978c-xvfgd" Nov 25 15:25:56 crc kubenswrapper[4890]: I1125 15:25:56.208948 4890 scope.go:117] "RemoveContainer" containerID="77f771985277fc92c381c09a36784838a56dce95cd158b58ae0dd79c46671845" Nov 25 15:25:56 crc kubenswrapper[4890]: I1125 15:25:56.223190 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-j6pqs" Nov 25 15:25:56 crc kubenswrapper[4890]: I1125 15:25:56.223883 4890 scope.go:117] "RemoveContainer" containerID="52f8db982762a517d8d96beb141da93c238d379ce39501ef137cb21b7fdcf14d" Nov 25 15:25:56 crc kubenswrapper[4890]: I1125 15:25:56.275487 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Nov 25 15:25:56 crc kubenswrapper[4890]: I1125 15:25:56.302251 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-hxzf4" Nov 25 15:25:56 crc kubenswrapper[4890]: I1125 15:25:56.303001 4890 scope.go:117] "RemoveContainer" containerID="6de55425fa91945cc6e8f62649df335dd495a944f0f7aed9121c87e80c8f3a96" Nov 25 15:25:56 crc kubenswrapper[4890]: I1125 15:25:56.316011 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-rs6hx" Nov 25 15:25:56 crc kubenswrapper[4890]: I1125 15:25:56.316976 4890 scope.go:117] "RemoveContainer" containerID="38d0123e0cedad944df1c84a66fc4aba91c298dc840c5489a45e6ec2fd23cc66" Nov 25 15:25:56 crc kubenswrapper[4890]: I1125 15:25:56.538251 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 25 15:25:56 crc kubenswrapper[4890]: I1125 15:25:56.567281 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-6zr4x" Nov 25 15:25:56 crc kubenswrapper[4890]: I1125 15:25:56.568441 4890 scope.go:117] "RemoveContainer" containerID="286d09d1c4a9ff164ba4164079699a9e512eed1984749d1fee1c487e3968209c" Nov 25 15:25:56 crc kubenswrapper[4890]: I1125 15:25:56.622440 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6jq7q" Nov 25 15:25:56 crc kubenswrapper[4890]: I1125 15:25:56.623373 4890 scope.go:117] "RemoveContainer" containerID="70634ed74a9c04a3a724fdb83aff92615249c19958b69dd169263857756bf093" Nov 25 15:25:56 crc kubenswrapper[4890]: I1125 15:25:56.649607 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-56bcr" Nov 25 15:25:56 crc kubenswrapper[4890]: I1125 15:25:56.653221 4890 scope.go:117] "RemoveContainer" containerID="09913842764d83958bb7259fb1a01e61d3da9512eb07f23da446059eb49a046a" Nov 25 15:25:56 crc kubenswrapper[4890]: E1125 15:25:56.654389 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=octavia-operator-controller-manager-fd75fd47d-56bcr_openstack-operators(a9efeabb-bf1c-4de2-8322-961b284e95cc)\"" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-56bcr" podUID="a9efeabb-bf1c-4de2-8322-961b284e95cc" Nov 25 15:25:56 crc kubenswrapper[4890]: I1125 15:25:56.775505 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-dzkgt" Nov 25 15:25:56 crc kubenswrapper[4890]: I1125 15:25:56.776467 4890 scope.go:117] "RemoveContainer" containerID="7ff93d6707b3f0b8fb868414ecf039dfc3ca37a71aa91bcc35b9114d425a199f" Nov 25 15:25:56 crc kubenswrapper[4890]: I1125 15:25:56.858446 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 25 15:25:56 crc kubenswrapper[4890]: I1125 15:25:56.938132 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 25 15:25:56 crc kubenswrapper[4890]: I1125 15:25:56.955244 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5cb74df96-6mqt2" Nov 25 15:25:56 crc kubenswrapper[4890]: I1125 15:25:56.982205 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-864885998-62ddz" Nov 25 15:25:56 crc kubenswrapper[4890]: I1125 15:25:56.982705 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 25 15:25:56 crc kubenswrapper[4890]: I1125 15:25:56.982960 4890 scope.go:117] "RemoveContainer" containerID="4dbafe9e16c939461a203c2af1cab51ba3a18f270889d1e522c1d51306fb473d" Nov 25 15:25:56 crc kubenswrapper[4890]: E1125 15:25:56.983258 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=watcher-operator-controller-manager-864885998-62ddz_openstack-operators(a9a69828-3f0d-4d55-a840-10c2c791444c)\"" pod="openstack-operators/watcher-operator-controller-manager-864885998-62ddz" podUID="a9a69828-3f0d-4d55-a840-10c2c791444c" Nov 25 15:25:57 crc kubenswrapper[4890]: I1125 15:25:57.006311 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 25 15:25:57 crc kubenswrapper[4890]: I1125 15:25:57.054745 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 25 15:25:57 crc kubenswrapper[4890]: I1125 15:25:57.107074 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 25 15:25:57 crc kubenswrapper[4890]: I1125 15:25:57.127252 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-nhckl" Nov 25 15:25:57 crc kubenswrapper[4890]: I1125 15:25:57.137286 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 25 15:25:57 crc kubenswrapper[4890]: I1125 15:25:57.159007 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 25 15:25:57 crc kubenswrapper[4890]: I1125 15:25:57.179725 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 25 15:25:57 crc kubenswrapper[4890]: I1125 15:25:57.222714 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 25 15:25:57 crc kubenswrapper[4890]: I1125 15:25:57.251329 4890 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Nov 25 15:25:57 crc kubenswrapper[4890]: I1125 15:25:57.251380 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Nov 25 15:25:57 crc kubenswrapper[4890]: I1125 15:25:57.287210 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 25 15:25:57 crc kubenswrapper[4890]: I1125 15:25:57.291632 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 25 15:25:57 crc kubenswrapper[4890]: I1125 15:25:57.337314 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Nov 25 15:25:57 crc kubenswrapper[4890]: I1125 15:25:57.383125 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 25 15:25:57 crc kubenswrapper[4890]: I1125 15:25:57.436375 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-ctl55" Nov 25 15:25:57 crc kubenswrapper[4890]: I1125 15:25:57.576808 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 25 15:25:57 crc kubenswrapper[4890]: I1125 15:25:57.634051 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 25 15:25:57 crc kubenswrapper[4890]: E1125 15:25:57.770920 4890 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd175a35b_8b9b_49c0_8fc2_02ef62f131d9.slice/crio-24f5582ca11aa9c0ac5997208248ab279725e93ef77197c94bd72f340f8bbab7.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod72c26c5b_3370_4994_a5f0_82128c62c2df.slice/crio-5085d0d9569ced239129365e8bcff54f0f41fba51b79ac5ddca34271333273d9.scope\": RecentStats: unable to find data in memory cache]" Nov 25 15:25:57 crc kubenswrapper[4890]: I1125 15:25:57.770989 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 25 15:25:57 crc kubenswrapper[4890]: I1125 15:25:57.776153 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-g5vrc" Nov 25 15:25:57 crc kubenswrapper[4890]: I1125 15:25:57.776915 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-pmmf2" event={"ID":"d175a35b-8b9b-49c0-8fc2-02ef62f131d9","Type":"ContainerStarted","Data":"24f5582ca11aa9c0ac5997208248ab279725e93ef77197c94bd72f340f8bbab7"} Nov 25 15:25:57 crc kubenswrapper[4890]: I1125 15:25:57.779821 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-hxzf4" event={"ID":"679b3bf3-1ebc-4d19-9ad2-588bd41565dd","Type":"ContainerStarted","Data":"7eed3f7f71023e9126d0de61d864c2912c3fd5e6e53615d8e82e9bb5d98d5081"} Nov 25 15:25:57 crc kubenswrapper[4890]: I1125 15:25:57.782562 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-6zr4x" event={"ID":"d50aec9d-1673-4b81-bb57-04cae418843e","Type":"ContainerStarted","Data":"2545dda572d458934e753cbbe89f8cbd4188e0160a44a1d5caf399b84c9fb657"} Nov 25 15:25:57 crc kubenswrapper[4890]: I1125 15:25:57.786650 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-t8zql" Nov 25 15:25:57 crc kubenswrapper[4890]: I1125 15:25:57.786949 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 25 15:25:57 crc kubenswrapper[4890]: I1125 15:25:57.788437 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-jcw7x" event={"ID":"0bdbada3-e08c-4f8a-bd55-bac955d4370d","Type":"ContainerStarted","Data":"ea03217bf38df8f664326633d6bc9735c9f7d24feebcf0fb7ac30106bfefa2d0"} Nov 25 15:25:57 crc kubenswrapper[4890]: I1125 15:25:57.792427 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6jq7q" event={"ID":"0769cde4-b49f-41c5-b720-e0dbc667eb4a","Type":"ContainerStarted","Data":"afdecad3ee6dca827e01ecfd5110e60bbd29e083f201918f083e2ccc5e1ba964"} Nov 25 15:25:57 crc kubenswrapper[4890]: I1125 15:25:57.794708 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-rs6hx" event={"ID":"a5d8a8ba-50c0-4fb6-823f-f28cf0d0f125","Type":"ContainerStarted","Data":"59bdc927190ff946a016c3b38ef86d4b991001a88772e7a4c7e5cc7cbf93b8d3"} Nov 25 15:25:57 crc kubenswrapper[4890]: I1125 15:25:57.798202 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-j6pqs" event={"ID":"c528348f-0907-423a-a0c2-f15cf18ff99a","Type":"ContainerStarted","Data":"dc8aff712070f60aaf997f341250476ab00d06875bc86901bf5c918d9c5a94d9"} Nov 25 15:25:57 crc kubenswrapper[4890]: I1125 15:25:57.798565 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-j6pqs" Nov 25 15:25:57 crc kubenswrapper[4890]: I1125 15:25:57.802590 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-dzkgt" event={"ID":"861317e1-222d-4f14-b931-e9e7d100ebd6","Type":"ContainerStarted","Data":"aabb0ea70edd64ac6611bc788292aa0806bdccd21d77f87d39aa68b59663cf81"} Nov 25 15:25:57 crc kubenswrapper[4890]: I1125 15:25:57.806417 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-774b86978c-xvfgd" event={"ID":"72c26c5b-3370-4994-a5f0-82128c62c2df","Type":"ContainerStarted","Data":"5085d0d9569ced239129365e8bcff54f0f41fba51b79ac5ddca34271333273d9"} Nov 25 15:25:57 crc kubenswrapper[4890]: I1125 15:25:57.810210 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"a47203e9-a88a-42f7-86e1-164c6ad532b5","Type":"ContainerStarted","Data":"57618fb9802f27a586bce5ed3b16f89465e88061eac4872f5ae34829c7d008e2"} Nov 25 15:25:57 crc kubenswrapper[4890]: I1125 15:25:57.815220 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-66fc8c7fbf-lnjxm" event={"ID":"dad943fe-2209-40f9-a9bb-0be7ba80cca2","Type":"ContainerStarted","Data":"137b37bbfac0ae56e61f31ca811e8cf420a7182f7941ed967187fc1ed8ac86f0"} Nov 25 15:25:57 crc kubenswrapper[4890]: I1125 15:25:57.830397 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Nov 25 15:25:57 crc kubenswrapper[4890]: I1125 15:25:57.835426 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Nov 25 15:25:57 crc kubenswrapper[4890]: I1125 15:25:57.859391 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Nov 25 15:25:57 crc kubenswrapper[4890]: I1125 15:25:57.917020 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-fdhzs" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.005122 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-6d88ccc4fc-9smnn" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.007702 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.018837 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.056401 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.058072 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.072822 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.096016 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.110843 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-ncrf8" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.168987 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.173023 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-5lr7r" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.272148 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.314144 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-bgppv" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.316047 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.334883 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.399603 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.400550 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.402971 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.403428 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-44bfh" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.486342 4890 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.487508 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.516149 4890 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-n9lgx" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.543553 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.552383 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-xznbp" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.552868 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-wtxp4" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.555328 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.569091 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-2j29m" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.598335 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.608997 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.619480 4890 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-hm6p5" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.664694 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.669751 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-xtbnh" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.693782 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.738700 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.762060 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.790843 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.798386 4890 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.827259 4890 generic.go:334] "Generic (PLEG): container finished" podID="dad943fe-2209-40f9-a9bb-0be7ba80cca2" containerID="137b37bbfac0ae56e61f31ca811e8cf420a7182f7941ed967187fc1ed8ac86f0" exitCode=1 Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.827353 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-66fc8c7fbf-lnjxm" event={"ID":"dad943fe-2209-40f9-a9bb-0be7ba80cca2","Type":"ContainerDied","Data":"137b37bbfac0ae56e61f31ca811e8cf420a7182f7941ed967187fc1ed8ac86f0"} Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.827416 4890 scope.go:117] "RemoveContainer" containerID="be30ac4807bb1b8659327603f58cc562ca5c318c9246dbbc3a89d446e0498389" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.828269 4890 scope.go:117] "RemoveContainer" containerID="137b37bbfac0ae56e61f31ca811e8cf420a7182f7941ed967187fc1ed8ac86f0" Nov 25 15:25:58 crc kubenswrapper[4890]: E1125 15:25:58.828759 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=metallb-operator-controller-manager-66fc8c7fbf-lnjxm_metallb-system(dad943fe-2209-40f9-a9bb-0be7ba80cca2)\"" pod="metallb-system/metallb-operator-controller-manager-66fc8c7fbf-lnjxm" podUID="dad943fe-2209-40f9-a9bb-0be7ba80cca2" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.830782 4890 generic.go:334] "Generic (PLEG): container finished" podID="d175a35b-8b9b-49c0-8fc2-02ef62f131d9" containerID="24f5582ca11aa9c0ac5997208248ab279725e93ef77197c94bd72f340f8bbab7" exitCode=1 Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.830828 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-pmmf2" event={"ID":"d175a35b-8b9b-49c0-8fc2-02ef62f131d9","Type":"ContainerDied","Data":"24f5582ca11aa9c0ac5997208248ab279725e93ef77197c94bd72f340f8bbab7"} Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.833912 4890 generic.go:334] "Generic (PLEG): container finished" podID="679b3bf3-1ebc-4d19-9ad2-588bd41565dd" containerID="7eed3f7f71023e9126d0de61d864c2912c3fd5e6e53615d8e82e9bb5d98d5081" exitCode=1 Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.833999 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-hxzf4" event={"ID":"679b3bf3-1ebc-4d19-9ad2-588bd41565dd","Type":"ContainerDied","Data":"7eed3f7f71023e9126d0de61d864c2912c3fd5e6e53615d8e82e9bb5d98d5081"} Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.834535 4890 scope.go:117] "RemoveContainer" containerID="7eed3f7f71023e9126d0de61d864c2912c3fd5e6e53615d8e82e9bb5d98d5081" Nov 25 15:25:58 crc kubenswrapper[4890]: E1125 15:25:58.834792 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=ironic-operator-controller-manager-5bfcdc958c-hxzf4_openstack-operators(679b3bf3-1ebc-4d19-9ad2-588bd41565dd)\"" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-hxzf4" podUID="679b3bf3-1ebc-4d19-9ad2-588bd41565dd" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.837025 4890 scope.go:117] "RemoveContainer" containerID="24f5582ca11aa9c0ac5997208248ab279725e93ef77197c94bd72f340f8bbab7" Nov 25 15:25:58 crc kubenswrapper[4890]: E1125 15:25:58.837414 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=barbican-operator-controller-manager-86dc4d89c8-pmmf2_openstack-operators(d175a35b-8b9b-49c0-8fc2-02ef62f131d9)\"" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-pmmf2" podUID="d175a35b-8b9b-49c0-8fc2-02ef62f131d9" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.839213 4890 generic.go:334] "Generic (PLEG): container finished" podID="72c26c5b-3370-4994-a5f0-82128c62c2df" containerID="5085d0d9569ced239129365e8bcff54f0f41fba51b79ac5ddca34271333273d9" exitCode=1 Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.839318 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-774b86978c-xvfgd" event={"ID":"72c26c5b-3370-4994-a5f0-82128c62c2df","Type":"ContainerDied","Data":"5085d0d9569ced239129365e8bcff54f0f41fba51b79ac5ddca34271333273d9"} Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.840445 4890 scope.go:117] "RemoveContainer" containerID="5085d0d9569ced239129365e8bcff54f0f41fba51b79ac5ddca34271333273d9" Nov 25 15:25:58 crc kubenswrapper[4890]: E1125 15:25:58.840853 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=heat-operator-controller-manager-774b86978c-xvfgd_openstack-operators(72c26c5b-3370-4994-a5f0-82128c62c2df)\"" pod="openstack-operators/heat-operator-controller-manager-774b86978c-xvfgd" podUID="72c26c5b-3370-4994-a5f0-82128c62c2df" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.843353 4890 generic.go:334] "Generic (PLEG): container finished" podID="d50aec9d-1673-4b81-bb57-04cae418843e" containerID="2545dda572d458934e753cbbe89f8cbd4188e0160a44a1d5caf399b84c9fb657" exitCode=1 Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.843433 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-6zr4x" event={"ID":"d50aec9d-1673-4b81-bb57-04cae418843e","Type":"ContainerDied","Data":"2545dda572d458934e753cbbe89f8cbd4188e0160a44a1d5caf399b84c9fb657"} Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.844105 4890 scope.go:117] "RemoveContainer" containerID="2545dda572d458934e753cbbe89f8cbd4188e0160a44a1d5caf399b84c9fb657" Nov 25 15:25:58 crc kubenswrapper[4890]: E1125 15:25:58.844500 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=telemetry-operator-controller-manager-567f98c9d-6zr4x_openstack-operators(d50aec9d-1673-4b81-bb57-04cae418843e)\"" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-6zr4x" podUID="d50aec9d-1673-4b81-bb57-04cae418843e" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.855058 4890 generic.go:334] "Generic (PLEG): container finished" podID="a47203e9-a88a-42f7-86e1-164c6ad532b5" containerID="57618fb9802f27a586bce5ed3b16f89465e88061eac4872f5ae34829c7d008e2" exitCode=1 Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.855698 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"a47203e9-a88a-42f7-86e1-164c6ad532b5","Type":"ContainerDied","Data":"57618fb9802f27a586bce5ed3b16f89465e88061eac4872f5ae34829c7d008e2"} Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.855894 4890 scope.go:117] "RemoveContainer" containerID="57618fb9802f27a586bce5ed3b16f89465e88061eac4872f5ae34829c7d008e2" Nov 25 15:25:58 crc kubenswrapper[4890]: E1125 15:25:58.856205 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-state-metrics pod=kube-state-metrics-0_openstack(a47203e9-a88a-42f7-86e1-164c6ad532b5)\"" pod="openstack/kube-state-metrics-0" podUID="a47203e9-a88a-42f7-86e1-164c6ad532b5" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.867542 4890 generic.go:334] "Generic (PLEG): container finished" podID="0bdbada3-e08c-4f8a-bd55-bac955d4370d" containerID="ea03217bf38df8f664326633d6bc9735c9f7d24feebcf0fb7ac30106bfefa2d0" exitCode=1 Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.867614 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-jcw7x" event={"ID":"0bdbada3-e08c-4f8a-bd55-bac955d4370d","Type":"ContainerDied","Data":"ea03217bf38df8f664326633d6bc9735c9f7d24feebcf0fb7ac30106bfefa2d0"} Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.868324 4890 scope.go:117] "RemoveContainer" containerID="ea03217bf38df8f664326633d6bc9735c9f7d24feebcf0fb7ac30106bfefa2d0" Nov 25 15:25:58 crc kubenswrapper[4890]: E1125 15:25:58.868614 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=cinder-operator-controller-manager-79856dc55c-jcw7x_openstack-operators(0bdbada3-e08c-4f8a-bd55-bac955d4370d)\"" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-jcw7x" podUID="0bdbada3-e08c-4f8a-bd55-bac955d4370d" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.873471 4890 generic.go:334] "Generic (PLEG): container finished" podID="c528348f-0907-423a-a0c2-f15cf18ff99a" containerID="dc8aff712070f60aaf997f341250476ab00d06875bc86901bf5c918d9c5a94d9" exitCode=1 Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.873548 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-j6pqs" event={"ID":"c528348f-0907-423a-a0c2-f15cf18ff99a","Type":"ContainerDied","Data":"dc8aff712070f60aaf997f341250476ab00d06875bc86901bf5c918d9c5a94d9"} Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.874670 4890 scope.go:117] "RemoveContainer" containerID="dc8aff712070f60aaf997f341250476ab00d06875bc86901bf5c918d9c5a94d9" Nov 25 15:25:58 crc kubenswrapper[4890]: E1125 15:25:58.875031 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=horizon-operator-controller-manager-68c9694994-j6pqs_openstack-operators(c528348f-0907-423a-a0c2-f15cf18ff99a)\"" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-j6pqs" podUID="c528348f-0907-423a-a0c2-f15cf18ff99a" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.880996 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.881384 4890 generic.go:334] "Generic (PLEG): container finished" podID="861317e1-222d-4f14-b931-e9e7d100ebd6" containerID="aabb0ea70edd64ac6611bc788292aa0806bdccd21d77f87d39aa68b59663cf81" exitCode=1 Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.881529 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-dzkgt" event={"ID":"861317e1-222d-4f14-b931-e9e7d100ebd6","Type":"ContainerDied","Data":"aabb0ea70edd64ac6611bc788292aa0806bdccd21d77f87d39aa68b59663cf81"} Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.882004 4890 scope.go:117] "RemoveContainer" containerID="aabb0ea70edd64ac6611bc788292aa0806bdccd21d77f87d39aa68b59663cf81" Nov 25 15:25:58 crc kubenswrapper[4890]: E1125 15:25:58.882264 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=swift-operator-controller-manager-6fdc4fcf86-dzkgt_openstack-operators(861317e1-222d-4f14-b931-e9e7d100ebd6)\"" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-dzkgt" podUID="861317e1-222d-4f14-b931-e9e7d100ebd6" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.890117 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.891934 4890 generic.go:334] "Generic (PLEG): container finished" podID="0769cde4-b49f-41c5-b720-e0dbc667eb4a" containerID="afdecad3ee6dca827e01ecfd5110e60bbd29e083f201918f083e2ccc5e1ba964" exitCode=1 Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.892091 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6jq7q" event={"ID":"0769cde4-b49f-41c5-b720-e0dbc667eb4a","Type":"ContainerDied","Data":"afdecad3ee6dca827e01ecfd5110e60bbd29e083f201918f083e2ccc5e1ba964"} Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.892705 4890 scope.go:117] "RemoveContainer" containerID="afdecad3ee6dca827e01ecfd5110e60bbd29e083f201918f083e2ccc5e1ba964" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.892973 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 25 15:25:58 crc kubenswrapper[4890]: E1125 15:25:58.892983 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=nova-operator-controller-manager-79556f57fc-6jq7q_openstack-operators(0769cde4-b49f-41c5-b720-e0dbc667eb4a)\"" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6jq7q" podUID="0769cde4-b49f-41c5-b720-e0dbc667eb4a" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.898973 4890 generic.go:334] "Generic (PLEG): container finished" podID="a5d8a8ba-50c0-4fb6-823f-f28cf0d0f125" containerID="59bdc927190ff946a016c3b38ef86d4b991001a88772e7a4c7e5cc7cbf93b8d3" exitCode=1 Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.899065 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-rs6hx" event={"ID":"a5d8a8ba-50c0-4fb6-823f-f28cf0d0f125","Type":"ContainerDied","Data":"59bdc927190ff946a016c3b38ef86d4b991001a88772e7a4c7e5cc7cbf93b8d3"} Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.900117 4890 scope.go:117] "RemoveContainer" containerID="59bdc927190ff946a016c3b38ef86d4b991001a88772e7a4c7e5cc7cbf93b8d3" Nov 25 15:25:58 crc kubenswrapper[4890]: E1125 15:25:58.900533 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=keystone-operator-controller-manager-748dc6576f-rs6hx_openstack-operators(a5d8a8ba-50c0-4fb6-823f-f28cf0d0f125)\"" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-rs6hx" podUID="a5d8a8ba-50c0-4fb6-823f-f28cf0d0f125" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.905827 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.936574 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.941644 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.976097 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Nov 25 15:25:58 crc kubenswrapper[4890]: I1125 15:25:58.994151 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Nov 25 15:25:59 crc kubenswrapper[4890]: I1125 15:25:59.015469 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 25 15:25:59 crc kubenswrapper[4890]: I1125 15:25:59.026538 4890 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-v4zm5" Nov 25 15:25:59 crc kubenswrapper[4890]: I1125 15:25:59.027775 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-cvp79" Nov 25 15:25:59 crc kubenswrapper[4890]: I1125 15:25:59.123607 4890 scope.go:117] "RemoveContainer" containerID="e0214f8d055e3e0e07dd810288a0139e23c1d5a5bcfdaa0457141e014e32d69d" Nov 25 15:25:59 crc kubenswrapper[4890]: I1125 15:25:59.159545 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 25 15:25:59 crc kubenswrapper[4890]: I1125 15:25:59.172971 4890 scope.go:117] "RemoveContainer" containerID="57328d9eeb31abecef2987ab0e98f3bd5485e45dbe190b251c30dc54effefe13" Nov 25 15:25:59 crc kubenswrapper[4890]: I1125 15:25:59.208065 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Nov 25 15:25:59 crc kubenswrapper[4890]: I1125 15:25:59.304994 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 25 15:25:59 crc kubenswrapper[4890]: I1125 15:25:59.314005 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 25 15:25:59 crc kubenswrapper[4890]: I1125 15:25:59.384088 4890 scope.go:117] "RemoveContainer" containerID="6de55425fa91945cc6e8f62649df335dd495a944f0f7aed9121c87e80c8f3a96" Nov 25 15:25:59 crc kubenswrapper[4890]: I1125 15:25:59.417364 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 25 15:25:59 crc kubenswrapper[4890]: I1125 15:25:59.419810 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 25 15:25:59 crc kubenswrapper[4890]: I1125 15:25:59.446731 4890 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 25 15:25:59 crc kubenswrapper[4890]: I1125 15:25:59.477424 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-f28kv" Nov 25 15:25:59 crc kubenswrapper[4890]: I1125 15:25:59.515908 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 25 15:25:59 crc kubenswrapper[4890]: I1125 15:25:59.560486 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 25 15:25:59 crc kubenswrapper[4890]: I1125 15:25:59.598085 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 25 15:25:59 crc kubenswrapper[4890]: I1125 15:25:59.611189 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 25 15:25:59 crc kubenswrapper[4890]: I1125 15:25:59.664231 4890 scope.go:117] "RemoveContainer" containerID="77f771985277fc92c381c09a36784838a56dce95cd158b58ae0dd79c46671845" Nov 25 15:25:59 crc kubenswrapper[4890]: I1125 15:25:59.681926 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 25 15:25:59 crc kubenswrapper[4890]: I1125 15:25:59.725698 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 25 15:25:59 crc kubenswrapper[4890]: I1125 15:25:59.728128 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 25 15:25:59 crc kubenswrapper[4890]: I1125 15:25:59.732858 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Nov 25 15:25:59 crc kubenswrapper[4890]: I1125 15:25:59.775474 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Nov 25 15:25:59 crc kubenswrapper[4890]: I1125 15:25:59.803802 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Nov 25 15:25:59 crc kubenswrapper[4890]: I1125 15:25:59.851499 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Nov 25 15:25:59 crc kubenswrapper[4890]: I1125 15:25:59.874572 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-qvvbs" Nov 25 15:25:59 crc kubenswrapper[4890]: I1125 15:25:59.912200 4890 scope.go:117] "RemoveContainer" containerID="7eed3f7f71023e9126d0de61d864c2912c3fd5e6e53615d8e82e9bb5d98d5081" Nov 25 15:25:59 crc kubenswrapper[4890]: E1125 15:25:59.912659 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=ironic-operator-controller-manager-5bfcdc958c-hxzf4_openstack-operators(679b3bf3-1ebc-4d19-9ad2-588bd41565dd)\"" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-hxzf4" podUID="679b3bf3-1ebc-4d19-9ad2-588bd41565dd" Nov 25 15:25:59 crc kubenswrapper[4890]: I1125 15:25:59.914456 4890 generic.go:334] "Generic (PLEG): container finished" podID="076c2a0e-7fdb-42e1-ab73-e296bba2b7a9" containerID="8052599ccdad737d107b842db95dca88d95c47c4640b7dfd30857525688810ee" exitCode=1 Nov 25 15:25:59 crc kubenswrapper[4890]: I1125 15:25:59.915505 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-lfnwl" event={"ID":"076c2a0e-7fdb-42e1-ab73-e296bba2b7a9","Type":"ContainerDied","Data":"8052599ccdad737d107b842db95dca88d95c47c4640b7dfd30857525688810ee"} Nov 25 15:25:59 crc kubenswrapper[4890]: I1125 15:25:59.915975 4890 scope.go:117] "RemoveContainer" containerID="8052599ccdad737d107b842db95dca88d95c47c4640b7dfd30857525688810ee" Nov 25 15:25:59 crc kubenswrapper[4890]: E1125 15:25:59.916258 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=glance-operator-controller-manager-68b95954c9-lfnwl_openstack-operators(076c2a0e-7fdb-42e1-ab73-e296bba2b7a9)\"" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-lfnwl" podUID="076c2a0e-7fdb-42e1-ab73-e296bba2b7a9" Nov 25 15:25:59 crc kubenswrapper[4890]: I1125 15:25:59.924294 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 25 15:25:59 crc kubenswrapper[4890]: I1125 15:25:59.928688 4890 scope.go:117] "RemoveContainer" containerID="137b37bbfac0ae56e61f31ca811e8cf420a7182f7941ed967187fc1ed8ac86f0" Nov 25 15:25:59 crc kubenswrapper[4890]: E1125 15:25:59.928937 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=metallb-operator-controller-manager-66fc8c7fbf-lnjxm_metallb-system(dad943fe-2209-40f9-a9bb-0be7ba80cca2)\"" pod="metallb-system/metallb-operator-controller-manager-66fc8c7fbf-lnjxm" podUID="dad943fe-2209-40f9-a9bb-0be7ba80cca2" Nov 25 15:25:59 crc kubenswrapper[4890]: I1125 15:25:59.933074 4890 scope.go:117] "RemoveContainer" containerID="5085d0d9569ced239129365e8bcff54f0f41fba51b79ac5ddca34271333273d9" Nov 25 15:25:59 crc kubenswrapper[4890]: I1125 15:25:59.933144 4890 scope.go:117] "RemoveContainer" containerID="ea03217bf38df8f664326633d6bc9735c9f7d24feebcf0fb7ac30106bfefa2d0" Nov 25 15:25:59 crc kubenswrapper[4890]: I1125 15:25:59.933679 4890 scope.go:117] "RemoveContainer" containerID="aabb0ea70edd64ac6611bc788292aa0806bdccd21d77f87d39aa68b59663cf81" Nov 25 15:25:59 crc kubenswrapper[4890]: I1125 15:25:59.933752 4890 scope.go:117] "RemoveContainer" containerID="afdecad3ee6dca827e01ecfd5110e60bbd29e083f201918f083e2ccc5e1ba964" Nov 25 15:25:59 crc kubenswrapper[4890]: E1125 15:25:59.933909 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=swift-operator-controller-manager-6fdc4fcf86-dzkgt_openstack-operators(861317e1-222d-4f14-b931-e9e7d100ebd6)\"" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-dzkgt" podUID="861317e1-222d-4f14-b931-e9e7d100ebd6" Nov 25 15:25:59 crc kubenswrapper[4890]: E1125 15:25:59.934031 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=nova-operator-controller-manager-79556f57fc-6jq7q_openstack-operators(0769cde4-b49f-41c5-b720-e0dbc667eb4a)\"" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6jq7q" podUID="0769cde4-b49f-41c5-b720-e0dbc667eb4a" Nov 25 15:25:59 crc kubenswrapper[4890]: I1125 15:25:59.934274 4890 scope.go:117] "RemoveContainer" containerID="59bdc927190ff946a016c3b38ef86d4b991001a88772e7a4c7e5cc7cbf93b8d3" Nov 25 15:25:59 crc kubenswrapper[4890]: I1125 15:25:59.934439 4890 scope.go:117] "RemoveContainer" containerID="57618fb9802f27a586bce5ed3b16f89465e88061eac4872f5ae34829c7d008e2" Nov 25 15:25:59 crc kubenswrapper[4890]: E1125 15:25:59.934498 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=cinder-operator-controller-manager-79856dc55c-jcw7x_openstack-operators(0bdbada3-e08c-4f8a-bd55-bac955d4370d)\"" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-jcw7x" podUID="0bdbada3-e08c-4f8a-bd55-bac955d4370d" Nov 25 15:25:59 crc kubenswrapper[4890]: E1125 15:25:59.934590 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=keystone-operator-controller-manager-748dc6576f-rs6hx_openstack-operators(a5d8a8ba-50c0-4fb6-823f-f28cf0d0f125)\"" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-rs6hx" podUID="a5d8a8ba-50c0-4fb6-823f-f28cf0d0f125" Nov 25 15:25:59 crc kubenswrapper[4890]: E1125 15:25:59.934639 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-state-metrics pod=kube-state-metrics-0_openstack(a47203e9-a88a-42f7-86e1-164c6ad532b5)\"" pod="openstack/kube-state-metrics-0" podUID="a47203e9-a88a-42f7-86e1-164c6ad532b5" Nov 25 15:25:59 crc kubenswrapper[4890]: E1125 15:25:59.934761 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=heat-operator-controller-manager-774b86978c-xvfgd_openstack-operators(72c26c5b-3370-4994-a5f0-82128c62c2df)\"" pod="openstack-operators/heat-operator-controller-manager-774b86978c-xvfgd" podUID="72c26c5b-3370-4994-a5f0-82128c62c2df" Nov 25 15:25:59 crc kubenswrapper[4890]: I1125 15:25:59.934788 4890 scope.go:117] "RemoveContainer" containerID="2545dda572d458934e753cbbe89f8cbd4188e0160a44a1d5caf399b84c9fb657" Nov 25 15:25:59 crc kubenswrapper[4890]: I1125 15:25:59.934873 4890 scope.go:117] "RemoveContainer" containerID="dc8aff712070f60aaf997f341250476ab00d06875bc86901bf5c918d9c5a94d9" Nov 25 15:25:59 crc kubenswrapper[4890]: I1125 15:25:59.934902 4890 scope.go:117] "RemoveContainer" containerID="24f5582ca11aa9c0ac5997208248ab279725e93ef77197c94bd72f340f8bbab7" Nov 25 15:25:59 crc kubenswrapper[4890]: E1125 15:25:59.935100 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=telemetry-operator-controller-manager-567f98c9d-6zr4x_openstack-operators(d50aec9d-1673-4b81-bb57-04cae418843e)\"" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-6zr4x" podUID="d50aec9d-1673-4b81-bb57-04cae418843e" Nov 25 15:25:59 crc kubenswrapper[4890]: E1125 15:25:59.935104 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=horizon-operator-controller-manager-68c9694994-j6pqs_openstack-operators(c528348f-0907-423a-a0c2-f15cf18ff99a)\"" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-j6pqs" podUID="c528348f-0907-423a-a0c2-f15cf18ff99a" Nov 25 15:25:59 crc kubenswrapper[4890]: E1125 15:25:59.935122 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=barbican-operator-controller-manager-86dc4d89c8-pmmf2_openstack-operators(d175a35b-8b9b-49c0-8fc2-02ef62f131d9)\"" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-pmmf2" podUID="d175a35b-8b9b-49c0-8fc2-02ef62f131d9" Nov 25 15:25:59 crc kubenswrapper[4890]: I1125 15:25:59.935545 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-66fc8c7fbf-lnjxm" Nov 25 15:25:59 crc kubenswrapper[4890]: I1125 15:25:59.942433 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Nov 25 15:25:59 crc kubenswrapper[4890]: I1125 15:25:59.961868 4890 scope.go:117] "RemoveContainer" containerID="286d09d1c4a9ff164ba4164079699a9e512eed1984749d1fee1c487e3968209c" Nov 25 15:25:59 crc kubenswrapper[4890]: I1125 15:25:59.965579 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 25 15:25:59 crc kubenswrapper[4890]: I1125 15:25:59.993352 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 25 15:26:00 crc kubenswrapper[4890]: I1125 15:26:00.008776 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 25 15:26:00 crc kubenswrapper[4890]: I1125 15:26:00.022339 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 25 15:26:00 crc kubenswrapper[4890]: I1125 15:26:00.063279 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Nov 25 15:26:00 crc kubenswrapper[4890]: I1125 15:26:00.083058 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 25 15:26:00 crc kubenswrapper[4890]: I1125 15:26:00.122851 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 25 15:26:00 crc kubenswrapper[4890]: I1125 15:26:00.134420 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Nov 25 15:26:00 crc kubenswrapper[4890]: I1125 15:26:00.167566 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Nov 25 15:26:00 crc kubenswrapper[4890]: I1125 15:26:00.175554 4890 scope.go:117] "RemoveContainer" containerID="60ab9e903ab7dd461c9fab060610f14563535916392274aabb88f8d7b90adbde" Nov 25 15:26:00 crc kubenswrapper[4890]: I1125 15:26:00.175764 4890 scope.go:117] "RemoveContainer" containerID="6328b6da81615173e561e6807d29b480ebef9c78fbca0d829f62256dd629eac3" Nov 25 15:26:00 crc kubenswrapper[4890]: I1125 15:26:00.184931 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Nov 25 15:26:00 crc kubenswrapper[4890]: I1125 15:26:00.233557 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 25 15:26:00 crc kubenswrapper[4890]: I1125 15:26:00.259295 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 25 15:26:00 crc kubenswrapper[4890]: I1125 15:26:00.271131 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 25 15:26:00 crc kubenswrapper[4890]: I1125 15:26:00.343966 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-ln68r" Nov 25 15:26:00 crc kubenswrapper[4890]: I1125 15:26:00.359476 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 25 15:26:00 crc kubenswrapper[4890]: I1125 15:26:00.370585 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-6h9n9" Nov 25 15:26:00 crc kubenswrapper[4890]: I1125 15:26:00.455521 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Nov 25 15:26:00 crc kubenswrapper[4890]: I1125 15:26:00.477363 4890 scope.go:117] "RemoveContainer" containerID="2e409da14a5c1cd59defa63e3a546ed3a94ed44d1a0de5a7202afe0f4832568b" Nov 25 15:26:00 crc kubenswrapper[4890]: I1125 15:26:00.520278 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 25 15:26:00 crc kubenswrapper[4890]: I1125 15:26:00.537477 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 25 15:26:00 crc kubenswrapper[4890]: I1125 15:26:00.564348 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Nov 25 15:26:00 crc kubenswrapper[4890]: I1125 15:26:00.633450 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 25 15:26:00 crc kubenswrapper[4890]: I1125 15:26:00.667268 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 25 15:26:00 crc kubenswrapper[4890]: I1125 15:26:00.670233 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Nov 25 15:26:00 crc kubenswrapper[4890]: I1125 15:26:00.678285 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 25 15:26:00 crc kubenswrapper[4890]: I1125 15:26:00.678889 4890 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 25 15:26:00 crc kubenswrapper[4890]: I1125 15:26:00.689273 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 25 15:26:00 crc kubenswrapper[4890]: I1125 15:26:00.699927 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 25 15:26:00 crc kubenswrapper[4890]: I1125 15:26:00.718711 4890 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 25 15:26:00 crc kubenswrapper[4890]: I1125 15:26:00.723921 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 25 15:26:00 crc kubenswrapper[4890]: I1125 15:26:00.753936 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Nov 25 15:26:00 crc kubenswrapper[4890]: I1125 15:26:00.761719 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 25 15:26:00 crc kubenswrapper[4890]: I1125 15:26:00.765846 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 25 15:26:00 crc kubenswrapper[4890]: I1125 15:26:00.777675 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-cn445" Nov 25 15:26:00 crc kubenswrapper[4890]: I1125 15:26:00.810586 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 25 15:26:00 crc kubenswrapper[4890]: I1125 15:26:00.864406 4890 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 25 15:26:00 crc kubenswrapper[4890]: I1125 15:26:00.864493 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 25 15:26:00 crc kubenswrapper[4890]: I1125 15:26:00.891206 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 25 15:26:00 crc kubenswrapper[4890]: I1125 15:26:00.919231 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 25 15:26:00 crc kubenswrapper[4890]: I1125 15:26:00.948263 4890 scope.go:117] "RemoveContainer" containerID="137b37bbfac0ae56e61f31ca811e8cf420a7182f7941ed967187fc1ed8ac86f0" Nov 25 15:26:00 crc kubenswrapper[4890]: E1125 15:26:00.948509 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=metallb-operator-controller-manager-66fc8c7fbf-lnjxm_metallb-system(dad943fe-2209-40f9-a9bb-0be7ba80cca2)\"" pod="metallb-system/metallb-operator-controller-manager-66fc8c7fbf-lnjxm" podUID="dad943fe-2209-40f9-a9bb-0be7ba80cca2" Nov 25 15:26:00 crc kubenswrapper[4890]: I1125 15:26:00.953711 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 25 15:26:00 crc kubenswrapper[4890]: I1125 15:26:00.982642 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 25 15:26:01 crc kubenswrapper[4890]: I1125 15:26:01.010365 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-t2qq4" Nov 25 15:26:01 crc kubenswrapper[4890]: I1125 15:26:01.034083 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 25 15:26:01 crc kubenswrapper[4890]: I1125 15:26:01.038732 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 25 15:26:01 crc kubenswrapper[4890]: I1125 15:26:01.110379 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 25 15:26:01 crc kubenswrapper[4890]: I1125 15:26:01.126902 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-q5l6w" Nov 25 15:26:01 crc kubenswrapper[4890]: I1125 15:26:01.143421 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 25 15:26:01 crc kubenswrapper[4890]: I1125 15:26:01.153873 4890 scope.go:117] "RemoveContainer" containerID="cce83adafd4303a3c4644332384918e2da8852bc205228c4dadd7e6a87ff1cf8" Nov 25 15:26:01 crc kubenswrapper[4890]: I1125 15:26:01.165009 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 25 15:26:01 crc kubenswrapper[4890]: I1125 15:26:01.184586 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 25 15:26:01 crc kubenswrapper[4890]: I1125 15:26:01.241812 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 25 15:26:01 crc kubenswrapper[4890]: I1125 15:26:01.270612 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 25 15:26:01 crc kubenswrapper[4890]: I1125 15:26:01.300326 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 25 15:26:01 crc kubenswrapper[4890]: I1125 15:26:01.359264 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 25 15:26:01 crc kubenswrapper[4890]: I1125 15:26:01.374837 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 25 15:26:01 crc kubenswrapper[4890]: I1125 15:26:01.394402 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-9z5kf" Nov 25 15:26:01 crc kubenswrapper[4890]: I1125 15:26:01.399679 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Nov 25 15:26:01 crc kubenswrapper[4890]: I1125 15:26:01.455821 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Nov 25 15:26:01 crc kubenswrapper[4890]: I1125 15:26:01.459598 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 25 15:26:01 crc kubenswrapper[4890]: I1125 15:26:01.488376 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 25 15:26:01 crc kubenswrapper[4890]: I1125 15:26:01.544875 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 25 15:26:01 crc kubenswrapper[4890]: I1125 15:26:01.558293 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 25 15:26:01 crc kubenswrapper[4890]: I1125 15:26:01.571635 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 25 15:26:01 crc kubenswrapper[4890]: I1125 15:26:01.594435 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 25 15:26:01 crc kubenswrapper[4890]: I1125 15:26:01.602475 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 25 15:26:01 crc kubenswrapper[4890]: I1125 15:26:01.747810 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 25 15:26:01 crc kubenswrapper[4890]: I1125 15:26:01.775122 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 25 15:26:01 crc kubenswrapper[4890]: I1125 15:26:01.847016 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-9cgrd" Nov 25 15:26:01 crc kubenswrapper[4890]: I1125 15:26:01.866568 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 25 15:26:01 crc kubenswrapper[4890]: I1125 15:26:01.872606 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 25 15:26:01 crc kubenswrapper[4890]: I1125 15:26:01.873812 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 25 15:26:01 crc kubenswrapper[4890]: I1125 15:26:01.902905 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 25 15:26:01 crc kubenswrapper[4890]: I1125 15:26:01.912839 4890 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 25 15:26:01 crc kubenswrapper[4890]: I1125 15:26:01.973235 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Nov 25 15:26:01 crc kubenswrapper[4890]: I1125 15:26:01.981428 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 25 15:26:02 crc kubenswrapper[4890]: I1125 15:26:02.062308 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 25 15:26:02 crc kubenswrapper[4890]: I1125 15:26:02.109695 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Nov 25 15:26:02 crc kubenswrapper[4890]: I1125 15:26:02.144245 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 25 15:26:02 crc kubenswrapper[4890]: I1125 15:26:02.199900 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 25 15:26:02 crc kubenswrapper[4890]: I1125 15:26:02.219870 4890 scope.go:117] "RemoveContainer" containerID="01cc934a2ae0af3a25c372a4e138035e62b92af8b615443d315c5b0d7695f5c1" Nov 25 15:26:02 crc kubenswrapper[4890]: I1125 15:26:02.220081 4890 scope.go:117] "RemoveContainer" containerID="5bdc10c6af2349a7635d97fa9053a19229cec02d80cae8cde73fc72b252787eb" Nov 25 15:26:02 crc kubenswrapper[4890]: I1125 15:26:02.220596 4890 scope.go:117] "RemoveContainer" containerID="45be6c4aa1f57c239660be552320d18f875cde9c70ec5abb3fa04a6ed3e9139e" Nov 25 15:26:02 crc kubenswrapper[4890]: I1125 15:26:02.221585 4890 scope.go:117] "RemoveContainer" containerID="f516cdea0e4e13c5f35f9cf192df13cf34b38f3a05d2fb0db0fb898037f8319c" Nov 25 15:26:02 crc kubenswrapper[4890]: I1125 15:26:02.252958 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Nov 25 15:26:02 crc kubenswrapper[4890]: I1125 15:26:02.268528 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Nov 25 15:26:02 crc kubenswrapper[4890]: I1125 15:26:02.310616 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 25 15:26:02 crc kubenswrapper[4890]: I1125 15:26:02.351637 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 25 15:26:02 crc kubenswrapper[4890]: I1125 15:26:02.363750 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 25 15:26:02 crc kubenswrapper[4890]: I1125 15:26:02.372586 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Nov 25 15:26:02 crc kubenswrapper[4890]: I1125 15:26:02.455730 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 25 15:26:02 crc kubenswrapper[4890]: I1125 15:26:02.462103 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 25 15:26:02 crc kubenswrapper[4890]: I1125 15:26:02.468285 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 25 15:26:02 crc kubenswrapper[4890]: I1125 15:26:02.501062 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 25 15:26:02 crc kubenswrapper[4890]: I1125 15:26:02.538914 4890 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 25 15:26:02 crc kubenswrapper[4890]: I1125 15:26:02.545893 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 25 15:26:02 crc kubenswrapper[4890]: I1125 15:26:02.587707 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 25 15:26:02 crc kubenswrapper[4890]: I1125 15:26:02.615707 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 25 15:26:02 crc kubenswrapper[4890]: I1125 15:26:02.680135 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 25 15:26:02 crc kubenswrapper[4890]: I1125 15:26:02.701679 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Nov 25 15:26:02 crc kubenswrapper[4890]: I1125 15:26:02.757358 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Nov 25 15:26:02 crc kubenswrapper[4890]: I1125 15:26:02.758712 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Nov 25 15:26:02 crc kubenswrapper[4890]: I1125 15:26:02.802727 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 25 15:26:02 crc kubenswrapper[4890]: I1125 15:26:02.803499 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 25 15:26:02 crc kubenswrapper[4890]: I1125 15:26:02.846105 4890 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 25 15:26:02 crc kubenswrapper[4890]: I1125 15:26:02.872121 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 25 15:26:02 crc kubenswrapper[4890]: I1125 15:26:02.887205 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 25 15:26:02 crc kubenswrapper[4890]: I1125 15:26:02.922341 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 25 15:26:02 crc kubenswrapper[4890]: I1125 15:26:02.954687 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 25 15:26:02 crc kubenswrapper[4890]: I1125 15:26:02.985495 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-slpt4" event={"ID":"f272410f-230d-43f2-946e-e34e4d9c05ea","Type":"ContainerStarted","Data":"f0d0bf516dcd47d736a52b93852c7285f0e58e2e35a41f80221eda17385548d1"} Nov 25 15:26:02 crc kubenswrapper[4890]: I1125 15:26:02.990288 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-nshw7" event={"ID":"9f0e3cf9-dcc0-405f-a2de-9148844ec3ae","Type":"ContainerStarted","Data":"a0074090c38ba41c6cd974ca26e67cc1c9cd2d25edd77a01b7f4e31a93593803"} Nov 25 15:26:02 crc kubenswrapper[4890]: I1125 15:26:02.992925 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Nov 25 15:26:03 crc kubenswrapper[4890]: I1125 15:26:03.033527 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 25 15:26:03 crc kubenswrapper[4890]: I1125 15:26:03.039812 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-9nwng" Nov 25 15:26:03 crc kubenswrapper[4890]: I1125 15:26:03.041192 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 25 15:26:03 crc kubenswrapper[4890]: I1125 15:26:03.054680 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 25 15:26:03 crc kubenswrapper[4890]: I1125 15:26:03.073903 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-zr9nn" Nov 25 15:26:03 crc kubenswrapper[4890]: I1125 15:26:03.082989 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-s7rlv" Nov 25 15:26:03 crc kubenswrapper[4890]: I1125 15:26:03.125886 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Nov 25 15:26:03 crc kubenswrapper[4890]: I1125 15:26:03.141089 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 25 15:26:03 crc kubenswrapper[4890]: I1125 15:26:03.147357 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 25 15:26:03 crc kubenswrapper[4890]: I1125 15:26:03.173145 4890 scope.go:117] "RemoveContainer" containerID="0770c2f791722b326d3323baa96eca63ec6eac08347883ea0d8f92a8abb2ed7d" Nov 25 15:26:03 crc kubenswrapper[4890]: I1125 15:26:03.173500 4890 scope.go:117] "RemoveContainer" containerID="34aad7779779b8785f00f0293acf127026ceeeb1d0bf6285415e611bc6e3be23" Nov 25 15:26:03 crc kubenswrapper[4890]: I1125 15:26:03.180034 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Nov 25 15:26:03 crc kubenswrapper[4890]: I1125 15:26:03.195410 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 25 15:26:03 crc kubenswrapper[4890]: I1125 15:26:03.211907 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 25 15:26:03 crc kubenswrapper[4890]: I1125 15:26:03.214287 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Nov 25 15:26:03 crc kubenswrapper[4890]: I1125 15:26:03.216857 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 25 15:26:03 crc kubenswrapper[4890]: I1125 15:26:03.308991 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 25 15:26:03 crc kubenswrapper[4890]: I1125 15:26:03.323792 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 25 15:26:03 crc kubenswrapper[4890]: I1125 15:26:03.333241 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 25 15:26:03 crc kubenswrapper[4890]: I1125 15:26:03.366123 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 25 15:26:03 crc kubenswrapper[4890]: I1125 15:26:03.375526 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 25 15:26:03 crc kubenswrapper[4890]: I1125 15:26:03.459969 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 25 15:26:03 crc kubenswrapper[4890]: I1125 15:26:03.538784 4890 scope.go:117] "RemoveContainer" containerID="52f8db982762a517d8d96beb141da93c238d379ce39501ef137cb21b7fdcf14d" Nov 25 15:26:03 crc kubenswrapper[4890]: I1125 15:26:03.566779 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 25 15:26:03 crc kubenswrapper[4890]: I1125 15:26:03.571835 4890 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 25 15:26:03 crc kubenswrapper[4890]: I1125 15:26:03.572548 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=64.572533522 podStartE2EDuration="1m4.572533522s" podCreationTimestamp="2025-11-25 15:24:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:25:44.181764054 +0000 UTC m=+1402.624226684" watchObservedRunningTime="2025-11-25 15:26:03.572533522 +0000 UTC m=+1422.014996132" Nov 25 15:26:03 crc kubenswrapper[4890]: I1125 15:26:03.577360 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=64.577343922 podStartE2EDuration="1m4.577343922s" podCreationTimestamp="2025-11-25 15:24:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:25:44.208567604 +0000 UTC m=+1402.651030214" watchObservedRunningTime="2025-11-25 15:26:03.577343922 +0000 UTC m=+1422.019806532" Nov 25 15:26:03 crc kubenswrapper[4890]: I1125 15:26:03.577806 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kqz9d" podStartSLOduration=42.163130368 podStartE2EDuration="50.577799404s" podCreationTimestamp="2025-11-25 15:25:13 +0000 UTC" firstStartedPulling="2025-11-25 15:25:15.708253048 +0000 UTC m=+1374.150715658" lastFinishedPulling="2025-11-25 15:25:24.122922084 +0000 UTC m=+1382.565384694" observedRunningTime="2025-11-25 15:25:44.232102243 +0000 UTC m=+1402.674564853" watchObservedRunningTime="2025-11-25 15:26:03.577799404 +0000 UTC m=+1422.020262014" Nov 25 15:26:03 crc kubenswrapper[4890]: I1125 15:26:03.586520 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 25 15:26:03 crc kubenswrapper[4890]: I1125 15:26:03.586573 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 25 15:26:03 crc kubenswrapper[4890]: I1125 15:26:03.587365 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-kr52s" Nov 25 15:26:03 crc kubenswrapper[4890]: I1125 15:26:03.590330 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 15:26:03 crc kubenswrapper[4890]: I1125 15:26:03.617600 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=22.617570748 podStartE2EDuration="22.617570748s" podCreationTimestamp="2025-11-25 15:25:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 15:26:03.610245735 +0000 UTC m=+1422.052708345" watchObservedRunningTime="2025-11-25 15:26:03.617570748 +0000 UTC m=+1422.060033358" Nov 25 15:26:03 crc kubenswrapper[4890]: I1125 15:26:03.698902 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-798gb" Nov 25 15:26:03 crc kubenswrapper[4890]: I1125 15:26:03.710671 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 25 15:26:03 crc kubenswrapper[4890]: I1125 15:26:03.716845 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 25 15:26:03 crc kubenswrapper[4890]: I1125 15:26:03.722620 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 25 15:26:03 crc kubenswrapper[4890]: I1125 15:26:03.747121 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Nov 25 15:26:03 crc kubenswrapper[4890]: I1125 15:26:03.757499 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 25 15:26:03 crc kubenswrapper[4890]: I1125 15:26:03.765478 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 25 15:26:03 crc kubenswrapper[4890]: I1125 15:26:03.800705 4890 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 25 15:26:03 crc kubenswrapper[4890]: I1125 15:26:03.803268 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 25 15:26:03 crc kubenswrapper[4890]: I1125 15:26:03.842339 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 25 15:26:03 crc kubenswrapper[4890]: I1125 15:26:03.893089 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 25 15:26:03 crc kubenswrapper[4890]: I1125 15:26:03.925729 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Nov 25 15:26:03 crc kubenswrapper[4890]: I1125 15:26:03.961768 4890 scope.go:117] "RemoveContainer" containerID="7ff93d6707b3f0b8fb868414ecf039dfc3ca37a71aa91bcc35b9114d425a199f" Nov 25 15:26:03 crc kubenswrapper[4890]: I1125 15:26:03.975638 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 25 15:26:03 crc kubenswrapper[4890]: I1125 15:26:03.992629 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.021264 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.021435 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.021629 4890 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-26pqz" Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.034192 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-829nk" event={"ID":"4364eec5-444a-41f6-af1c-fe16b24bd044","Type":"ContainerStarted","Data":"8453cefea24580a864e1b962ed8a682423dc9aaafad812309969556ac975663c"} Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.034482 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-829nk" Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.044082 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-wvjw7" event={"ID":"03353c8b-91f4-41ae-afc1-c522b1c73bb4","Type":"ContainerStarted","Data":"09ff68e9f7b5d7b971104be89b55602edbba51c72b2e776a35880be99ec6b53f"} Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.044258 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-wvjw7" Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.047814 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-stflb" event={"ID":"7e576602-19f5-4656-84dd-5ab80abca895","Type":"ContainerStarted","Data":"afe47ed941c0114eed272202844f29373d2d8c6c4f5c70cf119a83c56005a08d"} Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.049406 4890 scope.go:117] "RemoveContainer" containerID="afe47ed941c0114eed272202844f29373d2d8c6c4f5c70cf119a83c56005a08d" Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.053814 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.054985 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-vx75t" event={"ID":"29f250db-2a96-41ad-bb8f-5b503a9288d8","Type":"ContainerStarted","Data":"5d93b4988f6bdbdc8c47c62e1f25b2fcb1a65f5d3e949121c19272784f723e37"} Nov 25 15:26:04 crc kubenswrapper[4890]: E1125 15:26:04.055440 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=infra-operator-controller-manager-d5cc86f4b-stflb_openstack-operators(7e576602-19f5-4656-84dd-5ab80abca895)\"" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-stflb" podUID="7e576602-19f5-4656-84dd-5ab80abca895" Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.055855 4890 scope.go:117] "RemoveContainer" containerID="5d93b4988f6bdbdc8c47c62e1f25b2fcb1a65f5d3e949121c19272784f723e37" Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.057224 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 25 15:26:04 crc kubenswrapper[4890]: E1125 15:26:04.060360 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=ovn-operator-controller-manager-66cf5c67ff-vx75t_openstack-operators(29f250db-2a96-41ad-bb8f-5b503a9288d8)\"" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-vx75t" podUID="29f250db-2a96-41ad-bb8f-5b503a9288d8" Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.060686 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.062708 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.064115 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.064269 4890 generic.go:334] "Generic (PLEG): container finished" podID="f272410f-230d-43f2-946e-e34e4d9c05ea" containerID="f0d0bf516dcd47d736a52b93852c7285f0e58e2e35a41f80221eda17385548d1" exitCode=1 Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.064346 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-slpt4" event={"ID":"f272410f-230d-43f2-946e-e34e4d9c05ea","Type":"ContainerDied","Data":"f0d0bf516dcd47d736a52b93852c7285f0e58e2e35a41f80221eda17385548d1"} Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.065049 4890 scope.go:117] "RemoveContainer" containerID="f0d0bf516dcd47d736a52b93852c7285f0e58e2e35a41f80221eda17385548d1" Nov 25 15:26:04 crc kubenswrapper[4890]: E1125 15:26:04.065305 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=manila-operator-controller-manager-58bb8d67cc-slpt4_openstack-operators(f272410f-230d-43f2-946e-e34e4d9c05ea)\"" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-slpt4" podUID="f272410f-230d-43f2-946e-e34e4d9c05ea" Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.066894 4890 scope.go:117] "RemoveContainer" containerID="70634ed74a9c04a3a724fdb83aff92615249c19958b69dd169263857756bf093" Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.069034 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-68587559f4-9b4cf" event={"ID":"3e496c3d-9485-44bc-b3b1-dc1682a88434","Type":"ContainerStarted","Data":"aeb0938dec764a3788ac5a6992985db59cc25e8a9f8cd69f3f2408a7e27cdd24"} Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.069584 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-68587559f4-9b4cf" Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.071486 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-dmlx4" event={"ID":"e26594ec-4bc2-4051-8a49-fc490bd3b6c9","Type":"ContainerStarted","Data":"c186dc6ff6591b7b58925577cec707c6ed3796c30a9ae21bb9565049cea41856"} Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.081973 4890 generic.go:334] "Generic (PLEG): container finished" podID="9f0e3cf9-dcc0-405f-a2de-9148844ec3ae" containerID="a0074090c38ba41c6cd974ca26e67cc1c9cd2d25edd77a01b7f4e31a93593803" exitCode=1 Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.082366 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-nshw7" event={"ID":"9f0e3cf9-dcc0-405f-a2de-9148844ec3ae","Type":"ContainerDied","Data":"a0074090c38ba41c6cd974ca26e67cc1c9cd2d25edd77a01b7f4e31a93593803"} Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.082789 4890 scope.go:117] "RemoveContainer" containerID="a0074090c38ba41c6cd974ca26e67cc1c9cd2d25edd77a01b7f4e31a93593803" Nov 25 15:26:04 crc kubenswrapper[4890]: E1125 15:26:04.083118 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=placement-operator-controller-manager-5db546f9d9-nshw7_openstack-operators(9f0e3cf9-dcc0-405f-a2de-9148844ec3ae)\"" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-nshw7" podUID="9f0e3cf9-dcc0-405f-a2de-9148844ec3ae" Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.128680 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.172210 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-7fwnk" Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.216679 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.221894 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.236588 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.273073 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.315466 4890 scope.go:117] "RemoveContainer" containerID="38d0123e0cedad944df1c84a66fc4aba91c298dc840c5489a45e6ec2fd23cc66" Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.355356 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.401481 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.410411 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.415533 4890 scope.go:117] "RemoveContainer" containerID="57328d9eeb31abecef2987ab0e98f3bd5485e45dbe190b251c30dc54effefe13" Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.438276 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.530906 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.576298 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-wxzlp" Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.607111 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.622789 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.637232 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-44nwj" Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.665648 4890 scope.go:117] "RemoveContainer" containerID="6328b6da81615173e561e6807d29b480ebef9c78fbca0d829f62256dd629eac3" Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.738977 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.760745 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.762557 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.764156 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.795109 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.795246 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.849656 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.853031 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.881286 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.912133 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-4gbz6" Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.917108 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-hqgtf" Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.932422 4890 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-4kv2m" Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.932460 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.959949 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Nov 25 15:26:04 crc kubenswrapper[4890]: I1125 15:26:04.978091 4890 scope.go:117] "RemoveContainer" containerID="60ab9e903ab7dd461c9fab060610f14563535916392274aabb88f8d7b90adbde" Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.101096 4890 generic.go:334] "Generic (PLEG): container finished" podID="7e576602-19f5-4656-84dd-5ab80abca895" containerID="afe47ed941c0114eed272202844f29373d2d8c6c4f5c70cf119a83c56005a08d" exitCode=1 Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.101791 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-stflb" event={"ID":"7e576602-19f5-4656-84dd-5ab80abca895","Type":"ContainerDied","Data":"afe47ed941c0114eed272202844f29373d2d8c6c4f5c70cf119a83c56005a08d"} Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.101835 4890 scope.go:117] "RemoveContainer" containerID="01cc934a2ae0af3a25c372a4e138035e62b92af8b615443d315c5b0d7695f5c1" Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.102532 4890 scope.go:117] "RemoveContainer" containerID="afe47ed941c0114eed272202844f29373d2d8c6c4f5c70cf119a83c56005a08d" Nov 25 15:26:05 crc kubenswrapper[4890]: E1125 15:26:05.102860 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=infra-operator-controller-manager-d5cc86f4b-stflb_openstack-operators(7e576602-19f5-4656-84dd-5ab80abca895)\"" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-stflb" podUID="7e576602-19f5-4656-84dd-5ab80abca895" Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.113594 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.115915 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.117819 4890 generic.go:334] "Generic (PLEG): container finished" podID="e26594ec-4bc2-4051-8a49-fc490bd3b6c9" containerID="c186dc6ff6591b7b58925577cec707c6ed3796c30a9ae21bb9565049cea41856" exitCode=1 Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.117878 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-dmlx4" event={"ID":"e26594ec-4bc2-4051-8a49-fc490bd3b6c9","Type":"ContainerDied","Data":"c186dc6ff6591b7b58925577cec707c6ed3796c30a9ae21bb9565049cea41856"} Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.118518 4890 scope.go:117] "RemoveContainer" containerID="c186dc6ff6591b7b58925577cec707c6ed3796c30a9ae21bb9565049cea41856" Nov 25 15:26:05 crc kubenswrapper[4890]: E1125 15:26:05.118856 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with CrashLoopBackOff: \"back-off 20s restarting failed container=operator pod=rabbitmq-cluster-operator-manager-668c99d594-dmlx4_openstack-operators(e26594ec-4bc2-4051-8a49-fc490bd3b6c9)\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-dmlx4" podUID="e26594ec-4bc2-4051-8a49-fc490bd3b6c9" Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.124009 4890 generic.go:334] "Generic (PLEG): container finished" podID="4364eec5-444a-41f6-af1c-fe16b24bd044" containerID="8453cefea24580a864e1b962ed8a682423dc9aaafad812309969556ac975663c" exitCode=1 Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.124057 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-829nk" event={"ID":"4364eec5-444a-41f6-af1c-fe16b24bd044","Type":"ContainerDied","Data":"8453cefea24580a864e1b962ed8a682423dc9aaafad812309969556ac975663c"} Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.124634 4890 scope.go:117] "RemoveContainer" containerID="8453cefea24580a864e1b962ed8a682423dc9aaafad812309969556ac975663c" Nov 25 15:26:05 crc kubenswrapper[4890]: E1125 15:26:05.124906 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=neutron-operator-controller-manager-7c57c8bbc4-829nk_openstack-operators(4364eec5-444a-41f6-af1c-fe16b24bd044)\"" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-829nk" podUID="4364eec5-444a-41f6-af1c-fe16b24bd044" Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.127673 4890 scope.go:117] "RemoveContainer" containerID="a0074090c38ba41c6cd974ca26e67cc1c9cd2d25edd77a01b7f4e31a93593803" Nov 25 15:26:05 crc kubenswrapper[4890]: E1125 15:26:05.127949 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=placement-operator-controller-manager-5db546f9d9-nshw7_openstack-operators(9f0e3cf9-dcc0-405f-a2de-9148844ec3ae)\"" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-nshw7" podUID="9f0e3cf9-dcc0-405f-a2de-9148844ec3ae" Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.137714 4890 generic.go:334] "Generic (PLEG): container finished" podID="03353c8b-91f4-41ae-afc1-c522b1c73bb4" containerID="09ff68e9f7b5d7b971104be89b55602edbba51c72b2e776a35880be99ec6b53f" exitCode=1 Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.137782 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-wvjw7" event={"ID":"03353c8b-91f4-41ae-afc1-c522b1c73bb4","Type":"ContainerDied","Data":"09ff68e9f7b5d7b971104be89b55602edbba51c72b2e776a35880be99ec6b53f"} Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.138487 4890 scope.go:117] "RemoveContainer" containerID="09ff68e9f7b5d7b971104be89b55602edbba51c72b2e776a35880be99ec6b53f" Nov 25 15:26:05 crc kubenswrapper[4890]: E1125 15:26:05.138805 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=mariadb-operator-controller-manager-cb6c4fdb7-wvjw7_openstack-operators(03353c8b-91f4-41ae-afc1-c522b1c73bb4)\"" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-wvjw7" podUID="03353c8b-91f4-41ae-afc1-c522b1c73bb4" Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.144332 4890 generic.go:334] "Generic (PLEG): container finished" podID="29f250db-2a96-41ad-bb8f-5b503a9288d8" containerID="5d93b4988f6bdbdc8c47c62e1f25b2fcb1a65f5d3e949121c19272784f723e37" exitCode=1 Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.144427 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-vx75t" event={"ID":"29f250db-2a96-41ad-bb8f-5b503a9288d8","Type":"ContainerDied","Data":"5d93b4988f6bdbdc8c47c62e1f25b2fcb1a65f5d3e949121c19272784f723e37"} Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.145095 4890 scope.go:117] "RemoveContainer" containerID="5d93b4988f6bdbdc8c47c62e1f25b2fcb1a65f5d3e949121c19272784f723e37" Nov 25 15:26:05 crc kubenswrapper[4890]: E1125 15:26:05.145331 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=ovn-operator-controller-manager-66cf5c67ff-vx75t_openstack-operators(29f250db-2a96-41ad-bb8f-5b503a9288d8)\"" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-vx75t" podUID="29f250db-2a96-41ad-bb8f-5b503a9288d8" Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.155396 4890 scope.go:117] "RemoveContainer" containerID="f0d0bf516dcd47d736a52b93852c7285f0e58e2e35a41f80221eda17385548d1" Nov 25 15:26:05 crc kubenswrapper[4890]: E1125 15:26:05.156446 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=manila-operator-controller-manager-58bb8d67cc-slpt4_openstack-operators(f272410f-230d-43f2-946e-e34e4d9c05ea)\"" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-slpt4" podUID="f272410f-230d-43f2-946e-e34e4d9c05ea" Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.162488 4890 generic.go:334] "Generic (PLEG): container finished" podID="3e496c3d-9485-44bc-b3b1-dc1682a88434" containerID="aeb0938dec764a3788ac5a6992985db59cc25e8a9f8cd69f3f2408a7e27cdd24" exitCode=1 Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.162538 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-68587559f4-9b4cf" event={"ID":"3e496c3d-9485-44bc-b3b1-dc1682a88434","Type":"ContainerDied","Data":"aeb0938dec764a3788ac5a6992985db59cc25e8a9f8cd69f3f2408a7e27cdd24"} Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.164728 4890 scope.go:117] "RemoveContainer" containerID="aeb0938dec764a3788ac5a6992985db59cc25e8a9f8cd69f3f2408a7e27cdd24" Nov 25 15:26:05 crc kubenswrapper[4890]: E1125 15:26:05.165116 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=openstack-operator-controller-manager-68587559f4-9b4cf_openstack-operators(3e496c3d-9485-44bc-b3b1-dc1682a88434)\"" pod="openstack-operators/openstack-operator-controller-manager-68587559f4-9b4cf" podUID="3e496c3d-9485-44bc-b3b1-dc1682a88434" Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.198341 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.245970 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.251110 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.252840 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.258534 4890 scope.go:117] "RemoveContainer" containerID="45be6c4aa1f57c239660be552320d18f875cde9c70ec5abb3fa04a6ed3e9139e" Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.282527 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.305769 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.335704 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.349729 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.386386 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.458423 4890 scope.go:117] "RemoveContainer" containerID="f516cdea0e4e13c5f35f9cf192df13cf34b38f3a05d2fb0db0fb898037f8319c" Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.516718 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-lwcrw" Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.537790 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.590824 4890 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.591043 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://f8010cc46ab44d816173f84c66043f945180f00900ca208dc599fb56bcf061b2" gracePeriod=5 Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.600643 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.607142 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.652366 4890 scope.go:117] "RemoveContainer" containerID="34aad7779779b8785f00f0293acf127026ceeeb1d0bf6285415e611bc6e3be23" Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.660823 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.660845 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.739928 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.740737 4890 scope.go:117] "RemoveContainer" containerID="57618fb9802f27a586bce5ed3b16f89465e88061eac4872f5ae34829c7d008e2" Nov 25 15:26:05 crc kubenswrapper[4890]: E1125 15:26:05.742128 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-state-metrics pod=kube-state-metrics-0_openstack(a47203e9-a88a-42f7-86e1-164c6ad532b5)\"" pod="openstack/kube-state-metrics-0" podUID="a47203e9-a88a-42f7-86e1-164c6ad532b5" Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.744231 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.749128 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/kube-state-metrics-0" Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.751545 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.787711 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-2k6hj" Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.791433 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.852593 4890 scope.go:117] "RemoveContainer" containerID="0770c2f791722b326d3323baa96eca63ec6eac08347883ea0d8f92a8abb2ed7d" Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.855048 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.860508 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.861380 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.948222 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.959619 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 25 15:26:05 crc kubenswrapper[4890]: I1125 15:26:05.991284 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.000258 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.001451 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.032531 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.036384 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.042635 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-ldsnl" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.061675 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.061679 4890 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.074264 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.101863 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-6b6q8" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.109991 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-pmmf2" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.110699 4890 scope.go:117] "RemoveContainer" containerID="24f5582ca11aa9c0ac5997208248ab279725e93ef77197c94bd72f340f8bbab7" Nov 25 15:26:06 crc kubenswrapper[4890]: E1125 15:26:06.110923 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=barbican-operator-controller-manager-86dc4d89c8-pmmf2_openstack-operators(d175a35b-8b9b-49c0-8fc2-02ef62f131d9)\"" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-pmmf2" podUID="d175a35b-8b9b-49c0-8fc2-02ef62f131d9" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.111083 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-pmmf2" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.115091 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.115551 4890 scope.go:117] "RemoveContainer" containerID="5bdc10c6af2349a7635d97fa9053a19229cec02d80cae8cde73fc72b252787eb" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.116778 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-jcw7x" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.117186 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-jcw7x" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.117853 4890 scope.go:117] "RemoveContainer" containerID="ea03217bf38df8f664326633d6bc9735c9f7d24feebcf0fb7ac30106bfefa2d0" Nov 25 15:26:06 crc kubenswrapper[4890]: E1125 15:26:06.118081 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=cinder-operator-controller-manager-79856dc55c-jcw7x_openstack-operators(0bdbada3-e08c-4f8a-bd55-bac955d4370d)\"" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-jcw7x" podUID="0bdbada3-e08c-4f8a-bd55-bac955d4370d" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.123684 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-z4bgs" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.134207 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-stwjm" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.135289 4890 scope.go:117] "RemoveContainer" containerID="e7b7fcdef297e60c89760ed3c86c2d93e8c7a6a0088d065d1090410732523bb5" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.138046 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.157996 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-lfnwl" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.158047 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-lfnwl" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.158521 4890 scope.go:117] "RemoveContainer" containerID="8052599ccdad737d107b842db95dca88d95c47c4640b7dfd30857525688810ee" Nov 25 15:26:06 crc kubenswrapper[4890]: E1125 15:26:06.158766 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=glance-operator-controller-manager-68b95954c9-lfnwl_openstack-operators(076c2a0e-7fdb-42e1-ab73-e296bba2b7a9)\"" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-lfnwl" podUID="076c2a0e-7fdb-42e1-ab73-e296bba2b7a9" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.185776 4890 scope.go:117] "RemoveContainer" containerID="8453cefea24580a864e1b962ed8a682423dc9aaafad812309969556ac975663c" Nov 25 15:26:06 crc kubenswrapper[4890]: E1125 15:26:06.186012 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=neutron-operator-controller-manager-7c57c8bbc4-829nk_openstack-operators(4364eec5-444a-41f6-af1c-fe16b24bd044)\"" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-829nk" podUID="4364eec5-444a-41f6-af1c-fe16b24bd044" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.194498 4890 scope.go:117] "RemoveContainer" containerID="09ff68e9f7b5d7b971104be89b55602edbba51c72b2e776a35880be99ec6b53f" Nov 25 15:26:06 crc kubenswrapper[4890]: E1125 15:26:06.194742 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=mariadb-operator-controller-manager-cb6c4fdb7-wvjw7_openstack-operators(03353c8b-91f4-41ae-afc1-c522b1c73bb4)\"" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-wvjw7" podUID="03353c8b-91f4-41ae-afc1-c522b1c73bb4" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.199445 4890 scope.go:117] "RemoveContainer" containerID="aeb0938dec764a3788ac5a6992985db59cc25e8a9f8cd69f3f2408a7e27cdd24" Nov 25 15:26:06 crc kubenswrapper[4890]: E1125 15:26:06.199739 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=openstack-operator-controller-manager-68587559f4-9b4cf_openstack-operators(3e496c3d-9485-44bc-b3b1-dc1682a88434)\"" pod="openstack-operators/openstack-operator-controller-manager-68587559f4-9b4cf" podUID="3e496c3d-9485-44bc-b3b1-dc1682a88434" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.199813 4890 scope.go:117] "RemoveContainer" containerID="24f5582ca11aa9c0ac5997208248ab279725e93ef77197c94bd72f340f8bbab7" Nov 25 15:26:06 crc kubenswrapper[4890]: E1125 15:26:06.200149 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=barbican-operator-controller-manager-86dc4d89c8-pmmf2_openstack-operators(d175a35b-8b9b-49c0-8fc2-02ef62f131d9)\"" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-pmmf2" podUID="d175a35b-8b9b-49c0-8fc2-02ef62f131d9" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.200403 4890 scope.go:117] "RemoveContainer" containerID="57618fb9802f27a586bce5ed3b16f89465e88061eac4872f5ae34829c7d008e2" Nov 25 15:26:06 crc kubenswrapper[4890]: E1125 15:26:06.200580 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-state-metrics pod=kube-state-metrics-0_openstack(a47203e9-a88a-42f7-86e1-164c6ad532b5)\"" pod="openstack/kube-state-metrics-0" podUID="a47203e9-a88a-42f7-86e1-164c6ad532b5" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.200627 4890 scope.go:117] "RemoveContainer" containerID="ea03217bf38df8f664326633d6bc9735c9f7d24feebcf0fb7ac30106bfefa2d0" Nov 25 15:26:06 crc kubenswrapper[4890]: E1125 15:26:06.200793 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=cinder-operator-controller-manager-79856dc55c-jcw7x_openstack-operators(0bdbada3-e08c-4f8a-bd55-bac955d4370d)\"" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-jcw7x" podUID="0bdbada3-e08c-4f8a-bd55-bac955d4370d" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.207917 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-774b86978c-xvfgd" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.208928 4890 scope.go:117] "RemoveContainer" containerID="5085d0d9569ced239129365e8bcff54f0f41fba51b79ac5ddca34271333273d9" Nov 25 15:26:06 crc kubenswrapper[4890]: E1125 15:26:06.209153 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=heat-operator-controller-manager-774b86978c-xvfgd_openstack-operators(72c26c5b-3370-4994-a5f0-82128c62c2df)\"" pod="openstack-operators/heat-operator-controller-manager-774b86978c-xvfgd" podUID="72c26c5b-3370-4994-a5f0-82128c62c2df" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.209228 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/heat-operator-controller-manager-774b86978c-xvfgd" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.223674 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-j6pqs" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.224963 4890 scope.go:117] "RemoveContainer" containerID="dc8aff712070f60aaf997f341250476ab00d06875bc86901bf5c918d9c5a94d9" Nov 25 15:26:06 crc kubenswrapper[4890]: E1125 15:26:06.225226 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=horizon-operator-controller-manager-68c9694994-j6pqs_openstack-operators(c528348f-0907-423a-a0c2-f15cf18ff99a)\"" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-j6pqs" podUID="c528348f-0907-423a-a0c2-f15cf18ff99a" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.271468 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.276954 4890 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.285477 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-bgrtc" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.295544 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-68dvd" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.297267 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.302270 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-hxzf4" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.303004 4890 scope.go:117] "RemoveContainer" containerID="7eed3f7f71023e9126d0de61d864c2912c3fd5e6e53615d8e82e9bb5d98d5081" Nov 25 15:26:06 crc kubenswrapper[4890]: E1125 15:26:06.303275 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=ironic-operator-controller-manager-5bfcdc958c-hxzf4_openstack-operators(679b3bf3-1ebc-4d19-9ad2-588bd41565dd)\"" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-hxzf4" podUID="679b3bf3-1ebc-4d19-9ad2-588bd41565dd" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.303584 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-hxzf4" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.315711 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-rs6hx" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.316499 4890 scope.go:117] "RemoveContainer" containerID="59bdc927190ff946a016c3b38ef86d4b991001a88772e7a4c7e5cc7cbf93b8d3" Nov 25 15:26:06 crc kubenswrapper[4890]: E1125 15:26:06.316792 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=keystone-operator-controller-manager-748dc6576f-rs6hx_openstack-operators(a5d8a8ba-50c0-4fb6-823f-f28cf0d0f125)\"" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-rs6hx" podUID="a5d8a8ba-50c0-4fb6-823f-f28cf0d0f125" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.317193 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-rs6hx" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.338852 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.382270 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.421500 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.424504 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.436640 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-wvjw7" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.440307 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.455410 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.468950 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.492297 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.515884 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-slpt4" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.515935 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-slpt4" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.516693 4890 scope.go:117] "RemoveContainer" containerID="f0d0bf516dcd47d736a52b93852c7285f0e58e2e35a41f80221eda17385548d1" Nov 25 15:26:06 crc kubenswrapper[4890]: E1125 15:26:06.516975 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=manila-operator-controller-manager-58bb8d67cc-slpt4_openstack-operators(f272410f-230d-43f2-946e-e34e4d9c05ea)\"" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-slpt4" podUID="f272410f-230d-43f2-946e-e34e4d9c05ea" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.541456 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.555643 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.567356 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-6zr4x" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.568051 4890 scope.go:117] "RemoveContainer" containerID="2545dda572d458934e753cbbe89f8cbd4188e0160a44a1d5caf399b84c9fb657" Nov 25 15:26:06 crc kubenswrapper[4890]: E1125 15:26:06.568299 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=telemetry-operator-controller-manager-567f98c9d-6zr4x_openstack-operators(d50aec9d-1673-4b81-bb57-04cae418843e)\"" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-6zr4x" podUID="d50aec9d-1673-4b81-bb57-04cae418843e" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.569018 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-6zr4x" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.595922 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.617125 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.617988 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.619677 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.622509 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6jq7q" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.623230 4890 scope.go:117] "RemoveContainer" containerID="afdecad3ee6dca827e01ecfd5110e60bbd29e083f201918f083e2ccc5e1ba964" Nov 25 15:26:06 crc kubenswrapper[4890]: E1125 15:26:06.623475 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=nova-operator-controller-manager-79556f57fc-6jq7q_openstack-operators(0769cde4-b49f-41c5-b720-e0dbc667eb4a)\"" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6jq7q" podUID="0769cde4-b49f-41c5-b720-e0dbc667eb4a" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.623503 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6jq7q" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.633910 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.637045 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-qjthd" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.648941 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-56bcr" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.649615 4890 scope.go:117] "RemoveContainer" containerID="09913842764d83958bb7259fb1a01e61d3da9512eb07f23da446059eb49a046a" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.657943 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.662003 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.675786 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.719400 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-vx75t" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.720267 4890 scope.go:117] "RemoveContainer" containerID="5d93b4988f6bdbdc8c47c62e1f25b2fcb1a65f5d3e949121c19272784f723e37" Nov 25 15:26:06 crc kubenswrapper[4890]: E1125 15:26:06.720523 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=ovn-operator-controller-manager-66cf5c67ff-vx75t_openstack-operators(29f250db-2a96-41ad-bb8f-5b503a9288d8)\"" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-vx75t" podUID="29f250db-2a96-41ad-bb8f-5b503a9288d8" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.720967 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-vx75t" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.744769 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.756101 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-nshw7" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.756140 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-nshw7" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.756568 4890 scope.go:117] "RemoveContainer" containerID="a0074090c38ba41c6cd974ca26e67cc1c9cd2d25edd77a01b7f4e31a93593803" Nov 25 15:26:06 crc kubenswrapper[4890]: E1125 15:26:06.756887 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=placement-operator-controller-manager-5db546f9d9-nshw7_openstack-operators(9f0e3cf9-dcc0-405f-a2de-9148844ec3ae)\"" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-nshw7" podUID="9f0e3cf9-dcc0-405f-a2de-9148844ec3ae" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.775097 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-dzkgt" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.775696 4890 scope.go:117] "RemoveContainer" containerID="aabb0ea70edd64ac6611bc788292aa0806bdccd21d77f87d39aa68b59663cf81" Nov 25 15:26:06 crc kubenswrapper[4890]: E1125 15:26:06.775991 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=swift-operator-controller-manager-6fdc4fcf86-dzkgt_openstack-operators(861317e1-222d-4f14-b931-e9e7d100ebd6)\"" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-dzkgt" podUID="861317e1-222d-4f14-b931-e9e7d100ebd6" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.776140 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-dzkgt" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.790220 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-829nk" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.791075 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.822943 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.915484 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.919485 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.981829 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/watcher-operator-controller-manager-864885998-62ddz" Nov 25 15:26:06 crc kubenswrapper[4890]: I1125 15:26:06.982449 4890 scope.go:117] "RemoveContainer" containerID="4dbafe9e16c939461a203c2af1cab51ba3a18f270889d1e522c1d51306fb473d" Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.020006 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.071663 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.124048 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.126655 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.188193 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-htt2h" Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.211405 4890 generic.go:334] "Generic (PLEG): container finished" podID="a9efeabb-bf1c-4de2-8322-961b284e95cc" containerID="addcdee7c623c641d2c7c579c327f9f105ad3a6312c9860f69375f2691e21b26" exitCode=1 Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.211513 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-56bcr" event={"ID":"a9efeabb-bf1c-4de2-8322-961b284e95cc","Type":"ContainerDied","Data":"addcdee7c623c641d2c7c579c327f9f105ad3a6312c9860f69375f2691e21b26"} Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.211557 4890 scope.go:117] "RemoveContainer" containerID="09913842764d83958bb7259fb1a01e61d3da9512eb07f23da446059eb49a046a" Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.212810 4890 scope.go:117] "RemoveContainer" containerID="addcdee7c623c641d2c7c579c327f9f105ad3a6312c9860f69375f2691e21b26" Nov 25 15:26:07 crc kubenswrapper[4890]: E1125 15:26:07.213184 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=octavia-operator-controller-manager-fd75fd47d-56bcr_openstack-operators(a9efeabb-bf1c-4de2-8322-961b284e95cc)\"" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-56bcr" podUID="a9efeabb-bf1c-4de2-8322-961b284e95cc" Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.215614 4890 generic.go:334] "Generic (PLEG): container finished" podID="77a5e8f9-92ab-48b5-be19-29a7a8e5df49" containerID="a18bbc200fff39fe093ad099ce5322f28fa7e5a281377e8a1b83796f88c742cc" exitCode=1 Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.215671 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-stwjm" event={"ID":"77a5e8f9-92ab-48b5-be19-29a7a8e5df49","Type":"ContainerDied","Data":"a18bbc200fff39fe093ad099ce5322f28fa7e5a281377e8a1b83796f88c742cc"} Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.216604 4890 scope.go:117] "RemoveContainer" containerID="a18bbc200fff39fe093ad099ce5322f28fa7e5a281377e8a1b83796f88c742cc" Nov 25 15:26:07 crc kubenswrapper[4890]: E1125 15:26:07.216885 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=designate-operator-controller-manager-7d695c9b56-stwjm_openstack-operators(77a5e8f9-92ab-48b5-be19-29a7a8e5df49)\"" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-stwjm" podUID="77a5e8f9-92ab-48b5-be19-29a7a8e5df49" Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.219635 4890 scope.go:117] "RemoveContainer" containerID="2545dda572d458934e753cbbe89f8cbd4188e0160a44a1d5caf399b84c9fb657" Nov 25 15:26:07 crc kubenswrapper[4890]: E1125 15:26:07.219927 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=telemetry-operator-controller-manager-567f98c9d-6zr4x_openstack-operators(d50aec9d-1673-4b81-bb57-04cae418843e)\"" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-6zr4x" podUID="d50aec9d-1673-4b81-bb57-04cae418843e" Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.220334 4890 scope.go:117] "RemoveContainer" containerID="8453cefea24580a864e1b962ed8a682423dc9aaafad812309969556ac975663c" Nov 25 15:26:07 crc kubenswrapper[4890]: E1125 15:26:07.220517 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=neutron-operator-controller-manager-7c57c8bbc4-829nk_openstack-operators(4364eec5-444a-41f6-af1c-fe16b24bd044)\"" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-829nk" podUID="4364eec5-444a-41f6-af1c-fe16b24bd044" Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.220784 4890 scope.go:117] "RemoveContainer" containerID="09ff68e9f7b5d7b971104be89b55602edbba51c72b2e776a35880be99ec6b53f" Nov 25 15:26:07 crc kubenswrapper[4890]: E1125 15:26:07.220945 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=mariadb-operator-controller-manager-cb6c4fdb7-wvjw7_openstack-operators(03353c8b-91f4-41ae-afc1-c522b1c73bb4)\"" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-wvjw7" podUID="03353c8b-91f4-41ae-afc1-c522b1c73bb4" Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.221224 4890 scope.go:117] "RemoveContainer" containerID="5085d0d9569ced239129365e8bcff54f0f41fba51b79ac5ddca34271333273d9" Nov 25 15:26:07 crc kubenswrapper[4890]: E1125 15:26:07.221421 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=heat-operator-controller-manager-774b86978c-xvfgd_openstack-operators(72c26c5b-3370-4994-a5f0-82128c62c2df)\"" pod="openstack-operators/heat-operator-controller-manager-774b86978c-xvfgd" podUID="72c26c5b-3370-4994-a5f0-82128c62c2df" Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.221686 4890 scope.go:117] "RemoveContainer" containerID="afdecad3ee6dca827e01ecfd5110e60bbd29e083f201918f083e2ccc5e1ba964" Nov 25 15:26:07 crc kubenswrapper[4890]: E1125 15:26:07.221870 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=nova-operator-controller-manager-79556f57fc-6jq7q_openstack-operators(0769cde4-b49f-41c5-b720-e0dbc667eb4a)\"" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6jq7q" podUID="0769cde4-b49f-41c5-b720-e0dbc667eb4a" Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.222146 4890 scope.go:117] "RemoveContainer" containerID="5d93b4988f6bdbdc8c47c62e1f25b2fcb1a65f5d3e949121c19272784f723e37" Nov 25 15:26:07 crc kubenswrapper[4890]: E1125 15:26:07.222333 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=ovn-operator-controller-manager-66cf5c67ff-vx75t_openstack-operators(29f250db-2a96-41ad-bb8f-5b503a9288d8)\"" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-vx75t" podUID="29f250db-2a96-41ad-bb8f-5b503a9288d8" Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.222677 4890 scope.go:117] "RemoveContainer" containerID="59bdc927190ff946a016c3b38ef86d4b991001a88772e7a4c7e5cc7cbf93b8d3" Nov 25 15:26:07 crc kubenswrapper[4890]: E1125 15:26:07.222861 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=keystone-operator-controller-manager-748dc6576f-rs6hx_openstack-operators(a5d8a8ba-50c0-4fb6-823f-f28cf0d0f125)\"" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-rs6hx" podUID="a5d8a8ba-50c0-4fb6-823f-f28cf0d0f125" Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.223137 4890 scope.go:117] "RemoveContainer" containerID="7eed3f7f71023e9126d0de61d864c2912c3fd5e6e53615d8e82e9bb5d98d5081" Nov 25 15:26:07 crc kubenswrapper[4890]: E1125 15:26:07.223324 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=ironic-operator-controller-manager-5bfcdc958c-hxzf4_openstack-operators(679b3bf3-1ebc-4d19-9ad2-588bd41565dd)\"" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-hxzf4" podUID="679b3bf3-1ebc-4d19-9ad2-588bd41565dd" Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.223616 4890 scope.go:117] "RemoveContainer" containerID="aabb0ea70edd64ac6611bc788292aa0806bdccd21d77f87d39aa68b59663cf81" Nov 25 15:26:07 crc kubenswrapper[4890]: E1125 15:26:07.223817 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=swift-operator-controller-manager-6fdc4fcf86-dzkgt_openstack-operators(861317e1-222d-4f14-b931-e9e7d100ebd6)\"" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-dzkgt" podUID="861317e1-222d-4f14-b931-e9e7d100ebd6" Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.229729 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.238755 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.239051 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.243116 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.253733 4890 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.253799 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.253852 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.254728 4890 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-controller-manager" containerStatusID={"Type":"cri-o","ID":"b68b7a22193a1042f7b2a06df452033ab04e929a3b8f18b668593749e140c576"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container kube-controller-manager failed startup probe, will be restarted" Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.254859 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" containerID="cri-o://b68b7a22193a1042f7b2a06df452033ab04e929a3b8f18b668593749e140c576" gracePeriod=30 Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.260769 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.277484 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.349396 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.380314 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.411303 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.437080 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-xzflj" Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.486674 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.488409 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.614889 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.630823 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.640118 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.654519 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.677055 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.698910 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.732958 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-stflb" Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.733010 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-stflb" Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.733718 4890 scope.go:117] "RemoveContainer" containerID="afe47ed941c0114eed272202844f29373d2d8c6c4f5c70cf119a83c56005a08d" Nov 25 15:26:07 crc kubenswrapper[4890]: E1125 15:26:07.734077 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=infra-operator-controller-manager-d5cc86f4b-stflb_openstack-operators(7e576602-19f5-4656-84dd-5ab80abca895)\"" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-stflb" podUID="7e576602-19f5-4656-84dd-5ab80abca895" Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.764994 4890 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.793060 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.879922 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.886694 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.894765 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.896952 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.904491 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 25 15:26:07 crc kubenswrapper[4890]: I1125 15:26:07.976022 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 25 15:26:08 crc kubenswrapper[4890]: I1125 15:26:08.017504 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 25 15:26:08 crc kubenswrapper[4890]: I1125 15:26:08.083449 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 25 15:26:08 crc kubenswrapper[4890]: I1125 15:26:08.126782 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Nov 25 15:26:08 crc kubenswrapper[4890]: I1125 15:26:08.132786 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Nov 25 15:26:08 crc kubenswrapper[4890]: I1125 15:26:08.157474 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-8zbpv" Nov 25 15:26:08 crc kubenswrapper[4890]: I1125 15:26:08.224811 4890 scope.go:117] "RemoveContainer" containerID="e7b7fcdef297e60c89760ed3c86c2d93e8c7a6a0088d065d1090410732523bb5" Nov 25 15:26:08 crc kubenswrapper[4890]: I1125 15:26:08.244496 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-864885998-62ddz" event={"ID":"a9a69828-3f0d-4d55-a840-10c2c791444c","Type":"ContainerStarted","Data":"622bd013e72e58f2f4be193e9c8ece024e704c0c53f100d1086ed1f91444ddb8"} Nov 25 15:26:08 crc kubenswrapper[4890]: I1125 15:26:08.300747 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 25 15:26:08 crc kubenswrapper[4890]: I1125 15:26:08.376386 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 25 15:26:08 crc kubenswrapper[4890]: I1125 15:26:08.378832 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 25 15:26:08 crc kubenswrapper[4890]: I1125 15:26:08.404103 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-xngkr" Nov 25 15:26:08 crc kubenswrapper[4890]: I1125 15:26:08.501234 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 25 15:26:08 crc kubenswrapper[4890]: I1125 15:26:08.582647 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Nov 25 15:26:08 crc kubenswrapper[4890]: I1125 15:26:08.600748 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 25 15:26:08 crc kubenswrapper[4890]: I1125 15:26:08.664980 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 25 15:26:08 crc kubenswrapper[4890]: I1125 15:26:08.738517 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 25 15:26:08 crc kubenswrapper[4890]: I1125 15:26:08.746596 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 25 15:26:08 crc kubenswrapper[4890]: I1125 15:26:08.770469 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 25 15:26:08 crc kubenswrapper[4890]: I1125 15:26:08.845988 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 25 15:26:08 crc kubenswrapper[4890]: I1125 15:26:08.850488 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 25 15:26:08 crc kubenswrapper[4890]: I1125 15:26:08.868401 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 25 15:26:08 crc kubenswrapper[4890]: I1125 15:26:08.920264 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 25 15:26:08 crc kubenswrapper[4890]: I1125 15:26:08.950273 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 25 15:26:09 crc kubenswrapper[4890]: I1125 15:26:09.048782 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 25 15:26:09 crc kubenswrapper[4890]: I1125 15:26:09.127855 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 25 15:26:09 crc kubenswrapper[4890]: I1125 15:26:09.132532 4890 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-lv5hp" Nov 25 15:26:09 crc kubenswrapper[4890]: I1125 15:26:09.164357 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Nov 25 15:26:09 crc kubenswrapper[4890]: I1125 15:26:09.184877 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 25 15:26:09 crc kubenswrapper[4890]: I1125 15:26:09.206589 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-xj5wg" Nov 25 15:26:09 crc kubenswrapper[4890]: I1125 15:26:09.240105 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 25 15:26:09 crc kubenswrapper[4890]: I1125 15:26:09.256819 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-864885998-62ddz" Nov 25 15:26:09 crc kubenswrapper[4890]: I1125 15:26:09.321578 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 25 15:26:09 crc kubenswrapper[4890]: I1125 15:26:09.359224 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 25 15:26:09 crc kubenswrapper[4890]: I1125 15:26:09.374721 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 25 15:26:09 crc kubenswrapper[4890]: I1125 15:26:09.461654 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 25 15:26:09 crc kubenswrapper[4890]: I1125 15:26:09.608106 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 25 15:26:09 crc kubenswrapper[4890]: I1125 15:26:09.642332 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Nov 25 15:26:09 crc kubenswrapper[4890]: I1125 15:26:09.706078 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 25 15:26:09 crc kubenswrapper[4890]: I1125 15:26:09.726697 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Nov 25 15:26:09 crc kubenswrapper[4890]: I1125 15:26:09.901259 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 25 15:26:09 crc kubenswrapper[4890]: I1125 15:26:09.915385 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 25 15:26:10 crc kubenswrapper[4890]: I1125 15:26:10.389859 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 25 15:26:10 crc kubenswrapper[4890]: I1125 15:26:10.401317 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Nov 25 15:26:10 crc kubenswrapper[4890]: I1125 15:26:10.406565 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-94xds" Nov 25 15:26:10 crc kubenswrapper[4890]: I1125 15:26:10.552104 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 25 15:26:10 crc kubenswrapper[4890]: I1125 15:26:10.664096 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/openstack-operator-controller-manager-68587559f4-9b4cf" Nov 25 15:26:10 crc kubenswrapper[4890]: I1125 15:26:10.664876 4890 scope.go:117] "RemoveContainer" containerID="aeb0938dec764a3788ac5a6992985db59cc25e8a9f8cd69f3f2408a7e27cdd24" Nov 25 15:26:10 crc kubenswrapper[4890]: E1125 15:26:10.665123 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=openstack-operator-controller-manager-68587559f4-9b4cf_openstack-operators(3e496c3d-9485-44bc-b3b1-dc1682a88434)\"" pod="openstack-operators/openstack-operator-controller-manager-68587559f4-9b4cf" podUID="3e496c3d-9485-44bc-b3b1-dc1682a88434" Nov 25 15:26:10 crc kubenswrapper[4890]: I1125 15:26:10.884070 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 25 15:26:10 crc kubenswrapper[4890]: I1125 15:26:10.943806 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 25 15:26:10 crc kubenswrapper[4890]: I1125 15:26:10.968546 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 25 15:26:11 crc kubenswrapper[4890]: I1125 15:26:11.035865 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 25 15:26:11 crc kubenswrapper[4890]: I1125 15:26:11.274461 4890 generic.go:334] "Generic (PLEG): container finished" podID="e14d1071-6d9a-4543-b593-8e999b7dcad2" containerID="b8f2d09a82996fa87b1da07726363e7fa38591a8f4f4bdd13b0fdbbb73bb5c28" exitCode=1 Nov 25 15:26:11 crc kubenswrapper[4890]: I1125 15:26:11.274551 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-9c9wk" event={"ID":"e14d1071-6d9a-4543-b593-8e999b7dcad2","Type":"ContainerDied","Data":"b8f2d09a82996fa87b1da07726363e7fa38591a8f4f4bdd13b0fdbbb73bb5c28"} Nov 25 15:26:11 crc kubenswrapper[4890]: I1125 15:26:11.275923 4890 scope.go:117] "RemoveContainer" containerID="b8f2d09a82996fa87b1da07726363e7fa38591a8f4f4bdd13b0fdbbb73bb5c28" Nov 25 15:26:11 crc kubenswrapper[4890]: I1125 15:26:11.276605 4890 generic.go:334] "Generic (PLEG): container finished" podID="d0f0970d-430a-4ea0-b18c-6af23eb40c71" containerID="91bc437d2b50db9aa97f6cabfe86828900be0cbc84357d63d8f0d1f5d780d628" exitCode=1 Nov 25 15:26:11 crc kubenswrapper[4890]: I1125 15:26:11.276682 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-x7dwr" event={"ID":"d0f0970d-430a-4ea0-b18c-6af23eb40c71","Type":"ContainerDied","Data":"91bc437d2b50db9aa97f6cabfe86828900be0cbc84357d63d8f0d1f5d780d628"} Nov 25 15:26:11 crc kubenswrapper[4890]: I1125 15:26:11.277305 4890 scope.go:117] "RemoveContainer" containerID="91bc437d2b50db9aa97f6cabfe86828900be0cbc84357d63d8f0d1f5d780d628" Nov 25 15:26:11 crc kubenswrapper[4890]: I1125 15:26:11.280074 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 25 15:26:11 crc kubenswrapper[4890]: I1125 15:26:11.280120 4890 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="f8010cc46ab44d816173f84c66043f945180f00900ca208dc599fb56bcf061b2" exitCode=137 Nov 25 15:26:11 crc kubenswrapper[4890]: I1125 15:26:11.280146 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="216dc27f3c7236a012847fb5349779c9c44f1b480be3dc3c4f0ae8086ab78f23" Nov 25 15:26:11 crc kubenswrapper[4890]: I1125 15:26:11.328843 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 25 15:26:11 crc kubenswrapper[4890]: I1125 15:26:11.328925 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 15:26:11 crc kubenswrapper[4890]: I1125 15:26:11.442807 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 25 15:26:11 crc kubenswrapper[4890]: I1125 15:26:11.442910 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 15:26:11 crc kubenswrapper[4890]: I1125 15:26:11.443053 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 25 15:26:11 crc kubenswrapper[4890]: I1125 15:26:11.443115 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 25 15:26:11 crc kubenswrapper[4890]: I1125 15:26:11.443147 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 25 15:26:11 crc kubenswrapper[4890]: I1125 15:26:11.443149 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 15:26:11 crc kubenswrapper[4890]: I1125 15:26:11.443184 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 15:26:11 crc kubenswrapper[4890]: I1125 15:26:11.443193 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 25 15:26:11 crc kubenswrapper[4890]: I1125 15:26:11.443217 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 15:26:11 crc kubenswrapper[4890]: I1125 15:26:11.444017 4890 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Nov 25 15:26:11 crc kubenswrapper[4890]: I1125 15:26:11.444513 4890 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Nov 25 15:26:11 crc kubenswrapper[4890]: I1125 15:26:11.444528 4890 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 25 15:26:11 crc kubenswrapper[4890]: I1125 15:26:11.444543 4890 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Nov 25 15:26:11 crc kubenswrapper[4890]: I1125 15:26:11.449935 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 15:26:11 crc kubenswrapper[4890]: I1125 15:26:11.546202 4890 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 25 15:26:12 crc kubenswrapper[4890]: I1125 15:26:12.187444 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Nov 25 15:26:12 crc kubenswrapper[4890]: I1125 15:26:12.306675 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 15:26:12 crc kubenswrapper[4890]: I1125 15:26:12.307222 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-9c9wk" event={"ID":"e14d1071-6d9a-4543-b593-8e999b7dcad2","Type":"ContainerStarted","Data":"409a5ccdaeb3d3136014fc5aefe813e4327d87725fadba67070391ee28129ffe"} Nov 25 15:26:13 crc kubenswrapper[4890]: I1125 15:26:13.316896 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-x7dwr" event={"ID":"d0f0970d-430a-4ea0-b18c-6af23eb40c71","Type":"ContainerStarted","Data":"66a33405571cc4466164c57b8a6da0e71158deebe2b83df879cf229a66052133"} Nov 25 15:26:13 crc kubenswrapper[4890]: I1125 15:26:13.887126 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gh9mr"] Nov 25 15:26:13 crc kubenswrapper[4890]: E1125 15:26:13.887653 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 25 15:26:13 crc kubenswrapper[4890]: I1125 15:26:13.887673 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 25 15:26:13 crc kubenswrapper[4890]: E1125 15:26:13.887691 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="118dcc72-87d4-4aa3-9397-bee2c90eac63" containerName="installer" Nov 25 15:26:13 crc kubenswrapper[4890]: I1125 15:26:13.887699 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="118dcc72-87d4-4aa3-9397-bee2c90eac63" containerName="installer" Nov 25 15:26:13 crc kubenswrapper[4890]: I1125 15:26:13.887895 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 25 15:26:13 crc kubenswrapper[4890]: I1125 15:26:13.887932 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="118dcc72-87d4-4aa3-9397-bee2c90eac63" containerName="installer" Nov 25 15:26:13 crc kubenswrapper[4890]: I1125 15:26:13.889305 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gh9mr" Nov 25 15:26:13 crc kubenswrapper[4890]: I1125 15:26:13.898059 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gh9mr"] Nov 25 15:26:13 crc kubenswrapper[4890]: I1125 15:26:13.991168 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c9a571e-db77-4303-96cc-c111add7626d-catalog-content\") pod \"community-operators-gh9mr\" (UID: \"1c9a571e-db77-4303-96cc-c111add7626d\") " pod="openshift-marketplace/community-operators-gh9mr" Nov 25 15:26:13 crc kubenswrapper[4890]: I1125 15:26:13.991362 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c9a571e-db77-4303-96cc-c111add7626d-utilities\") pod \"community-operators-gh9mr\" (UID: \"1c9a571e-db77-4303-96cc-c111add7626d\") " pod="openshift-marketplace/community-operators-gh9mr" Nov 25 15:26:13 crc kubenswrapper[4890]: I1125 15:26:13.991431 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76jx8\" (UniqueName: \"kubernetes.io/projected/1c9a571e-db77-4303-96cc-c111add7626d-kube-api-access-76jx8\") pod \"community-operators-gh9mr\" (UID: \"1c9a571e-db77-4303-96cc-c111add7626d\") " pod="openshift-marketplace/community-operators-gh9mr" Nov 25 15:26:14 crc kubenswrapper[4890]: I1125 15:26:14.093746 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c9a571e-db77-4303-96cc-c111add7626d-catalog-content\") pod \"community-operators-gh9mr\" (UID: \"1c9a571e-db77-4303-96cc-c111add7626d\") " pod="openshift-marketplace/community-operators-gh9mr" Nov 25 15:26:14 crc kubenswrapper[4890]: I1125 15:26:14.093817 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c9a571e-db77-4303-96cc-c111add7626d-utilities\") pod \"community-operators-gh9mr\" (UID: \"1c9a571e-db77-4303-96cc-c111add7626d\") " pod="openshift-marketplace/community-operators-gh9mr" Nov 25 15:26:14 crc kubenswrapper[4890]: I1125 15:26:14.093844 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76jx8\" (UniqueName: \"kubernetes.io/projected/1c9a571e-db77-4303-96cc-c111add7626d-kube-api-access-76jx8\") pod \"community-operators-gh9mr\" (UID: \"1c9a571e-db77-4303-96cc-c111add7626d\") " pod="openshift-marketplace/community-operators-gh9mr" Nov 25 15:26:14 crc kubenswrapper[4890]: I1125 15:26:14.094310 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c9a571e-db77-4303-96cc-c111add7626d-catalog-content\") pod \"community-operators-gh9mr\" (UID: \"1c9a571e-db77-4303-96cc-c111add7626d\") " pod="openshift-marketplace/community-operators-gh9mr" Nov 25 15:26:14 crc kubenswrapper[4890]: I1125 15:26:14.094446 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c9a571e-db77-4303-96cc-c111add7626d-utilities\") pod \"community-operators-gh9mr\" (UID: \"1c9a571e-db77-4303-96cc-c111add7626d\") " pod="openshift-marketplace/community-operators-gh9mr" Nov 25 15:26:14 crc kubenswrapper[4890]: I1125 15:26:14.121204 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76jx8\" (UniqueName: \"kubernetes.io/projected/1c9a571e-db77-4303-96cc-c111add7626d-kube-api-access-76jx8\") pod \"community-operators-gh9mr\" (UID: \"1c9a571e-db77-4303-96cc-c111add7626d\") " pod="openshift-marketplace/community-operators-gh9mr" Nov 25 15:26:14 crc kubenswrapper[4890]: I1125 15:26:14.247655 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gh9mr" Nov 25 15:26:14 crc kubenswrapper[4890]: I1125 15:26:14.699089 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gh9mr"] Nov 25 15:26:15 crc kubenswrapper[4890]: I1125 15:26:15.172503 4890 scope.go:117] "RemoveContainer" containerID="137b37bbfac0ae56e61f31ca811e8cf420a7182f7941ed967187fc1ed8ac86f0" Nov 25 15:26:15 crc kubenswrapper[4890]: E1125 15:26:15.172827 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=metallb-operator-controller-manager-66fc8c7fbf-lnjxm_metallb-system(dad943fe-2209-40f9-a9bb-0be7ba80cca2)\"" pod="metallb-system/metallb-operator-controller-manager-66fc8c7fbf-lnjxm" podUID="dad943fe-2209-40f9-a9bb-0be7ba80cca2" Nov 25 15:26:15 crc kubenswrapper[4890]: I1125 15:26:15.338071 4890 generic.go:334] "Generic (PLEG): container finished" podID="1c9a571e-db77-4303-96cc-c111add7626d" containerID="8c20e9b751cfadd82d3cf4b5f40aaf1d3b45c26081dd57e7fcf8fe32e1140f99" exitCode=0 Nov 25 15:26:15 crc kubenswrapper[4890]: I1125 15:26:15.338129 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gh9mr" event={"ID":"1c9a571e-db77-4303-96cc-c111add7626d","Type":"ContainerDied","Data":"8c20e9b751cfadd82d3cf4b5f40aaf1d3b45c26081dd57e7fcf8fe32e1140f99"} Nov 25 15:26:15 crc kubenswrapper[4890]: I1125 15:26:15.338430 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gh9mr" event={"ID":"1c9a571e-db77-4303-96cc-c111add7626d","Type":"ContainerStarted","Data":"88fe258dfd5b32f740f82794e6235d890a94e9f5a650318106758fa7930b9a8b"} Nov 25 15:26:16 crc kubenswrapper[4890]: I1125 15:26:16.127881 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-stwjm" Nov 25 15:26:16 crc kubenswrapper[4890]: I1125 15:26:16.128674 4890 scope.go:117] "RemoveContainer" containerID="a18bbc200fff39fe093ad099ce5322f28fa7e5a281377e8a1b83796f88c742cc" Nov 25 15:26:16 crc kubenswrapper[4890]: E1125 15:26:16.128928 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=designate-operator-controller-manager-7d695c9b56-stwjm_openstack-operators(77a5e8f9-92ab-48b5-be19-29a7a8e5df49)\"" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-stwjm" podUID="77a5e8f9-92ab-48b5-be19-29a7a8e5df49" Nov 25 15:26:16 crc kubenswrapper[4890]: I1125 15:26:16.648959 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-56bcr" Nov 25 15:26:16 crc kubenswrapper[4890]: I1125 15:26:16.649809 4890 scope.go:117] "RemoveContainer" containerID="addcdee7c623c641d2c7c579c327f9f105ad3a6312c9860f69375f2691e21b26" Nov 25 15:26:16 crc kubenswrapper[4890]: E1125 15:26:16.650274 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=octavia-operator-controller-manager-fd75fd47d-56bcr_openstack-operators(a9efeabb-bf1c-4de2-8322-961b284e95cc)\"" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-56bcr" podUID="a9efeabb-bf1c-4de2-8322-961b284e95cc" Nov 25 15:26:16 crc kubenswrapper[4890]: I1125 15:26:16.984473 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-864885998-62ddz" Nov 25 15:26:17 crc kubenswrapper[4890]: I1125 15:26:17.173510 4890 scope.go:117] "RemoveContainer" containerID="57618fb9802f27a586bce5ed3b16f89465e88061eac4872f5ae34829c7d008e2" Nov 25 15:26:17 crc kubenswrapper[4890]: I1125 15:26:17.173649 4890 scope.go:117] "RemoveContainer" containerID="8052599ccdad737d107b842db95dca88d95c47c4640b7dfd30857525688810ee" Nov 25 15:26:17 crc kubenswrapper[4890]: E1125 15:26:17.173916 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=glance-operator-controller-manager-68b95954c9-lfnwl_openstack-operators(076c2a0e-7fdb-42e1-ab73-e296bba2b7a9)\"" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-lfnwl" podUID="076c2a0e-7fdb-42e1-ab73-e296bba2b7a9" Nov 25 15:26:18 crc kubenswrapper[4890]: I1125 15:26:18.173388 4890 scope.go:117] "RemoveContainer" containerID="f0d0bf516dcd47d736a52b93852c7285f0e58e2e35a41f80221eda17385548d1" Nov 25 15:26:18 crc kubenswrapper[4890]: E1125 15:26:18.173738 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=manila-operator-controller-manager-58bb8d67cc-slpt4_openstack-operators(f272410f-230d-43f2-946e-e34e4d9c05ea)\"" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-slpt4" podUID="f272410f-230d-43f2-946e-e34e4d9c05ea" Nov 25 15:26:18 crc kubenswrapper[4890]: I1125 15:26:18.174439 4890 scope.go:117] "RemoveContainer" containerID="c186dc6ff6591b7b58925577cec707c6ed3796c30a9ae21bb9565049cea41856" Nov 25 15:26:18 crc kubenswrapper[4890]: E1125 15:26:18.174836 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with CrashLoopBackOff: \"back-off 20s restarting failed container=operator pod=rabbitmq-cluster-operator-manager-668c99d594-dmlx4_openstack-operators(e26594ec-4bc2-4051-8a49-fc490bd3b6c9)\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-dmlx4" podUID="e26594ec-4bc2-4051-8a49-fc490bd3b6c9" Nov 25 15:26:19 crc kubenswrapper[4890]: I1125 15:26:19.173746 4890 scope.go:117] "RemoveContainer" containerID="59bdc927190ff946a016c3b38ef86d4b991001a88772e7a4c7e5cc7cbf93b8d3" Nov 25 15:26:19 crc kubenswrapper[4890]: I1125 15:26:19.174135 4890 scope.go:117] "RemoveContainer" containerID="5085d0d9569ced239129365e8bcff54f0f41fba51b79ac5ddca34271333273d9" Nov 25 15:26:20 crc kubenswrapper[4890]: I1125 15:26:20.173405 4890 scope.go:117] "RemoveContainer" containerID="ea03217bf38df8f664326633d6bc9735c9f7d24feebcf0fb7ac30106bfefa2d0" Nov 25 15:26:20 crc kubenswrapper[4890]: I1125 15:26:20.173589 4890 scope.go:117] "RemoveContainer" containerID="24f5582ca11aa9c0ac5997208248ab279725e93ef77197c94bd72f340f8bbab7" Nov 25 15:26:20 crc kubenswrapper[4890]: I1125 15:26:20.173722 4890 scope.go:117] "RemoveContainer" containerID="7eed3f7f71023e9126d0de61d864c2912c3fd5e6e53615d8e82e9bb5d98d5081" Nov 25 15:26:20 crc kubenswrapper[4890]: I1125 15:26:20.174271 4890 scope.go:117] "RemoveContainer" containerID="afdecad3ee6dca827e01ecfd5110e60bbd29e083f201918f083e2ccc5e1ba964" Nov 25 15:26:21 crc kubenswrapper[4890]: I1125 15:26:21.173936 4890 scope.go:117] "RemoveContainer" containerID="aabb0ea70edd64ac6611bc788292aa0806bdccd21d77f87d39aa68b59663cf81" Nov 25 15:26:21 crc kubenswrapper[4890]: I1125 15:26:21.175068 4890 scope.go:117] "RemoveContainer" containerID="a0074090c38ba41c6cd974ca26e67cc1c9cd2d25edd77a01b7f4e31a93593803" Nov 25 15:26:21 crc kubenswrapper[4890]: I1125 15:26:21.175200 4890 scope.go:117] "RemoveContainer" containerID="dc8aff712070f60aaf997f341250476ab00d06875bc86901bf5c918d9c5a94d9" Nov 25 15:26:21 crc kubenswrapper[4890]: E1125 15:26:21.175331 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=placement-operator-controller-manager-5db546f9d9-nshw7_openstack-operators(9f0e3cf9-dcc0-405f-a2de-9148844ec3ae)\"" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-nshw7" podUID="9f0e3cf9-dcc0-405f-a2de-9148844ec3ae" Nov 25 15:26:21 crc kubenswrapper[4890]: I1125 15:26:21.175348 4890 scope.go:117] "RemoveContainer" containerID="2545dda572d458934e753cbbe89f8cbd4188e0160a44a1d5caf399b84c9fb657" Nov 25 15:26:21 crc kubenswrapper[4890]: I1125 15:26:21.175804 4890 scope.go:117] "RemoveContainer" containerID="09ff68e9f7b5d7b971104be89b55602edbba51c72b2e776a35880be99ec6b53f" Nov 25 15:26:21 crc kubenswrapper[4890]: E1125 15:26:21.175984 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=mariadb-operator-controller-manager-cb6c4fdb7-wvjw7_openstack-operators(03353c8b-91f4-41ae-afc1-c522b1c73bb4)\"" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-wvjw7" podUID="03353c8b-91f4-41ae-afc1-c522b1c73bb4" Nov 25 15:26:22 crc kubenswrapper[4890]: I1125 15:26:22.182729 4890 scope.go:117] "RemoveContainer" containerID="afe47ed941c0114eed272202844f29373d2d8c6c4f5c70cf119a83c56005a08d" Nov 25 15:26:22 crc kubenswrapper[4890]: E1125 15:26:22.183196 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=infra-operator-controller-manager-d5cc86f4b-stflb_openstack-operators(7e576602-19f5-4656-84dd-5ab80abca895)\"" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-stflb" podUID="7e576602-19f5-4656-84dd-5ab80abca895" Nov 25 15:26:22 crc kubenswrapper[4890]: I1125 15:26:22.183650 4890 scope.go:117] "RemoveContainer" containerID="8453cefea24580a864e1b962ed8a682423dc9aaafad812309969556ac975663c" Nov 25 15:26:22 crc kubenswrapper[4890]: I1125 15:26:22.183809 4890 scope.go:117] "RemoveContainer" containerID="5d93b4988f6bdbdc8c47c62e1f25b2fcb1a65f5d3e949121c19272784f723e37" Nov 25 15:26:22 crc kubenswrapper[4890]: E1125 15:26:22.183961 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=neutron-operator-controller-manager-7c57c8bbc4-829nk_openstack-operators(4364eec5-444a-41f6-af1c-fe16b24bd044)\"" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-829nk" podUID="4364eec5-444a-41f6-af1c-fe16b24bd044" Nov 25 15:26:22 crc kubenswrapper[4890]: E1125 15:26:22.184063 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=ovn-operator-controller-manager-66cf5c67ff-vx75t_openstack-operators(29f250db-2a96-41ad-bb8f-5b503a9288d8)\"" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-vx75t" podUID="29f250db-2a96-41ad-bb8f-5b503a9288d8" Nov 25 15:26:23 crc kubenswrapper[4890]: I1125 15:26:23.175753 4890 scope.go:117] "RemoveContainer" containerID="aeb0938dec764a3788ac5a6992985db59cc25e8a9f8cd69f3f2408a7e27cdd24" Nov 25 15:26:23 crc kubenswrapper[4890]: E1125 15:26:23.176524 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=openstack-operator-controller-manager-68587559f4-9b4cf_openstack-operators(3e496c3d-9485-44bc-b3b1-dc1682a88434)\"" pod="openstack-operators/openstack-operator-controller-manager-68587559f4-9b4cf" podUID="3e496c3d-9485-44bc-b3b1-dc1682a88434" Nov 25 15:26:23 crc kubenswrapper[4890]: I1125 15:26:23.419826 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-hxzf4" event={"ID":"679b3bf3-1ebc-4d19-9ad2-588bd41565dd","Type":"ContainerStarted","Data":"30c29a88d4d305ddd8058edad08b87a432f763ae4946cfafd9734c6d4c8265eb"} Nov 25 15:26:24 crc kubenswrapper[4890]: I1125 15:26:24.431199 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-j6pqs" event={"ID":"c528348f-0907-423a-a0c2-f15cf18ff99a","Type":"ContainerStarted","Data":"c0a6b58647c04df36505de9c5f2cacae65fb2cd5d9c3859e31c7c4ea80f6829f"} Nov 25 15:26:24 crc kubenswrapper[4890]: I1125 15:26:24.431412 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-j6pqs" Nov 25 15:26:24 crc kubenswrapper[4890]: I1125 15:26:24.434408 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-pmmf2" event={"ID":"d175a35b-8b9b-49c0-8fc2-02ef62f131d9","Type":"ContainerStarted","Data":"a2eefc2d3dfaa01562b20b8386d3f8ae0982be0a82da400c47e4d014eb0269f1"} Nov 25 15:26:24 crc kubenswrapper[4890]: I1125 15:26:24.434904 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-pmmf2" Nov 25 15:26:24 crc kubenswrapper[4890]: I1125 15:26:24.436775 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-774b86978c-xvfgd" event={"ID":"72c26c5b-3370-4994-a5f0-82128c62c2df","Type":"ContainerStarted","Data":"1d4de5017f32f32f1ffbf0c36be4672d476bc8be8ccd73c3ece15d3e73762356"} Nov 25 15:26:24 crc kubenswrapper[4890]: I1125 15:26:24.437002 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-774b86978c-xvfgd" Nov 25 15:26:24 crc kubenswrapper[4890]: I1125 15:26:24.438926 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-6zr4x" event={"ID":"d50aec9d-1673-4b81-bb57-04cae418843e","Type":"ContainerStarted","Data":"c1a28710df3d62d779ce4dd6c92827bde5071caa4977becabf952d2eb88adc67"} Nov 25 15:26:24 crc kubenswrapper[4890]: I1125 15:26:24.439139 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-6zr4x" Nov 25 15:26:24 crc kubenswrapper[4890]: I1125 15:26:24.441074 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"a47203e9-a88a-42f7-86e1-164c6ad532b5","Type":"ContainerStarted","Data":"a98489ff547fdc2009e9da7cd1bba6e921f7c623994135bbe3cac44fdf7a6c56"} Nov 25 15:26:24 crc kubenswrapper[4890]: I1125 15:26:24.441323 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 25 15:26:24 crc kubenswrapper[4890]: I1125 15:26:24.443282 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-jcw7x" event={"ID":"0bdbada3-e08c-4f8a-bd55-bac955d4370d","Type":"ContainerStarted","Data":"759612cba4397005ef692e943350cb6a75e94c12ce67e3596d582fd04e0080a1"} Nov 25 15:26:24 crc kubenswrapper[4890]: I1125 15:26:24.443463 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-jcw7x" Nov 25 15:26:24 crc kubenswrapper[4890]: I1125 15:26:24.445916 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6jq7q" event={"ID":"0769cde4-b49f-41c5-b720-e0dbc667eb4a","Type":"ContainerStarted","Data":"ef9e65d4e41996a7c7fa6344a4cffa7a01d07f8f1696e60addb885b5f77564ff"} Nov 25 15:26:24 crc kubenswrapper[4890]: I1125 15:26:24.446114 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6jq7q" Nov 25 15:26:24 crc kubenswrapper[4890]: I1125 15:26:24.449294 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-rs6hx" event={"ID":"a5d8a8ba-50c0-4fb6-823f-f28cf0d0f125","Type":"ContainerStarted","Data":"fb3941f0b7e3281c4e6cc8d197db9bb60170e6801d7c224a85b1167bf8d3ec58"} Nov 25 15:26:24 crc kubenswrapper[4890]: I1125 15:26:24.450220 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-rs6hx" Nov 25 15:26:24 crc kubenswrapper[4890]: I1125 15:26:24.453342 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-dzkgt" event={"ID":"861317e1-222d-4f14-b931-e9e7d100ebd6","Type":"ContainerStarted","Data":"8ee8cd472d6719925e4bc57c06d0aa3ee12684399603226fc26b0c7cef99e7f1"} Nov 25 15:26:24 crc kubenswrapper[4890]: I1125 15:26:24.453418 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-hxzf4" Nov 25 15:26:24 crc kubenswrapper[4890]: I1125 15:26:24.453757 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-dzkgt" Nov 25 15:26:26 crc kubenswrapper[4890]: I1125 15:26:26.130793 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-stwjm" Nov 25 15:26:26 crc kubenswrapper[4890]: I1125 15:26:26.131315 4890 scope.go:117] "RemoveContainer" containerID="a18bbc200fff39fe093ad099ce5322f28fa7e5a281377e8a1b83796f88c742cc" Nov 25 15:26:26 crc kubenswrapper[4890]: E1125 15:26:26.131535 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=designate-operator-controller-manager-7d695c9b56-stwjm_openstack-operators(77a5e8f9-92ab-48b5-be19-29a7a8e5df49)\"" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-stwjm" podUID="77a5e8f9-92ab-48b5-be19-29a7a8e5df49" Nov 25 15:26:26 crc kubenswrapper[4890]: I1125 15:26:26.648487 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-56bcr" Nov 25 15:26:26 crc kubenswrapper[4890]: I1125 15:26:26.649718 4890 scope.go:117] "RemoveContainer" containerID="addcdee7c623c641d2c7c579c327f9f105ad3a6312c9860f69375f2691e21b26" Nov 25 15:26:26 crc kubenswrapper[4890]: E1125 15:26:26.650003 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=octavia-operator-controller-manager-fd75fd47d-56bcr_openstack-operators(a9efeabb-bf1c-4de2-8322-961b284e95cc)\"" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-56bcr" podUID="a9efeabb-bf1c-4de2-8322-961b284e95cc" Nov 25 15:26:27 crc kubenswrapper[4890]: I1125 15:26:27.484005 4890 generic.go:334] "Generic (PLEG): container finished" podID="1c9a571e-db77-4303-96cc-c111add7626d" containerID="bd5877a3900425fe22977396b122dbfc9504e980f0f73e0be858e810a14aab51" exitCode=0 Nov 25 15:26:27 crc kubenswrapper[4890]: I1125 15:26:27.484195 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gh9mr" event={"ID":"1c9a571e-db77-4303-96cc-c111add7626d","Type":"ContainerDied","Data":"bd5877a3900425fe22977396b122dbfc9504e980f0f73e0be858e810a14aab51"} Nov 25 15:26:28 crc kubenswrapper[4890]: I1125 15:26:28.174283 4890 scope.go:117] "RemoveContainer" containerID="137b37bbfac0ae56e61f31ca811e8cf420a7182f7941ed967187fc1ed8ac86f0" Nov 25 15:26:29 crc kubenswrapper[4890]: I1125 15:26:29.507546 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-66fc8c7fbf-lnjxm" event={"ID":"dad943fe-2209-40f9-a9bb-0be7ba80cca2","Type":"ContainerStarted","Data":"59100515b0d2e8157e3887d52b4ef671238ffc9d1cbc4b6d968ce26002868813"} Nov 25 15:26:29 crc kubenswrapper[4890]: I1125 15:26:29.507801 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-66fc8c7fbf-lnjxm" Nov 25 15:26:29 crc kubenswrapper[4890]: I1125 15:26:29.514708 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gh9mr" event={"ID":"1c9a571e-db77-4303-96cc-c111add7626d","Type":"ContainerStarted","Data":"d11854ae7e58b3739a5b1b33d4e46c1c1b48da778db1b7e3f005b331cb2434b6"} Nov 25 15:26:29 crc kubenswrapper[4890]: I1125 15:26:29.547915 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gh9mr" podStartSLOduration=3.504646204 podStartE2EDuration="16.547896618s" podCreationTimestamp="2025-11-25 15:26:13 +0000 UTC" firstStartedPulling="2025-11-25 15:26:15.340100545 +0000 UTC m=+1433.782563165" lastFinishedPulling="2025-11-25 15:26:28.383350969 +0000 UTC m=+1446.825813579" observedRunningTime="2025-11-25 15:26:29.540132084 +0000 UTC m=+1447.982594694" watchObservedRunningTime="2025-11-25 15:26:29.547896618 +0000 UTC m=+1447.990359228" Nov 25 15:26:30 crc kubenswrapper[4890]: I1125 15:26:30.173544 4890 scope.go:117] "RemoveContainer" containerID="f0d0bf516dcd47d736a52b93852c7285f0e58e2e35a41f80221eda17385548d1" Nov 25 15:26:31 crc kubenswrapper[4890]: I1125 15:26:31.173517 4890 scope.go:117] "RemoveContainer" containerID="8052599ccdad737d107b842db95dca88d95c47c4640b7dfd30857525688810ee" Nov 25 15:26:31 crc kubenswrapper[4890]: I1125 15:26:31.540827 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-slpt4" event={"ID":"f272410f-230d-43f2-946e-e34e4d9c05ea","Type":"ContainerStarted","Data":"d4a955e088086e86142cb4cd474083bb60970a3ea04ab43c1e5e2c9804079e62"} Nov 25 15:26:31 crc kubenswrapper[4890]: I1125 15:26:31.541801 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-slpt4" Nov 25 15:26:32 crc kubenswrapper[4890]: I1125 15:26:32.180352 4890 scope.go:117] "RemoveContainer" containerID="09ff68e9f7b5d7b971104be89b55602edbba51c72b2e776a35880be99ec6b53f" Nov 25 15:26:32 crc kubenswrapper[4890]: I1125 15:26:32.552028 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-lfnwl" event={"ID":"076c2a0e-7fdb-42e1-ab73-e296bba2b7a9","Type":"ContainerStarted","Data":"64007d634fb09cf5a80f6e776780dc93aabf6c2655fb99afbb725e7a605c1adb"} Nov 25 15:26:32 crc kubenswrapper[4890]: I1125 15:26:32.552510 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-lfnwl" Nov 25 15:26:33 crc kubenswrapper[4890]: I1125 15:26:33.173868 4890 scope.go:117] "RemoveContainer" containerID="c186dc6ff6591b7b58925577cec707c6ed3796c30a9ae21bb9565049cea41856" Nov 25 15:26:33 crc kubenswrapper[4890]: I1125 15:26:33.174065 4890 scope.go:117] "RemoveContainer" containerID="afe47ed941c0114eed272202844f29373d2d8c6c4f5c70cf119a83c56005a08d" Nov 25 15:26:33 crc kubenswrapper[4890]: I1125 15:26:33.571257 4890 generic.go:334] "Generic (PLEG): container finished" podID="4bbe88e6-b859-452a-baf1-8a963b475014" containerID="17a73befb091b6d575f44bfa544a3b79cec0c8a8c327edbac49c7aebfe563aae" exitCode=0 Nov 25 15:26:33 crc kubenswrapper[4890]: I1125 15:26:33.571339 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2xzsf" event={"ID":"4bbe88e6-b859-452a-baf1-8a963b475014","Type":"ContainerDied","Data":"17a73befb091b6d575f44bfa544a3b79cec0c8a8c327edbac49c7aebfe563aae"} Nov 25 15:26:33 crc kubenswrapper[4890]: I1125 15:26:33.572251 4890 scope.go:117] "RemoveContainer" containerID="17a73befb091b6d575f44bfa544a3b79cec0c8a8c327edbac49c7aebfe563aae" Nov 25 15:26:33 crc kubenswrapper[4890]: I1125 15:26:33.590363 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-wvjw7" event={"ID":"03353c8b-91f4-41ae-afc1-c522b1c73bb4","Type":"ContainerStarted","Data":"6dd60af535ad66fe44e35836adbf8d249695d2c229efa5c6e8f13ecdc32fc2ec"} Nov 25 15:26:33 crc kubenswrapper[4890]: I1125 15:26:33.591254 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-wvjw7" Nov 25 15:26:33 crc kubenswrapper[4890]: I1125 15:26:33.595422 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-stflb" event={"ID":"7e576602-19f5-4656-84dd-5ab80abca895","Type":"ContainerStarted","Data":"24eca631555b6e2fa97dbb7f54598143b7603117e799126d0495411d25204f2b"} Nov 25 15:26:33 crc kubenswrapper[4890]: I1125 15:26:33.595626 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-stflb" Nov 25 15:26:34 crc kubenswrapper[4890]: I1125 15:26:34.247872 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gh9mr" Nov 25 15:26:34 crc kubenswrapper[4890]: I1125 15:26:34.248521 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gh9mr" Nov 25 15:26:34 crc kubenswrapper[4890]: I1125 15:26:34.298336 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gh9mr" Nov 25 15:26:34 crc kubenswrapper[4890]: I1125 15:26:34.607251 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-dmlx4" event={"ID":"e26594ec-4bc2-4051-8a49-fc490bd3b6c9","Type":"ContainerStarted","Data":"2e57c25de62398e54932faa6809c637287bb68873c6cee08badff2e057f9015b"} Nov 25 15:26:34 crc kubenswrapper[4890]: I1125 15:26:34.656674 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gh9mr" Nov 25 15:26:35 crc kubenswrapper[4890]: I1125 15:26:35.173030 4890 scope.go:117] "RemoveContainer" containerID="5d93b4988f6bdbdc8c47c62e1f25b2fcb1a65f5d3e949121c19272784f723e37" Nov 25 15:26:35 crc kubenswrapper[4890]: I1125 15:26:35.173349 4890 scope.go:117] "RemoveContainer" containerID="a0074090c38ba41c6cd974ca26e67cc1c9cd2d25edd77a01b7f4e31a93593803" Nov 25 15:26:35 crc kubenswrapper[4890]: I1125 15:26:35.173480 4890 scope.go:117] "RemoveContainer" containerID="aeb0938dec764a3788ac5a6992985db59cc25e8a9f8cd69f3f2408a7e27cdd24" Nov 25 15:26:35 crc kubenswrapper[4890]: I1125 15:26:35.623121 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2xzsf" event={"ID":"4bbe88e6-b859-452a-baf1-8a963b475014","Type":"ContainerStarted","Data":"1b09b13d15f6c7bc9c5c5543ccd439f51c112eac192d2a5c7dfb1491f5441e75"} Nov 25 15:26:35 crc kubenswrapper[4890]: I1125 15:26:35.624360 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-2xzsf" Nov 25 15:26:35 crc kubenswrapper[4890]: I1125 15:26:35.627902 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-2xzsf" Nov 25 15:26:35 crc kubenswrapper[4890]: I1125 15:26:35.750144 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 25 15:26:35 crc kubenswrapper[4890]: I1125 15:26:35.818362 4890 scope.go:117] "RemoveContainer" containerID="f28158f6c45d4687b53cc873efafe9bd6ddd6a433e6e23b764315cf714baff16" Nov 25 15:26:35 crc kubenswrapper[4890]: I1125 15:26:35.859327 4890 scope.go:117] "RemoveContainer" containerID="97df7bc09544ad16b4be8d113257f79ac856eb0352b81415283734caeb5030bb" Nov 25 15:26:36 crc kubenswrapper[4890]: I1125 15:26:36.053622 4890 scope.go:117] "RemoveContainer" containerID="d1b1f601af90078f6742d375b3a9c3a0ce2fefdf39c2cde7a1057b3e2a3f8950" Nov 25 15:26:36 crc kubenswrapper[4890]: I1125 15:26:36.072419 4890 scope.go:117] "RemoveContainer" containerID="9400aa82e81acded3c8c82ed72b89555165e618f7269b989ffebeca39eb383dc" Nov 25 15:26:36 crc kubenswrapper[4890]: I1125 15:26:36.107072 4890 scope.go:117] "RemoveContainer" containerID="90c930538f8ad9440ca2e8fefaa6d6abf1268230e5c9ba07b74a553ef0141b4d" Nov 25 15:26:36 crc kubenswrapper[4890]: I1125 15:26:36.112574 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-pmmf2" Nov 25 15:26:36 crc kubenswrapper[4890]: I1125 15:26:36.122863 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-jcw7x" Nov 25 15:26:36 crc kubenswrapper[4890]: I1125 15:26:36.159778 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-lfnwl" Nov 25 15:26:36 crc kubenswrapper[4890]: I1125 15:26:36.211412 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-774b86978c-xvfgd" Nov 25 15:26:36 crc kubenswrapper[4890]: I1125 15:26:36.228667 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-j6pqs" Nov 25 15:26:36 crc kubenswrapper[4890]: I1125 15:26:36.306454 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-hxzf4" Nov 25 15:26:36 crc kubenswrapper[4890]: I1125 15:26:36.317418 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-rs6hx" Nov 25 15:26:36 crc kubenswrapper[4890]: I1125 15:26:36.518096 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-slpt4" Nov 25 15:26:36 crc kubenswrapper[4890]: I1125 15:26:36.572373 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-6zr4x" Nov 25 15:26:36 crc kubenswrapper[4890]: I1125 15:26:36.625887 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6jq7q" Nov 25 15:26:36 crc kubenswrapper[4890]: I1125 15:26:36.633537 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-68587559f4-9b4cf" event={"ID":"3e496c3d-9485-44bc-b3b1-dc1682a88434","Type":"ContainerStarted","Data":"b3456141b4e2caadcc48eba1c1d2eaa2b020e652320a97b8cd49de0e62f0fb83"} Nov 25 15:26:36 crc kubenswrapper[4890]: I1125 15:26:36.633802 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-68587559f4-9b4cf" Nov 25 15:26:36 crc kubenswrapper[4890]: I1125 15:26:36.637824 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-nshw7" event={"ID":"9f0e3cf9-dcc0-405f-a2de-9148844ec3ae","Type":"ContainerStarted","Data":"d8bb01be1e1d0da44f767d71996eff9a0980be607a604a6b70d04eeaca302679"} Nov 25 15:26:36 crc kubenswrapper[4890]: I1125 15:26:36.638404 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-nshw7" Nov 25 15:26:36 crc kubenswrapper[4890]: I1125 15:26:36.640578 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-vx75t" event={"ID":"29f250db-2a96-41ad-bb8f-5b503a9288d8","Type":"ContainerStarted","Data":"b38bfdbfd03de03a6719654a1adc4771fb6fd6e931526c4b29d986b812f7cf54"} Nov 25 15:26:36 crc kubenswrapper[4890]: I1125 15:26:36.719392 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-vx75t" Nov 25 15:26:36 crc kubenswrapper[4890]: I1125 15:26:36.777788 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-dzkgt" Nov 25 15:26:37 crc kubenswrapper[4890]: I1125 15:26:37.173749 4890 scope.go:117] "RemoveContainer" containerID="8453cefea24580a864e1b962ed8a682423dc9aaafad812309969556ac975663c" Nov 25 15:26:37 crc kubenswrapper[4890]: I1125 15:26:37.650239 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Nov 25 15:26:37 crc kubenswrapper[4890]: I1125 15:26:37.651968 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 25 15:26:37 crc kubenswrapper[4890]: I1125 15:26:37.652133 4890 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="b68b7a22193a1042f7b2a06df452033ab04e929a3b8f18b668593749e140c576" exitCode=137 Nov 25 15:26:37 crc kubenswrapper[4890]: I1125 15:26:37.652432 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"b68b7a22193a1042f7b2a06df452033ab04e929a3b8f18b668593749e140c576"} Nov 25 15:26:37 crc kubenswrapper[4890]: I1125 15:26:37.652501 4890 scope.go:117] "RemoveContainer" containerID="93325a6407389c4a7db0ba3c1a9c84a6980db3c3277cdf3860583a564b562bbe" Nov 25 15:26:37 crc kubenswrapper[4890]: I1125 15:26:37.654999 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-829nk" event={"ID":"4364eec5-444a-41f6-af1c-fe16b24bd044","Type":"ContainerStarted","Data":"09d743feb96a54b8366e2bdcaa858a163701f80769f8bdf31bbdf56b6f8045b7"} Nov 25 15:26:38 crc kubenswrapper[4890]: I1125 15:26:38.174125 4890 scope.go:117] "RemoveContainer" containerID="a18bbc200fff39fe093ad099ce5322f28fa7e5a281377e8a1b83796f88c742cc" Nov 25 15:26:38 crc kubenswrapper[4890]: I1125 15:26:38.664387 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Nov 25 15:26:40 crc kubenswrapper[4890]: I1125 15:26:40.671635 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-68587559f4-9b4cf" Nov 25 15:26:40 crc kubenswrapper[4890]: I1125 15:26:40.707469 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-stwjm" event={"ID":"77a5e8f9-92ab-48b5-be19-29a7a8e5df49","Type":"ContainerStarted","Data":"a52947cbc84743c4c8e294c9daacf28724301167324c1d16f500b37de9154ad7"} Nov 25 15:26:40 crc kubenswrapper[4890]: I1125 15:26:40.708858 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-stwjm" Nov 25 15:26:41 crc kubenswrapper[4890]: I1125 15:26:41.173391 4890 scope.go:117] "RemoveContainer" containerID="addcdee7c623c641d2c7c579c327f9f105ad3a6312c9860f69375f2691e21b26" Nov 25 15:26:41 crc kubenswrapper[4890]: I1125 15:26:41.719687 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-56bcr" event={"ID":"a9efeabb-bf1c-4de2-8322-961b284e95cc","Type":"ContainerStarted","Data":"185fcb733dc05ae402d719e2ef8f22a9aac66a7d735604a9659924689d3a123c"} Nov 25 15:26:41 crc kubenswrapper[4890]: I1125 15:26:41.720308 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-56bcr" Nov 25 15:26:41 crc kubenswrapper[4890]: I1125 15:26:41.722639 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Nov 25 15:26:41 crc kubenswrapper[4890]: I1125 15:26:41.723665 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"45181d3d91d8c382307c2244ff55b0e7bf873c0551eddb077f4890ba98b1a556"} Nov 25 15:26:44 crc kubenswrapper[4890]: I1125 15:26:44.344489 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 15:26:45 crc kubenswrapper[4890]: I1125 15:26:45.783810 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6g727"] Nov 25 15:26:45 crc kubenswrapper[4890]: I1125 15:26:45.786421 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6g727" Nov 25 15:26:45 crc kubenswrapper[4890]: I1125 15:26:45.791278 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6g727"] Nov 25 15:26:45 crc kubenswrapper[4890]: I1125 15:26:45.818546 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pv9s7\" (UniqueName: \"kubernetes.io/projected/b53e3920-fe5e-45fe-9b2d-495a5cb16b2c-kube-api-access-pv9s7\") pod \"community-operators-6g727\" (UID: \"b53e3920-fe5e-45fe-9b2d-495a5cb16b2c\") " pod="openshift-marketplace/community-operators-6g727" Nov 25 15:26:45 crc kubenswrapper[4890]: I1125 15:26:45.818602 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b53e3920-fe5e-45fe-9b2d-495a5cb16b2c-catalog-content\") pod \"community-operators-6g727\" (UID: \"b53e3920-fe5e-45fe-9b2d-495a5cb16b2c\") " pod="openshift-marketplace/community-operators-6g727" Nov 25 15:26:45 crc kubenswrapper[4890]: I1125 15:26:45.818650 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b53e3920-fe5e-45fe-9b2d-495a5cb16b2c-utilities\") pod \"community-operators-6g727\" (UID: \"b53e3920-fe5e-45fe-9b2d-495a5cb16b2c\") " pod="openshift-marketplace/community-operators-6g727" Nov 25 15:26:45 crc kubenswrapper[4890]: I1125 15:26:45.921346 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b53e3920-fe5e-45fe-9b2d-495a5cb16b2c-catalog-content\") pod \"community-operators-6g727\" (UID: \"b53e3920-fe5e-45fe-9b2d-495a5cb16b2c\") " pod="openshift-marketplace/community-operators-6g727" Nov 25 15:26:45 crc kubenswrapper[4890]: I1125 15:26:45.921465 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b53e3920-fe5e-45fe-9b2d-495a5cb16b2c-utilities\") pod \"community-operators-6g727\" (UID: \"b53e3920-fe5e-45fe-9b2d-495a5cb16b2c\") " pod="openshift-marketplace/community-operators-6g727" Nov 25 15:26:45 crc kubenswrapper[4890]: I1125 15:26:45.921607 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pv9s7\" (UniqueName: \"kubernetes.io/projected/b53e3920-fe5e-45fe-9b2d-495a5cb16b2c-kube-api-access-pv9s7\") pod \"community-operators-6g727\" (UID: \"b53e3920-fe5e-45fe-9b2d-495a5cb16b2c\") " pod="openshift-marketplace/community-operators-6g727" Nov 25 15:26:45 crc kubenswrapper[4890]: I1125 15:26:45.921979 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b53e3920-fe5e-45fe-9b2d-495a5cb16b2c-utilities\") pod \"community-operators-6g727\" (UID: \"b53e3920-fe5e-45fe-9b2d-495a5cb16b2c\") " pod="openshift-marketplace/community-operators-6g727" Nov 25 15:26:45 crc kubenswrapper[4890]: I1125 15:26:45.921980 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b53e3920-fe5e-45fe-9b2d-495a5cb16b2c-catalog-content\") pod \"community-operators-6g727\" (UID: \"b53e3920-fe5e-45fe-9b2d-495a5cb16b2c\") " pod="openshift-marketplace/community-operators-6g727" Nov 25 15:26:45 crc kubenswrapper[4890]: I1125 15:26:45.952076 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pv9s7\" (UniqueName: \"kubernetes.io/projected/b53e3920-fe5e-45fe-9b2d-495a5cb16b2c-kube-api-access-pv9s7\") pod \"community-operators-6g727\" (UID: \"b53e3920-fe5e-45fe-9b2d-495a5cb16b2c\") " pod="openshift-marketplace/community-operators-6g727" Nov 25 15:26:46 crc kubenswrapper[4890]: I1125 15:26:46.108490 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6g727" Nov 25 15:26:46 crc kubenswrapper[4890]: I1125 15:26:46.129412 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-stwjm" Nov 25 15:26:46 crc kubenswrapper[4890]: I1125 15:26:46.442861 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-wvjw7" Nov 25 15:26:46 crc kubenswrapper[4890]: I1125 15:26:46.576846 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6g727"] Nov 25 15:26:46 crc kubenswrapper[4890]: W1125 15:26:46.577195 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb53e3920_fe5e_45fe_9b2d_495a5cb16b2c.slice/crio-ad3647df1b566cc7c46012aacf4ae79d207d5a05e18409b5af4d99ed9799d7ec WatchSource:0}: Error finding container ad3647df1b566cc7c46012aacf4ae79d207d5a05e18409b5af4d99ed9799d7ec: Status 404 returned error can't find the container with id ad3647df1b566cc7c46012aacf4ae79d207d5a05e18409b5af4d99ed9799d7ec Nov 25 15:26:46 crc kubenswrapper[4890]: I1125 15:26:46.651944 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-56bcr" Nov 25 15:26:46 crc kubenswrapper[4890]: I1125 15:26:46.722508 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-vx75t" Nov 25 15:26:46 crc kubenswrapper[4890]: I1125 15:26:46.758141 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-nshw7" Nov 25 15:26:46 crc kubenswrapper[4890]: I1125 15:26:46.772959 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6g727" event={"ID":"b53e3920-fe5e-45fe-9b2d-495a5cb16b2c","Type":"ContainerStarted","Data":"ad3647df1b566cc7c46012aacf4ae79d207d5a05e18409b5af4d99ed9799d7ec"} Nov 25 15:26:46 crc kubenswrapper[4890]: I1125 15:26:46.791931 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-829nk" Nov 25 15:26:46 crc kubenswrapper[4890]: I1125 15:26:46.797218 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-829nk" Nov 25 15:26:47 crc kubenswrapper[4890]: I1125 15:26:47.121705 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-492nv"] Nov 25 15:26:47 crc kubenswrapper[4890]: I1125 15:26:47.124732 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-492nv" Nov 25 15:26:47 crc kubenswrapper[4890]: I1125 15:26:47.151768 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-492nv"] Nov 25 15:26:47 crc kubenswrapper[4890]: I1125 15:26:47.203260 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fcc307a-b17f-46ab-aabe-0a0361656701-utilities\") pod \"community-operators-492nv\" (UID: \"1fcc307a-b17f-46ab-aabe-0a0361656701\") " pod="openshift-marketplace/community-operators-492nv" Nov 25 15:26:47 crc kubenswrapper[4890]: I1125 15:26:47.203336 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fcc307a-b17f-46ab-aabe-0a0361656701-catalog-content\") pod \"community-operators-492nv\" (UID: \"1fcc307a-b17f-46ab-aabe-0a0361656701\") " pod="openshift-marketplace/community-operators-492nv" Nov 25 15:26:47 crc kubenswrapper[4890]: I1125 15:26:47.203364 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxd9v\" (UniqueName: \"kubernetes.io/projected/1fcc307a-b17f-46ab-aabe-0a0361656701-kube-api-access-gxd9v\") pod \"community-operators-492nv\" (UID: \"1fcc307a-b17f-46ab-aabe-0a0361656701\") " pod="openshift-marketplace/community-operators-492nv" Nov 25 15:26:47 crc kubenswrapper[4890]: I1125 15:26:47.252524 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 15:26:47 crc kubenswrapper[4890]: I1125 15:26:47.257477 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 15:26:47 crc kubenswrapper[4890]: I1125 15:26:47.306387 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fcc307a-b17f-46ab-aabe-0a0361656701-utilities\") pod \"community-operators-492nv\" (UID: \"1fcc307a-b17f-46ab-aabe-0a0361656701\") " pod="openshift-marketplace/community-operators-492nv" Nov 25 15:26:47 crc kubenswrapper[4890]: I1125 15:26:47.306461 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fcc307a-b17f-46ab-aabe-0a0361656701-catalog-content\") pod \"community-operators-492nv\" (UID: \"1fcc307a-b17f-46ab-aabe-0a0361656701\") " pod="openshift-marketplace/community-operators-492nv" Nov 25 15:26:47 crc kubenswrapper[4890]: I1125 15:26:47.306489 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxd9v\" (UniqueName: \"kubernetes.io/projected/1fcc307a-b17f-46ab-aabe-0a0361656701-kube-api-access-gxd9v\") pod \"community-operators-492nv\" (UID: \"1fcc307a-b17f-46ab-aabe-0a0361656701\") " pod="openshift-marketplace/community-operators-492nv" Nov 25 15:26:47 crc kubenswrapper[4890]: I1125 15:26:47.307066 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fcc307a-b17f-46ab-aabe-0a0361656701-utilities\") pod \"community-operators-492nv\" (UID: \"1fcc307a-b17f-46ab-aabe-0a0361656701\") " pod="openshift-marketplace/community-operators-492nv" Nov 25 15:26:47 crc kubenswrapper[4890]: I1125 15:26:47.307221 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fcc307a-b17f-46ab-aabe-0a0361656701-catalog-content\") pod \"community-operators-492nv\" (UID: \"1fcc307a-b17f-46ab-aabe-0a0361656701\") " pod="openshift-marketplace/community-operators-492nv" Nov 25 15:26:47 crc kubenswrapper[4890]: I1125 15:26:47.336670 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxd9v\" (UniqueName: \"kubernetes.io/projected/1fcc307a-b17f-46ab-aabe-0a0361656701-kube-api-access-gxd9v\") pod \"community-operators-492nv\" (UID: \"1fcc307a-b17f-46ab-aabe-0a0361656701\") " pod="openshift-marketplace/community-operators-492nv" Nov 25 15:26:47 crc kubenswrapper[4890]: I1125 15:26:47.455218 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-492nv" Nov 25 15:26:47 crc kubenswrapper[4890]: I1125 15:26:47.739432 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-stflb" Nov 25 15:26:47 crc kubenswrapper[4890]: I1125 15:26:47.952657 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-492nv"] Nov 25 15:26:48 crc kubenswrapper[4890]: I1125 15:26:48.174093 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-kr7td"] Nov 25 15:26:48 crc kubenswrapper[4890]: I1125 15:26:48.177772 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kr7td" Nov 25 15:26:48 crc kubenswrapper[4890]: I1125 15:26:48.255747 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kr7td"] Nov 25 15:26:48 crc kubenswrapper[4890]: I1125 15:26:48.353366 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb551926-2f33-4616-a423-caee14e79f74-catalog-content\") pod \"community-operators-kr7td\" (UID: \"cb551926-2f33-4616-a423-caee14e79f74\") " pod="openshift-marketplace/community-operators-kr7td" Nov 25 15:26:48 crc kubenswrapper[4890]: I1125 15:26:48.353444 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2plp\" (UniqueName: \"kubernetes.io/projected/cb551926-2f33-4616-a423-caee14e79f74-kube-api-access-c2plp\") pod \"community-operators-kr7td\" (UID: \"cb551926-2f33-4616-a423-caee14e79f74\") " pod="openshift-marketplace/community-operators-kr7td" Nov 25 15:26:48 crc kubenswrapper[4890]: I1125 15:26:48.354697 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb551926-2f33-4616-a423-caee14e79f74-utilities\") pod \"community-operators-kr7td\" (UID: \"cb551926-2f33-4616-a423-caee14e79f74\") " pod="openshift-marketplace/community-operators-kr7td" Nov 25 15:26:48 crc kubenswrapper[4890]: I1125 15:26:48.457098 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb551926-2f33-4616-a423-caee14e79f74-utilities\") pod \"community-operators-kr7td\" (UID: \"cb551926-2f33-4616-a423-caee14e79f74\") " pod="openshift-marketplace/community-operators-kr7td" Nov 25 15:26:48 crc kubenswrapper[4890]: I1125 15:26:48.457208 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb551926-2f33-4616-a423-caee14e79f74-catalog-content\") pod \"community-operators-kr7td\" (UID: \"cb551926-2f33-4616-a423-caee14e79f74\") " pod="openshift-marketplace/community-operators-kr7td" Nov 25 15:26:48 crc kubenswrapper[4890]: I1125 15:26:48.457234 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2plp\" (UniqueName: \"kubernetes.io/projected/cb551926-2f33-4616-a423-caee14e79f74-kube-api-access-c2plp\") pod \"community-operators-kr7td\" (UID: \"cb551926-2f33-4616-a423-caee14e79f74\") " pod="openshift-marketplace/community-operators-kr7td" Nov 25 15:26:48 crc kubenswrapper[4890]: I1125 15:26:48.457929 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb551926-2f33-4616-a423-caee14e79f74-catalog-content\") pod \"community-operators-kr7td\" (UID: \"cb551926-2f33-4616-a423-caee14e79f74\") " pod="openshift-marketplace/community-operators-kr7td" Nov 25 15:26:48 crc kubenswrapper[4890]: I1125 15:26:48.458053 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb551926-2f33-4616-a423-caee14e79f74-utilities\") pod \"community-operators-kr7td\" (UID: \"cb551926-2f33-4616-a423-caee14e79f74\") " pod="openshift-marketplace/community-operators-kr7td" Nov 25 15:26:48 crc kubenswrapper[4890]: I1125 15:26:48.479640 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2plp\" (UniqueName: \"kubernetes.io/projected/cb551926-2f33-4616-a423-caee14e79f74-kube-api-access-c2plp\") pod \"community-operators-kr7td\" (UID: \"cb551926-2f33-4616-a423-caee14e79f74\") " pod="openshift-marketplace/community-operators-kr7td" Nov 25 15:26:48 crc kubenswrapper[4890]: I1125 15:26:48.580357 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kr7td" Nov 25 15:26:48 crc kubenswrapper[4890]: I1125 15:26:48.808106 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-492nv" event={"ID":"1fcc307a-b17f-46ab-aabe-0a0361656701","Type":"ContainerStarted","Data":"bc3d2402820722971a63507a69a3578ea47f648835295734093bfd05126b245b"} Nov 25 15:26:48 crc kubenswrapper[4890]: I1125 15:26:48.811303 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6g727" event={"ID":"b53e3920-fe5e-45fe-9b2d-495a5cb16b2c","Type":"ContainerStarted","Data":"82e3cdda2919b11d5bf605afb47f59af1d0184f90723c6570cf19eb01d175be1"} Nov 25 15:26:49 crc kubenswrapper[4890]: I1125 15:26:49.113890 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kr7td"] Nov 25 15:26:49 crc kubenswrapper[4890]: I1125 15:26:49.370513 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-g8wbt"] Nov 25 15:26:49 crc kubenswrapper[4890]: I1125 15:26:49.373393 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g8wbt" Nov 25 15:26:49 crc kubenswrapper[4890]: I1125 15:26:49.382283 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g8wbt"] Nov 25 15:26:49 crc kubenswrapper[4890]: I1125 15:26:49.482902 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c070ef9-daba-4ae9-ac05-dff449baf01a-catalog-content\") pod \"community-operators-g8wbt\" (UID: \"7c070ef9-daba-4ae9-ac05-dff449baf01a\") " pod="openshift-marketplace/community-operators-g8wbt" Nov 25 15:26:49 crc kubenswrapper[4890]: I1125 15:26:49.482949 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzqtm\" (UniqueName: \"kubernetes.io/projected/7c070ef9-daba-4ae9-ac05-dff449baf01a-kube-api-access-rzqtm\") pod \"community-operators-g8wbt\" (UID: \"7c070ef9-daba-4ae9-ac05-dff449baf01a\") " pod="openshift-marketplace/community-operators-g8wbt" Nov 25 15:26:49 crc kubenswrapper[4890]: I1125 15:26:49.483378 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c070ef9-daba-4ae9-ac05-dff449baf01a-utilities\") pod \"community-operators-g8wbt\" (UID: \"7c070ef9-daba-4ae9-ac05-dff449baf01a\") " pod="openshift-marketplace/community-operators-g8wbt" Nov 25 15:26:49 crc kubenswrapper[4890]: I1125 15:26:49.585404 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c070ef9-daba-4ae9-ac05-dff449baf01a-utilities\") pod \"community-operators-g8wbt\" (UID: \"7c070ef9-daba-4ae9-ac05-dff449baf01a\") " pod="openshift-marketplace/community-operators-g8wbt" Nov 25 15:26:49 crc kubenswrapper[4890]: I1125 15:26:49.585684 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c070ef9-daba-4ae9-ac05-dff449baf01a-catalog-content\") pod \"community-operators-g8wbt\" (UID: \"7c070ef9-daba-4ae9-ac05-dff449baf01a\") " pod="openshift-marketplace/community-operators-g8wbt" Nov 25 15:26:49 crc kubenswrapper[4890]: I1125 15:26:49.585765 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzqtm\" (UniqueName: \"kubernetes.io/projected/7c070ef9-daba-4ae9-ac05-dff449baf01a-kube-api-access-rzqtm\") pod \"community-operators-g8wbt\" (UID: \"7c070ef9-daba-4ae9-ac05-dff449baf01a\") " pod="openshift-marketplace/community-operators-g8wbt" Nov 25 15:26:49 crc kubenswrapper[4890]: I1125 15:26:49.585929 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c070ef9-daba-4ae9-ac05-dff449baf01a-utilities\") pod \"community-operators-g8wbt\" (UID: \"7c070ef9-daba-4ae9-ac05-dff449baf01a\") " pod="openshift-marketplace/community-operators-g8wbt" Nov 25 15:26:49 crc kubenswrapper[4890]: I1125 15:26:49.586074 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c070ef9-daba-4ae9-ac05-dff449baf01a-catalog-content\") pod \"community-operators-g8wbt\" (UID: \"7c070ef9-daba-4ae9-ac05-dff449baf01a\") " pod="openshift-marketplace/community-operators-g8wbt" Nov 25 15:26:49 crc kubenswrapper[4890]: I1125 15:26:49.604317 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzqtm\" (UniqueName: \"kubernetes.io/projected/7c070ef9-daba-4ae9-ac05-dff449baf01a-kube-api-access-rzqtm\") pod \"community-operators-g8wbt\" (UID: \"7c070ef9-daba-4ae9-ac05-dff449baf01a\") " pod="openshift-marketplace/community-operators-g8wbt" Nov 25 15:26:49 crc kubenswrapper[4890]: I1125 15:26:49.698151 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g8wbt" Nov 25 15:26:49 crc kubenswrapper[4890]: I1125 15:26:49.832835 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kr7td" event={"ID":"cb551926-2f33-4616-a423-caee14e79f74","Type":"ContainerStarted","Data":"50abe642979368e933bd97e0e71a88affb9cd7aad1c209c05f9aa209991764ef"} Nov 25 15:26:49 crc kubenswrapper[4890]: I1125 15:26:49.836038 4890 generic.go:334] "Generic (PLEG): container finished" podID="b53e3920-fe5e-45fe-9b2d-495a5cb16b2c" containerID="82e3cdda2919b11d5bf605afb47f59af1d0184f90723c6570cf19eb01d175be1" exitCode=0 Nov 25 15:26:49 crc kubenswrapper[4890]: I1125 15:26:49.836064 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6g727" event={"ID":"b53e3920-fe5e-45fe-9b2d-495a5cb16b2c","Type":"ContainerDied","Data":"82e3cdda2919b11d5bf605afb47f59af1d0184f90723c6570cf19eb01d175be1"} Nov 25 15:26:50 crc kubenswrapper[4890]: I1125 15:26:50.157966 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g8wbt"] Nov 25 15:26:50 crc kubenswrapper[4890]: I1125 15:26:50.571560 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gxfqw"] Nov 25 15:26:50 crc kubenswrapper[4890]: I1125 15:26:50.574373 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gxfqw" Nov 25 15:26:50 crc kubenswrapper[4890]: I1125 15:26:50.584276 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gxfqw"] Nov 25 15:26:50 crc kubenswrapper[4890]: I1125 15:26:50.716941 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc5b6ee3-033f-4b8f-82b7-46e28498f625-utilities\") pod \"community-operators-gxfqw\" (UID: \"fc5b6ee3-033f-4b8f-82b7-46e28498f625\") " pod="openshift-marketplace/community-operators-gxfqw" Nov 25 15:26:50 crc kubenswrapper[4890]: I1125 15:26:50.717578 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc5b6ee3-033f-4b8f-82b7-46e28498f625-catalog-content\") pod \"community-operators-gxfqw\" (UID: \"fc5b6ee3-033f-4b8f-82b7-46e28498f625\") " pod="openshift-marketplace/community-operators-gxfqw" Nov 25 15:26:50 crc kubenswrapper[4890]: I1125 15:26:50.717778 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzzl6\" (UniqueName: \"kubernetes.io/projected/fc5b6ee3-033f-4b8f-82b7-46e28498f625-kube-api-access-nzzl6\") pod \"community-operators-gxfqw\" (UID: \"fc5b6ee3-033f-4b8f-82b7-46e28498f625\") " pod="openshift-marketplace/community-operators-gxfqw" Nov 25 15:26:50 crc kubenswrapper[4890]: I1125 15:26:50.819347 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc5b6ee3-033f-4b8f-82b7-46e28498f625-utilities\") pod \"community-operators-gxfqw\" (UID: \"fc5b6ee3-033f-4b8f-82b7-46e28498f625\") " pod="openshift-marketplace/community-operators-gxfqw" Nov 25 15:26:50 crc kubenswrapper[4890]: I1125 15:26:50.819885 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc5b6ee3-033f-4b8f-82b7-46e28498f625-catalog-content\") pod \"community-operators-gxfqw\" (UID: \"fc5b6ee3-033f-4b8f-82b7-46e28498f625\") " pod="openshift-marketplace/community-operators-gxfqw" Nov 25 15:26:50 crc kubenswrapper[4890]: I1125 15:26:50.819934 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc5b6ee3-033f-4b8f-82b7-46e28498f625-utilities\") pod \"community-operators-gxfqw\" (UID: \"fc5b6ee3-033f-4b8f-82b7-46e28498f625\") " pod="openshift-marketplace/community-operators-gxfqw" Nov 25 15:26:50 crc kubenswrapper[4890]: I1125 15:26:50.819948 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzzl6\" (UniqueName: \"kubernetes.io/projected/fc5b6ee3-033f-4b8f-82b7-46e28498f625-kube-api-access-nzzl6\") pod \"community-operators-gxfqw\" (UID: \"fc5b6ee3-033f-4b8f-82b7-46e28498f625\") " pod="openshift-marketplace/community-operators-gxfqw" Nov 25 15:26:50 crc kubenswrapper[4890]: I1125 15:26:50.820558 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc5b6ee3-033f-4b8f-82b7-46e28498f625-catalog-content\") pod \"community-operators-gxfqw\" (UID: \"fc5b6ee3-033f-4b8f-82b7-46e28498f625\") " pod="openshift-marketplace/community-operators-gxfqw" Nov 25 15:26:50 crc kubenswrapper[4890]: I1125 15:26:50.846732 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g8wbt" event={"ID":"7c070ef9-daba-4ae9-ac05-dff449baf01a","Type":"ContainerStarted","Data":"811aa1444ea14e66014986dfab6ee4501d180834bd5711daad15c99b109dc58c"} Nov 25 15:26:50 crc kubenswrapper[4890]: I1125 15:26:50.848025 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzzl6\" (UniqueName: \"kubernetes.io/projected/fc5b6ee3-033f-4b8f-82b7-46e28498f625-kube-api-access-nzzl6\") pod \"community-operators-gxfqw\" (UID: \"fc5b6ee3-033f-4b8f-82b7-46e28498f625\") " pod="openshift-marketplace/community-operators-gxfqw" Nov 25 15:26:50 crc kubenswrapper[4890]: I1125 15:26:50.898225 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gxfqw" Nov 25 15:26:51 crc kubenswrapper[4890]: I1125 15:26:51.371841 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gxfqw"] Nov 25 15:26:51 crc kubenswrapper[4890]: W1125 15:26:51.375686 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfc5b6ee3_033f_4b8f_82b7_46e28498f625.slice/crio-945c2abc9b2224e5901c60b5487bf24102fd493a5019b956b10c3fd51008abe6 WatchSource:0}: Error finding container 945c2abc9b2224e5901c60b5487bf24102fd493a5019b956b10c3fd51008abe6: Status 404 returned error can't find the container with id 945c2abc9b2224e5901c60b5487bf24102fd493a5019b956b10c3fd51008abe6 Nov 25 15:26:51 crc kubenswrapper[4890]: I1125 15:26:51.773407 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ddmtt"] Nov 25 15:26:51 crc kubenswrapper[4890]: I1125 15:26:51.777404 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ddmtt" Nov 25 15:26:51 crc kubenswrapper[4890]: I1125 15:26:51.789619 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ddmtt"] Nov 25 15:26:51 crc kubenswrapper[4890]: I1125 15:26:51.856755 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gxfqw" event={"ID":"fc5b6ee3-033f-4b8f-82b7-46e28498f625","Type":"ContainerStarted","Data":"945c2abc9b2224e5901c60b5487bf24102fd493a5019b956b10c3fd51008abe6"} Nov 25 15:26:51 crc kubenswrapper[4890]: I1125 15:26:51.858944 4890 generic.go:334] "Generic (PLEG): container finished" podID="cb551926-2f33-4616-a423-caee14e79f74" containerID="324d3b471e383e8d75a04f41b4d279d80982950aa20b6522732f1ccd7ba394a8" exitCode=0 Nov 25 15:26:51 crc kubenswrapper[4890]: I1125 15:26:51.858997 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kr7td" event={"ID":"cb551926-2f33-4616-a423-caee14e79f74","Type":"ContainerDied","Data":"324d3b471e383e8d75a04f41b4d279d80982950aa20b6522732f1ccd7ba394a8"} Nov 25 15:26:51 crc kubenswrapper[4890]: I1125 15:26:51.861415 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-492nv" event={"ID":"1fcc307a-b17f-46ab-aabe-0a0361656701","Type":"ContainerStarted","Data":"23a8a7f031d8d7fb3e4865029b2f6864c91388d8980dd59c503a7a2bc53430d6"} Nov 25 15:26:51 crc kubenswrapper[4890]: I1125 15:26:51.865660 4890 generic.go:334] "Generic (PLEG): container finished" podID="7c070ef9-daba-4ae9-ac05-dff449baf01a" containerID="27512991ab22e87d4234e5b73728a5adba06ff97504b8aea926363411d19c00f" exitCode=0 Nov 25 15:26:51 crc kubenswrapper[4890]: I1125 15:26:51.865786 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g8wbt" event={"ID":"7c070ef9-daba-4ae9-ac05-dff449baf01a","Type":"ContainerDied","Data":"27512991ab22e87d4234e5b73728a5adba06ff97504b8aea926363411d19c00f"} Nov 25 15:26:51 crc kubenswrapper[4890]: I1125 15:26:51.940396 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9tvq\" (UniqueName: \"kubernetes.io/projected/b9fbb88d-f5f9-41c6-9719-15be794dfb6c-kube-api-access-j9tvq\") pod \"community-operators-ddmtt\" (UID: \"b9fbb88d-f5f9-41c6-9719-15be794dfb6c\") " pod="openshift-marketplace/community-operators-ddmtt" Nov 25 15:26:51 crc kubenswrapper[4890]: I1125 15:26:51.940464 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9fbb88d-f5f9-41c6-9719-15be794dfb6c-utilities\") pod \"community-operators-ddmtt\" (UID: \"b9fbb88d-f5f9-41c6-9719-15be794dfb6c\") " pod="openshift-marketplace/community-operators-ddmtt" Nov 25 15:26:51 crc kubenswrapper[4890]: I1125 15:26:51.940542 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9fbb88d-f5f9-41c6-9719-15be794dfb6c-catalog-content\") pod \"community-operators-ddmtt\" (UID: \"b9fbb88d-f5f9-41c6-9719-15be794dfb6c\") " pod="openshift-marketplace/community-operators-ddmtt" Nov 25 15:26:52 crc kubenswrapper[4890]: I1125 15:26:52.044293 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9tvq\" (UniqueName: \"kubernetes.io/projected/b9fbb88d-f5f9-41c6-9719-15be794dfb6c-kube-api-access-j9tvq\") pod \"community-operators-ddmtt\" (UID: \"b9fbb88d-f5f9-41c6-9719-15be794dfb6c\") " pod="openshift-marketplace/community-operators-ddmtt" Nov 25 15:26:52 crc kubenswrapper[4890]: I1125 15:26:52.044395 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9fbb88d-f5f9-41c6-9719-15be794dfb6c-utilities\") pod \"community-operators-ddmtt\" (UID: \"b9fbb88d-f5f9-41c6-9719-15be794dfb6c\") " pod="openshift-marketplace/community-operators-ddmtt" Nov 25 15:26:52 crc kubenswrapper[4890]: I1125 15:26:52.044511 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9fbb88d-f5f9-41c6-9719-15be794dfb6c-catalog-content\") pod \"community-operators-ddmtt\" (UID: \"b9fbb88d-f5f9-41c6-9719-15be794dfb6c\") " pod="openshift-marketplace/community-operators-ddmtt" Nov 25 15:26:52 crc kubenswrapper[4890]: I1125 15:26:52.046233 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9fbb88d-f5f9-41c6-9719-15be794dfb6c-utilities\") pod \"community-operators-ddmtt\" (UID: \"b9fbb88d-f5f9-41c6-9719-15be794dfb6c\") " pod="openshift-marketplace/community-operators-ddmtt" Nov 25 15:26:52 crc kubenswrapper[4890]: I1125 15:26:52.046824 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9fbb88d-f5f9-41c6-9719-15be794dfb6c-catalog-content\") pod \"community-operators-ddmtt\" (UID: \"b9fbb88d-f5f9-41c6-9719-15be794dfb6c\") " pod="openshift-marketplace/community-operators-ddmtt" Nov 25 15:26:52 crc kubenswrapper[4890]: I1125 15:26:52.074399 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9tvq\" (UniqueName: \"kubernetes.io/projected/b9fbb88d-f5f9-41c6-9719-15be794dfb6c-kube-api-access-j9tvq\") pod \"community-operators-ddmtt\" (UID: \"b9fbb88d-f5f9-41c6-9719-15be794dfb6c\") " pod="openshift-marketplace/community-operators-ddmtt" Nov 25 15:26:52 crc kubenswrapper[4890]: I1125 15:26:52.102963 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ddmtt" Nov 25 15:26:52 crc kubenswrapper[4890]: I1125 15:26:52.588116 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ddmtt"] Nov 25 15:26:52 crc kubenswrapper[4890]: W1125 15:26:52.594375 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb9fbb88d_f5f9_41c6_9719_15be794dfb6c.slice/crio-98c4050c253d51989b1845008d34976e3e890d2b80a459e1ba7166dbd9890e80 WatchSource:0}: Error finding container 98c4050c253d51989b1845008d34976e3e890d2b80a459e1ba7166dbd9890e80: Status 404 returned error can't find the container with id 98c4050c253d51989b1845008d34976e3e890d2b80a459e1ba7166dbd9890e80 Nov 25 15:26:52 crc kubenswrapper[4890]: I1125 15:26:52.894376 4890 generic.go:334] "Generic (PLEG): container finished" podID="fc5b6ee3-033f-4b8f-82b7-46e28498f625" containerID="88c913329b26945b6331ba5d2f4f26f2dd514f71abb8ac40652b3fecfc7fca00" exitCode=0 Nov 25 15:26:52 crc kubenswrapper[4890]: I1125 15:26:52.895673 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gxfqw" event={"ID":"fc5b6ee3-033f-4b8f-82b7-46e28498f625","Type":"ContainerDied","Data":"88c913329b26945b6331ba5d2f4f26f2dd514f71abb8ac40652b3fecfc7fca00"} Nov 25 15:26:52 crc kubenswrapper[4890]: I1125 15:26:52.902698 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ddmtt" event={"ID":"b9fbb88d-f5f9-41c6-9719-15be794dfb6c","Type":"ContainerStarted","Data":"98c4050c253d51989b1845008d34976e3e890d2b80a459e1ba7166dbd9890e80"} Nov 25 15:26:52 crc kubenswrapper[4890]: I1125 15:26:52.906967 4890 generic.go:334] "Generic (PLEG): container finished" podID="1fcc307a-b17f-46ab-aabe-0a0361656701" containerID="23a8a7f031d8d7fb3e4865029b2f6864c91388d8980dd59c503a7a2bc53430d6" exitCode=0 Nov 25 15:26:52 crc kubenswrapper[4890]: I1125 15:26:52.907021 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-492nv" event={"ID":"1fcc307a-b17f-46ab-aabe-0a0361656701","Type":"ContainerDied","Data":"23a8a7f031d8d7fb3e4865029b2f6864c91388d8980dd59c503a7a2bc53430d6"} Nov 25 15:26:52 crc kubenswrapper[4890]: I1125 15:26:52.969217 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-v8mj2"] Nov 25 15:26:52 crc kubenswrapper[4890]: I1125 15:26:52.971860 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v8mj2" Nov 25 15:26:52 crc kubenswrapper[4890]: I1125 15:26:52.981585 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v8mj2"] Nov 25 15:26:53 crc kubenswrapper[4890]: I1125 15:26:53.084665 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7x9w4\" (UniqueName: \"kubernetes.io/projected/00771e9e-5583-42ab-b834-4643d9bbdbec-kube-api-access-7x9w4\") pod \"community-operators-v8mj2\" (UID: \"00771e9e-5583-42ab-b834-4643d9bbdbec\") " pod="openshift-marketplace/community-operators-v8mj2" Nov 25 15:26:53 crc kubenswrapper[4890]: I1125 15:26:53.084744 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00771e9e-5583-42ab-b834-4643d9bbdbec-utilities\") pod \"community-operators-v8mj2\" (UID: \"00771e9e-5583-42ab-b834-4643d9bbdbec\") " pod="openshift-marketplace/community-operators-v8mj2" Nov 25 15:26:53 crc kubenswrapper[4890]: I1125 15:26:53.084785 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00771e9e-5583-42ab-b834-4643d9bbdbec-catalog-content\") pod \"community-operators-v8mj2\" (UID: \"00771e9e-5583-42ab-b834-4643d9bbdbec\") " pod="openshift-marketplace/community-operators-v8mj2" Nov 25 15:26:53 crc kubenswrapper[4890]: I1125 15:26:53.186215 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7x9w4\" (UniqueName: \"kubernetes.io/projected/00771e9e-5583-42ab-b834-4643d9bbdbec-kube-api-access-7x9w4\") pod \"community-operators-v8mj2\" (UID: \"00771e9e-5583-42ab-b834-4643d9bbdbec\") " pod="openshift-marketplace/community-operators-v8mj2" Nov 25 15:26:53 crc kubenswrapper[4890]: I1125 15:26:53.186290 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00771e9e-5583-42ab-b834-4643d9bbdbec-utilities\") pod \"community-operators-v8mj2\" (UID: \"00771e9e-5583-42ab-b834-4643d9bbdbec\") " pod="openshift-marketplace/community-operators-v8mj2" Nov 25 15:26:53 crc kubenswrapper[4890]: I1125 15:26:53.186330 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00771e9e-5583-42ab-b834-4643d9bbdbec-catalog-content\") pod \"community-operators-v8mj2\" (UID: \"00771e9e-5583-42ab-b834-4643d9bbdbec\") " pod="openshift-marketplace/community-operators-v8mj2" Nov 25 15:26:53 crc kubenswrapper[4890]: I1125 15:26:53.186789 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00771e9e-5583-42ab-b834-4643d9bbdbec-catalog-content\") pod \"community-operators-v8mj2\" (UID: \"00771e9e-5583-42ab-b834-4643d9bbdbec\") " pod="openshift-marketplace/community-operators-v8mj2" Nov 25 15:26:53 crc kubenswrapper[4890]: I1125 15:26:53.186948 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00771e9e-5583-42ab-b834-4643d9bbdbec-utilities\") pod \"community-operators-v8mj2\" (UID: \"00771e9e-5583-42ab-b834-4643d9bbdbec\") " pod="openshift-marketplace/community-operators-v8mj2" Nov 25 15:26:53 crc kubenswrapper[4890]: I1125 15:26:53.204945 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7x9w4\" (UniqueName: \"kubernetes.io/projected/00771e9e-5583-42ab-b834-4643d9bbdbec-kube-api-access-7x9w4\") pod \"community-operators-v8mj2\" (UID: \"00771e9e-5583-42ab-b834-4643d9bbdbec\") " pod="openshift-marketplace/community-operators-v8mj2" Nov 25 15:26:53 crc kubenswrapper[4890]: I1125 15:26:53.313664 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v8mj2" Nov 25 15:26:53 crc kubenswrapper[4890]: I1125 15:26:53.872664 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v8mj2"] Nov 25 15:26:53 crc kubenswrapper[4890]: W1125 15:26:53.877445 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod00771e9e_5583_42ab_b834_4643d9bbdbec.slice/crio-49174bc1bc2916c23aad34ea827d1d0da668cc5ce08d9282197851218bf6fc0e WatchSource:0}: Error finding container 49174bc1bc2916c23aad34ea827d1d0da668cc5ce08d9282197851218bf6fc0e: Status 404 returned error can't find the container with id 49174bc1bc2916c23aad34ea827d1d0da668cc5ce08d9282197851218bf6fc0e Nov 25 15:26:53 crc kubenswrapper[4890]: I1125 15:26:53.947325 4890 generic.go:334] "Generic (PLEG): container finished" podID="b9fbb88d-f5f9-41c6-9719-15be794dfb6c" containerID="50d3fcd65f895ac4b64138cb2a54b5f4ab83b31085f788a228b32f5683880aff" exitCode=0 Nov 25 15:26:53 crc kubenswrapper[4890]: I1125 15:26:53.947650 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ddmtt" event={"ID":"b9fbb88d-f5f9-41c6-9719-15be794dfb6c","Type":"ContainerDied","Data":"50d3fcd65f895ac4b64138cb2a54b5f4ab83b31085f788a228b32f5683880aff"} Nov 25 15:26:53 crc kubenswrapper[4890]: I1125 15:26:53.950192 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v8mj2" event={"ID":"00771e9e-5583-42ab-b834-4643d9bbdbec","Type":"ContainerStarted","Data":"49174bc1bc2916c23aad34ea827d1d0da668cc5ce08d9282197851218bf6fc0e"} Nov 25 15:26:54 crc kubenswrapper[4890]: I1125 15:26:54.184850 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-h44zl"] Nov 25 15:26:54 crc kubenswrapper[4890]: I1125 15:26:54.187350 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h44zl" Nov 25 15:26:54 crc kubenswrapper[4890]: I1125 15:26:54.190072 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-h44zl"] Nov 25 15:26:54 crc kubenswrapper[4890]: I1125 15:26:54.205302 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08bbb8ce-17b9-4797-80a7-de3beca270f4-catalog-content\") pod \"community-operators-h44zl\" (UID: \"08bbb8ce-17b9-4797-80a7-de3beca270f4\") " pod="openshift-marketplace/community-operators-h44zl" Nov 25 15:26:54 crc kubenswrapper[4890]: I1125 15:26:54.205363 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49srl\" (UniqueName: \"kubernetes.io/projected/08bbb8ce-17b9-4797-80a7-de3beca270f4-kube-api-access-49srl\") pod \"community-operators-h44zl\" (UID: \"08bbb8ce-17b9-4797-80a7-de3beca270f4\") " pod="openshift-marketplace/community-operators-h44zl" Nov 25 15:26:54 crc kubenswrapper[4890]: I1125 15:26:54.205462 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08bbb8ce-17b9-4797-80a7-de3beca270f4-utilities\") pod \"community-operators-h44zl\" (UID: \"08bbb8ce-17b9-4797-80a7-de3beca270f4\") " pod="openshift-marketplace/community-operators-h44zl" Nov 25 15:26:54 crc kubenswrapper[4890]: I1125 15:26:54.307487 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08bbb8ce-17b9-4797-80a7-de3beca270f4-utilities\") pod \"community-operators-h44zl\" (UID: \"08bbb8ce-17b9-4797-80a7-de3beca270f4\") " pod="openshift-marketplace/community-operators-h44zl" Nov 25 15:26:54 crc kubenswrapper[4890]: I1125 15:26:54.307653 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08bbb8ce-17b9-4797-80a7-de3beca270f4-catalog-content\") pod \"community-operators-h44zl\" (UID: \"08bbb8ce-17b9-4797-80a7-de3beca270f4\") " pod="openshift-marketplace/community-operators-h44zl" Nov 25 15:26:54 crc kubenswrapper[4890]: I1125 15:26:54.307700 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49srl\" (UniqueName: \"kubernetes.io/projected/08bbb8ce-17b9-4797-80a7-de3beca270f4-kube-api-access-49srl\") pod \"community-operators-h44zl\" (UID: \"08bbb8ce-17b9-4797-80a7-de3beca270f4\") " pod="openshift-marketplace/community-operators-h44zl" Nov 25 15:26:54 crc kubenswrapper[4890]: I1125 15:26:54.308064 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08bbb8ce-17b9-4797-80a7-de3beca270f4-utilities\") pod \"community-operators-h44zl\" (UID: \"08bbb8ce-17b9-4797-80a7-de3beca270f4\") " pod="openshift-marketplace/community-operators-h44zl" Nov 25 15:26:54 crc kubenswrapper[4890]: I1125 15:26:54.308125 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08bbb8ce-17b9-4797-80a7-de3beca270f4-catalog-content\") pod \"community-operators-h44zl\" (UID: \"08bbb8ce-17b9-4797-80a7-de3beca270f4\") " pod="openshift-marketplace/community-operators-h44zl" Nov 25 15:26:54 crc kubenswrapper[4890]: I1125 15:26:54.329487 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49srl\" (UniqueName: \"kubernetes.io/projected/08bbb8ce-17b9-4797-80a7-de3beca270f4-kube-api-access-49srl\") pod \"community-operators-h44zl\" (UID: \"08bbb8ce-17b9-4797-80a7-de3beca270f4\") " pod="openshift-marketplace/community-operators-h44zl" Nov 25 15:26:54 crc kubenswrapper[4890]: I1125 15:26:54.349143 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 15:26:54 crc kubenswrapper[4890]: I1125 15:26:54.624305 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h44zl" Nov 25 15:26:54 crc kubenswrapper[4890]: I1125 15:26:54.963894 4890 generic.go:334] "Generic (PLEG): container finished" podID="00771e9e-5583-42ab-b834-4643d9bbdbec" containerID="aba18652c1832a9d5351cea01a0e87328ddbaca9fafe0ce1be9bb66182cfebae" exitCode=0 Nov 25 15:26:54 crc kubenswrapper[4890]: I1125 15:26:54.964005 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v8mj2" event={"ID":"00771e9e-5583-42ab-b834-4643d9bbdbec","Type":"ContainerDied","Data":"aba18652c1832a9d5351cea01a0e87328ddbaca9fafe0ce1be9bb66182cfebae"} Nov 25 15:26:55 crc kubenswrapper[4890]: I1125 15:26:55.368245 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rcbhc"] Nov 25 15:26:55 crc kubenswrapper[4890]: I1125 15:26:55.370284 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rcbhc" Nov 25 15:26:55 crc kubenswrapper[4890]: I1125 15:26:55.378064 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rcbhc"] Nov 25 15:26:55 crc kubenswrapper[4890]: I1125 15:26:55.429151 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b18eb2e9-77b4-46ba-9e32-31c1e9b73963-catalog-content\") pod \"community-operators-rcbhc\" (UID: \"b18eb2e9-77b4-46ba-9e32-31c1e9b73963\") " pod="openshift-marketplace/community-operators-rcbhc" Nov 25 15:26:55 crc kubenswrapper[4890]: I1125 15:26:55.429267 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b18eb2e9-77b4-46ba-9e32-31c1e9b73963-utilities\") pod \"community-operators-rcbhc\" (UID: \"b18eb2e9-77b4-46ba-9e32-31c1e9b73963\") " pod="openshift-marketplace/community-operators-rcbhc" Nov 25 15:26:55 crc kubenswrapper[4890]: I1125 15:26:55.429359 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hr2md\" (UniqueName: \"kubernetes.io/projected/b18eb2e9-77b4-46ba-9e32-31c1e9b73963-kube-api-access-hr2md\") pod \"community-operators-rcbhc\" (UID: \"b18eb2e9-77b4-46ba-9e32-31c1e9b73963\") " pod="openshift-marketplace/community-operators-rcbhc" Nov 25 15:26:55 crc kubenswrapper[4890]: I1125 15:26:55.531761 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b18eb2e9-77b4-46ba-9e32-31c1e9b73963-catalog-content\") pod \"community-operators-rcbhc\" (UID: \"b18eb2e9-77b4-46ba-9e32-31c1e9b73963\") " pod="openshift-marketplace/community-operators-rcbhc" Nov 25 15:26:55 crc kubenswrapper[4890]: I1125 15:26:55.531880 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b18eb2e9-77b4-46ba-9e32-31c1e9b73963-utilities\") pod \"community-operators-rcbhc\" (UID: \"b18eb2e9-77b4-46ba-9e32-31c1e9b73963\") " pod="openshift-marketplace/community-operators-rcbhc" Nov 25 15:26:55 crc kubenswrapper[4890]: I1125 15:26:55.532098 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hr2md\" (UniqueName: \"kubernetes.io/projected/b18eb2e9-77b4-46ba-9e32-31c1e9b73963-kube-api-access-hr2md\") pod \"community-operators-rcbhc\" (UID: \"b18eb2e9-77b4-46ba-9e32-31c1e9b73963\") " pod="openshift-marketplace/community-operators-rcbhc" Nov 25 15:26:55 crc kubenswrapper[4890]: I1125 15:26:55.532596 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b18eb2e9-77b4-46ba-9e32-31c1e9b73963-catalog-content\") pod \"community-operators-rcbhc\" (UID: \"b18eb2e9-77b4-46ba-9e32-31c1e9b73963\") " pod="openshift-marketplace/community-operators-rcbhc" Nov 25 15:26:55 crc kubenswrapper[4890]: I1125 15:26:55.532710 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b18eb2e9-77b4-46ba-9e32-31c1e9b73963-utilities\") pod \"community-operators-rcbhc\" (UID: \"b18eb2e9-77b4-46ba-9e32-31c1e9b73963\") " pod="openshift-marketplace/community-operators-rcbhc" Nov 25 15:26:55 crc kubenswrapper[4890]: I1125 15:26:55.551336 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hr2md\" (UniqueName: \"kubernetes.io/projected/b18eb2e9-77b4-46ba-9e32-31c1e9b73963-kube-api-access-hr2md\") pod \"community-operators-rcbhc\" (UID: \"b18eb2e9-77b4-46ba-9e32-31c1e9b73963\") " pod="openshift-marketplace/community-operators-rcbhc" Nov 25 15:26:55 crc kubenswrapper[4890]: I1125 15:26:55.721950 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rcbhc" Nov 25 15:26:56 crc kubenswrapper[4890]: I1125 15:26:56.448371 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 15:26:56 crc kubenswrapper[4890]: I1125 15:26:56.448827 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 15:26:56 crc kubenswrapper[4890]: I1125 15:26:56.594807 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-h44zl"] Nov 25 15:26:56 crc kubenswrapper[4890]: I1125 15:26:56.985080 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h44zl" event={"ID":"08bbb8ce-17b9-4797-80a7-de3beca270f4","Type":"ContainerStarted","Data":"32d6e83eadedd458f04dc0727d7bb04a7014d5a7650ee8ec9b4afbb8ce0f29b4"} Nov 25 15:26:58 crc kubenswrapper[4890]: I1125 15:26:58.423278 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rcbhc"] Nov 25 15:26:58 crc kubenswrapper[4890]: W1125 15:26:58.516027 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb18eb2e9_77b4_46ba_9e32_31c1e9b73963.slice/crio-a758e8482e3d183aa59e2885cbf5a7e5cd809f48c007a4712f6979acf2a9337d WatchSource:0}: Error finding container a758e8482e3d183aa59e2885cbf5a7e5cd809f48c007a4712f6979acf2a9337d: Status 404 returned error can't find the container with id a758e8482e3d183aa59e2885cbf5a7e5cd809f48c007a4712f6979acf2a9337d Nov 25 15:26:59 crc kubenswrapper[4890]: I1125 15:26:59.067319 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kr7td" event={"ID":"cb551926-2f33-4616-a423-caee14e79f74","Type":"ContainerStarted","Data":"b2e5c93300d37e63edce4624180363409c7c85e5902027f073ae462124ddb757"} Nov 25 15:26:59 crc kubenswrapper[4890]: I1125 15:26:59.084716 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6g727" event={"ID":"b53e3920-fe5e-45fe-9b2d-495a5cb16b2c","Type":"ContainerStarted","Data":"32ed94c9ed63590d671f1a4e6483de3d71759ff85d32e388abb90a0e81bd77fe"} Nov 25 15:26:59 crc kubenswrapper[4890]: I1125 15:26:59.098735 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rcbhc" event={"ID":"b18eb2e9-77b4-46ba-9e32-31c1e9b73963","Type":"ContainerStarted","Data":"a758e8482e3d183aa59e2885cbf5a7e5cd809f48c007a4712f6979acf2a9337d"} Nov 25 15:26:59 crc kubenswrapper[4890]: E1125 15:26:59.405562 4890 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcb551926_2f33_4616_a423_caee14e79f74.slice/crio-b2e5c93300d37e63edce4624180363409c7c85e5902027f073ae462124ddb757.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb53e3920_fe5e_45fe_9b2d_495a5cb16b2c.slice/crio-conmon-32ed94c9ed63590d671f1a4e6483de3d71759ff85d32e388abb90a0e81bd77fe.scope\": RecentStats: unable to find data in memory cache]" Nov 25 15:26:59 crc kubenswrapper[4890]: I1125 15:26:59.937137 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-66fc8c7fbf-lnjxm" Nov 25 15:27:00 crc kubenswrapper[4890]: I1125 15:27:00.122444 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ddmtt" event={"ID":"b9fbb88d-f5f9-41c6-9719-15be794dfb6c","Type":"ContainerStarted","Data":"5cb9b523a30d43bc28da23a27df37463d35f8bac93e2c391c98b08c3b39391b6"} Nov 25 15:27:00 crc kubenswrapper[4890]: I1125 15:27:00.126360 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-492nv" event={"ID":"1fcc307a-b17f-46ab-aabe-0a0361656701","Type":"ContainerStarted","Data":"60e586812f3d16610b2f2890383a9106ad815412ccc1d6c56e39851322fd6bb3"} Nov 25 15:27:00 crc kubenswrapper[4890]: I1125 15:27:00.132938 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g8wbt" event={"ID":"7c070ef9-daba-4ae9-ac05-dff449baf01a","Type":"ContainerStarted","Data":"46a904cda6ea2d958239b096888450e7894e0315371384e6fb9a6ae541dc625b"} Nov 25 15:27:00 crc kubenswrapper[4890]: I1125 15:27:00.143616 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gxfqw" event={"ID":"fc5b6ee3-033f-4b8f-82b7-46e28498f625","Type":"ContainerStarted","Data":"ec6ac43bf61445fbec47e7b8dc51b5fccdc266550c9b100a10b6282659496fe4"} Nov 25 15:27:00 crc kubenswrapper[4890]: I1125 15:27:00.153233 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v8mj2" event={"ID":"00771e9e-5583-42ab-b834-4643d9bbdbec","Type":"ContainerStarted","Data":"25383107f01eedb261016e63aaa8d7e88ddff63b3170d5bdc523e921f50e8778"} Nov 25 15:27:00 crc kubenswrapper[4890]: I1125 15:27:00.166623 4890 generic.go:334] "Generic (PLEG): container finished" podID="cb551926-2f33-4616-a423-caee14e79f74" containerID="b2e5c93300d37e63edce4624180363409c7c85e5902027f073ae462124ddb757" exitCode=0 Nov 25 15:27:00 crc kubenswrapper[4890]: I1125 15:27:00.166705 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kr7td" event={"ID":"cb551926-2f33-4616-a423-caee14e79f74","Type":"ContainerDied","Data":"b2e5c93300d37e63edce4624180363409c7c85e5902027f073ae462124ddb757"} Nov 25 15:27:00 crc kubenswrapper[4890]: I1125 15:27:00.177459 4890 generic.go:334] "Generic (PLEG): container finished" podID="b53e3920-fe5e-45fe-9b2d-495a5cb16b2c" containerID="32ed94c9ed63590d671f1a4e6483de3d71759ff85d32e388abb90a0e81bd77fe" exitCode=0 Nov 25 15:27:00 crc kubenswrapper[4890]: I1125 15:27:00.193065 4890 generic.go:334] "Generic (PLEG): container finished" podID="b18eb2e9-77b4-46ba-9e32-31c1e9b73963" containerID="3b02fe5096c64727128fcb100ea22d7cfae1d8624fbb52019491522ac365aa2b" exitCode=0 Nov 25 15:27:00 crc kubenswrapper[4890]: I1125 15:27:00.216661 4890 generic.go:334] "Generic (PLEG): container finished" podID="08bbb8ce-17b9-4797-80a7-de3beca270f4" containerID="a8e112324cbebe6d570d7610b4ef244ecf7bb4b8aab01e300595ae4f47c1ee92" exitCode=0 Nov 25 15:27:00 crc kubenswrapper[4890]: I1125 15:27:00.216713 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6g727" event={"ID":"b53e3920-fe5e-45fe-9b2d-495a5cb16b2c","Type":"ContainerDied","Data":"32ed94c9ed63590d671f1a4e6483de3d71759ff85d32e388abb90a0e81bd77fe"} Nov 25 15:27:00 crc kubenswrapper[4890]: I1125 15:27:00.216766 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rcbhc" event={"ID":"b18eb2e9-77b4-46ba-9e32-31c1e9b73963","Type":"ContainerDied","Data":"3b02fe5096c64727128fcb100ea22d7cfae1d8624fbb52019491522ac365aa2b"} Nov 25 15:27:00 crc kubenswrapper[4890]: I1125 15:27:00.216782 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h44zl" event={"ID":"08bbb8ce-17b9-4797-80a7-de3beca270f4","Type":"ContainerDied","Data":"a8e112324cbebe6d570d7610b4ef244ecf7bb4b8aab01e300595ae4f47c1ee92"} Nov 25 15:27:01 crc kubenswrapper[4890]: I1125 15:27:01.230652 4890 generic.go:334] "Generic (PLEG): container finished" podID="b9fbb88d-f5f9-41c6-9719-15be794dfb6c" containerID="5cb9b523a30d43bc28da23a27df37463d35f8bac93e2c391c98b08c3b39391b6" exitCode=0 Nov 25 15:27:01 crc kubenswrapper[4890]: I1125 15:27:01.232394 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ddmtt" event={"ID":"b9fbb88d-f5f9-41c6-9719-15be794dfb6c","Type":"ContainerDied","Data":"5cb9b523a30d43bc28da23a27df37463d35f8bac93e2c391c98b08c3b39391b6"} Nov 25 15:27:01 crc kubenswrapper[4890]: I1125 15:27:01.238364 4890 generic.go:334] "Generic (PLEG): container finished" podID="1fcc307a-b17f-46ab-aabe-0a0361656701" containerID="60e586812f3d16610b2f2890383a9106ad815412ccc1d6c56e39851322fd6bb3" exitCode=0 Nov 25 15:27:01 crc kubenswrapper[4890]: I1125 15:27:01.239148 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-492nv" event={"ID":"1fcc307a-b17f-46ab-aabe-0a0361656701","Type":"ContainerDied","Data":"60e586812f3d16610b2f2890383a9106ad815412ccc1d6c56e39851322fd6bb3"} Nov 25 15:27:01 crc kubenswrapper[4890]: I1125 15:27:01.246069 4890 generic.go:334] "Generic (PLEG): container finished" podID="7c070ef9-daba-4ae9-ac05-dff449baf01a" containerID="46a904cda6ea2d958239b096888450e7894e0315371384e6fb9a6ae541dc625b" exitCode=0 Nov 25 15:27:01 crc kubenswrapper[4890]: I1125 15:27:01.246724 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g8wbt" event={"ID":"7c070ef9-daba-4ae9-ac05-dff449baf01a","Type":"ContainerDied","Data":"46a904cda6ea2d958239b096888450e7894e0315371384e6fb9a6ae541dc625b"} Nov 25 15:27:01 crc kubenswrapper[4890]: I1125 15:27:01.255981 4890 generic.go:334] "Generic (PLEG): container finished" podID="fc5b6ee3-033f-4b8f-82b7-46e28498f625" containerID="ec6ac43bf61445fbec47e7b8dc51b5fccdc266550c9b100a10b6282659496fe4" exitCode=0 Nov 25 15:27:01 crc kubenswrapper[4890]: I1125 15:27:01.256042 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gxfqw" event={"ID":"fc5b6ee3-033f-4b8f-82b7-46e28498f625","Type":"ContainerDied","Data":"ec6ac43bf61445fbec47e7b8dc51b5fccdc266550c9b100a10b6282659496fe4"} Nov 25 15:27:01 crc kubenswrapper[4890]: I1125 15:27:01.263271 4890 generic.go:334] "Generic (PLEG): container finished" podID="00771e9e-5583-42ab-b834-4643d9bbdbec" containerID="25383107f01eedb261016e63aaa8d7e88ddff63b3170d5bdc523e921f50e8778" exitCode=0 Nov 25 15:27:01 crc kubenswrapper[4890]: I1125 15:27:01.263305 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v8mj2" event={"ID":"00771e9e-5583-42ab-b834-4643d9bbdbec","Type":"ContainerDied","Data":"25383107f01eedb261016e63aaa8d7e88ddff63b3170d5bdc523e921f50e8778"} Nov 25 15:27:14 crc kubenswrapper[4890]: I1125 15:27:14.103719 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cr58b"] Nov 25 15:27:14 crc kubenswrapper[4890]: I1125 15:27:14.105627 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cr58b" Nov 25 15:27:14 crc kubenswrapper[4890]: I1125 15:27:14.108821 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-22zzp" Nov 25 15:27:14 crc kubenswrapper[4890]: I1125 15:27:14.109049 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 15:27:14 crc kubenswrapper[4890]: I1125 15:27:14.109129 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 15:27:14 crc kubenswrapper[4890]: I1125 15:27:14.109317 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 15:27:14 crc kubenswrapper[4890]: I1125 15:27:14.118902 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cr58b"] Nov 25 15:27:14 crc kubenswrapper[4890]: I1125 15:27:14.165541 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6d15a3f2-bbd4-43ae-b620-93251f74369e-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-cr58b\" (UID: \"6d15a3f2-bbd4-43ae-b620-93251f74369e\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cr58b" Nov 25 15:27:14 crc kubenswrapper[4890]: I1125 15:27:14.165656 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xq2st\" (UniqueName: \"kubernetes.io/projected/6d15a3f2-bbd4-43ae-b620-93251f74369e-kube-api-access-xq2st\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-cr58b\" (UID: \"6d15a3f2-bbd4-43ae-b620-93251f74369e\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cr58b" Nov 25 15:27:14 crc kubenswrapper[4890]: I1125 15:27:14.165701 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6d15a3f2-bbd4-43ae-b620-93251f74369e-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-cr58b\" (UID: \"6d15a3f2-bbd4-43ae-b620-93251f74369e\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cr58b" Nov 25 15:27:14 crc kubenswrapper[4890]: I1125 15:27:14.165792 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d15a3f2-bbd4-43ae-b620-93251f74369e-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-cr58b\" (UID: \"6d15a3f2-bbd4-43ae-b620-93251f74369e\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cr58b" Nov 25 15:27:14 crc kubenswrapper[4890]: I1125 15:27:14.267335 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xq2st\" (UniqueName: \"kubernetes.io/projected/6d15a3f2-bbd4-43ae-b620-93251f74369e-kube-api-access-xq2st\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-cr58b\" (UID: \"6d15a3f2-bbd4-43ae-b620-93251f74369e\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cr58b" Nov 25 15:27:14 crc kubenswrapper[4890]: I1125 15:27:14.267413 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6d15a3f2-bbd4-43ae-b620-93251f74369e-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-cr58b\" (UID: \"6d15a3f2-bbd4-43ae-b620-93251f74369e\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cr58b" Nov 25 15:27:14 crc kubenswrapper[4890]: I1125 15:27:14.267518 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d15a3f2-bbd4-43ae-b620-93251f74369e-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-cr58b\" (UID: \"6d15a3f2-bbd4-43ae-b620-93251f74369e\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cr58b" Nov 25 15:27:14 crc kubenswrapper[4890]: I1125 15:27:14.267650 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6d15a3f2-bbd4-43ae-b620-93251f74369e-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-cr58b\" (UID: \"6d15a3f2-bbd4-43ae-b620-93251f74369e\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cr58b" Nov 25 15:27:14 crc kubenswrapper[4890]: I1125 15:27:14.273820 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6d15a3f2-bbd4-43ae-b620-93251f74369e-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-cr58b\" (UID: \"6d15a3f2-bbd4-43ae-b620-93251f74369e\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cr58b" Nov 25 15:27:14 crc kubenswrapper[4890]: I1125 15:27:14.274390 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6d15a3f2-bbd4-43ae-b620-93251f74369e-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-cr58b\" (UID: \"6d15a3f2-bbd4-43ae-b620-93251f74369e\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cr58b" Nov 25 15:27:14 crc kubenswrapper[4890]: I1125 15:27:14.274727 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d15a3f2-bbd4-43ae-b620-93251f74369e-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-cr58b\" (UID: \"6d15a3f2-bbd4-43ae-b620-93251f74369e\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cr58b" Nov 25 15:27:14 crc kubenswrapper[4890]: I1125 15:27:14.294244 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xq2st\" (UniqueName: \"kubernetes.io/projected/6d15a3f2-bbd4-43ae-b620-93251f74369e-kube-api-access-xq2st\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-cr58b\" (UID: \"6d15a3f2-bbd4-43ae-b620-93251f74369e\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cr58b" Nov 25 15:27:14 crc kubenswrapper[4890]: I1125 15:27:14.467736 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cr58b" Nov 25 15:27:17 crc kubenswrapper[4890]: I1125 15:27:17.804336 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-48z2b"] Nov 25 15:27:17 crc kubenswrapper[4890]: I1125 15:27:17.806904 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-48z2b" Nov 25 15:27:17 crc kubenswrapper[4890]: I1125 15:27:17.821747 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-48z2b"] Nov 25 15:27:17 crc kubenswrapper[4890]: I1125 15:27:17.839676 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gvv6\" (UniqueName: \"kubernetes.io/projected/eacc1f05-701e-49af-b2d7-9c049f6cddb9-kube-api-access-2gvv6\") pod \"redhat-operators-48z2b\" (UID: \"eacc1f05-701e-49af-b2d7-9c049f6cddb9\") " pod="openshift-marketplace/redhat-operators-48z2b" Nov 25 15:27:17 crc kubenswrapper[4890]: I1125 15:27:17.839822 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eacc1f05-701e-49af-b2d7-9c049f6cddb9-catalog-content\") pod \"redhat-operators-48z2b\" (UID: \"eacc1f05-701e-49af-b2d7-9c049f6cddb9\") " pod="openshift-marketplace/redhat-operators-48z2b" Nov 25 15:27:17 crc kubenswrapper[4890]: I1125 15:27:17.839880 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eacc1f05-701e-49af-b2d7-9c049f6cddb9-utilities\") pod \"redhat-operators-48z2b\" (UID: \"eacc1f05-701e-49af-b2d7-9c049f6cddb9\") " pod="openshift-marketplace/redhat-operators-48z2b" Nov 25 15:27:17 crc kubenswrapper[4890]: I1125 15:27:17.941845 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gvv6\" (UniqueName: \"kubernetes.io/projected/eacc1f05-701e-49af-b2d7-9c049f6cddb9-kube-api-access-2gvv6\") pod \"redhat-operators-48z2b\" (UID: \"eacc1f05-701e-49af-b2d7-9c049f6cddb9\") " pod="openshift-marketplace/redhat-operators-48z2b" Nov 25 15:27:17 crc kubenswrapper[4890]: I1125 15:27:17.941975 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eacc1f05-701e-49af-b2d7-9c049f6cddb9-catalog-content\") pod \"redhat-operators-48z2b\" (UID: \"eacc1f05-701e-49af-b2d7-9c049f6cddb9\") " pod="openshift-marketplace/redhat-operators-48z2b" Nov 25 15:27:17 crc kubenswrapper[4890]: I1125 15:27:17.942017 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eacc1f05-701e-49af-b2d7-9c049f6cddb9-utilities\") pod \"redhat-operators-48z2b\" (UID: \"eacc1f05-701e-49af-b2d7-9c049f6cddb9\") " pod="openshift-marketplace/redhat-operators-48z2b" Nov 25 15:27:17 crc kubenswrapper[4890]: I1125 15:27:17.942450 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eacc1f05-701e-49af-b2d7-9c049f6cddb9-catalog-content\") pod \"redhat-operators-48z2b\" (UID: \"eacc1f05-701e-49af-b2d7-9c049f6cddb9\") " pod="openshift-marketplace/redhat-operators-48z2b" Nov 25 15:27:17 crc kubenswrapper[4890]: I1125 15:27:17.942544 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eacc1f05-701e-49af-b2d7-9c049f6cddb9-utilities\") pod \"redhat-operators-48z2b\" (UID: \"eacc1f05-701e-49af-b2d7-9c049f6cddb9\") " pod="openshift-marketplace/redhat-operators-48z2b" Nov 25 15:27:17 crc kubenswrapper[4890]: I1125 15:27:17.960867 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gvv6\" (UniqueName: \"kubernetes.io/projected/eacc1f05-701e-49af-b2d7-9c049f6cddb9-kube-api-access-2gvv6\") pod \"redhat-operators-48z2b\" (UID: \"eacc1f05-701e-49af-b2d7-9c049f6cddb9\") " pod="openshift-marketplace/redhat-operators-48z2b" Nov 25 15:27:18 crc kubenswrapper[4890]: I1125 15:27:18.177510 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-48z2b" Nov 25 15:27:18 crc kubenswrapper[4890]: I1125 15:27:18.776986 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mtfgl"] Nov 25 15:27:18 crc kubenswrapper[4890]: I1125 15:27:18.779058 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mtfgl" Nov 25 15:27:18 crc kubenswrapper[4890]: I1125 15:27:18.789369 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mtfgl"] Nov 25 15:27:18 crc kubenswrapper[4890]: I1125 15:27:18.858684 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qk5x8\" (UniqueName: \"kubernetes.io/projected/f7235b7e-d6fe-45a2-93fd-66b0a77b4b76-kube-api-access-qk5x8\") pod \"redhat-operators-mtfgl\" (UID: \"f7235b7e-d6fe-45a2-93fd-66b0a77b4b76\") " pod="openshift-marketplace/redhat-operators-mtfgl" Nov 25 15:27:18 crc kubenswrapper[4890]: I1125 15:27:18.858773 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7235b7e-d6fe-45a2-93fd-66b0a77b4b76-utilities\") pod \"redhat-operators-mtfgl\" (UID: \"f7235b7e-d6fe-45a2-93fd-66b0a77b4b76\") " pod="openshift-marketplace/redhat-operators-mtfgl" Nov 25 15:27:18 crc kubenswrapper[4890]: I1125 15:27:18.858862 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7235b7e-d6fe-45a2-93fd-66b0a77b4b76-catalog-content\") pod \"redhat-operators-mtfgl\" (UID: \"f7235b7e-d6fe-45a2-93fd-66b0a77b4b76\") " pod="openshift-marketplace/redhat-operators-mtfgl" Nov 25 15:27:18 crc kubenswrapper[4890]: I1125 15:27:18.960575 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qk5x8\" (UniqueName: \"kubernetes.io/projected/f7235b7e-d6fe-45a2-93fd-66b0a77b4b76-kube-api-access-qk5x8\") pod \"redhat-operators-mtfgl\" (UID: \"f7235b7e-d6fe-45a2-93fd-66b0a77b4b76\") " pod="openshift-marketplace/redhat-operators-mtfgl" Nov 25 15:27:18 crc kubenswrapper[4890]: I1125 15:27:18.960978 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7235b7e-d6fe-45a2-93fd-66b0a77b4b76-utilities\") pod \"redhat-operators-mtfgl\" (UID: \"f7235b7e-d6fe-45a2-93fd-66b0a77b4b76\") " pod="openshift-marketplace/redhat-operators-mtfgl" Nov 25 15:27:18 crc kubenswrapper[4890]: I1125 15:27:18.961098 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7235b7e-d6fe-45a2-93fd-66b0a77b4b76-catalog-content\") pod \"redhat-operators-mtfgl\" (UID: \"f7235b7e-d6fe-45a2-93fd-66b0a77b4b76\") " pod="openshift-marketplace/redhat-operators-mtfgl" Nov 25 15:27:18 crc kubenswrapper[4890]: I1125 15:27:18.961726 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7235b7e-d6fe-45a2-93fd-66b0a77b4b76-utilities\") pod \"redhat-operators-mtfgl\" (UID: \"f7235b7e-d6fe-45a2-93fd-66b0a77b4b76\") " pod="openshift-marketplace/redhat-operators-mtfgl" Nov 25 15:27:18 crc kubenswrapper[4890]: I1125 15:27:18.961748 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7235b7e-d6fe-45a2-93fd-66b0a77b4b76-catalog-content\") pod \"redhat-operators-mtfgl\" (UID: \"f7235b7e-d6fe-45a2-93fd-66b0a77b4b76\") " pod="openshift-marketplace/redhat-operators-mtfgl" Nov 25 15:27:18 crc kubenswrapper[4890]: I1125 15:27:18.988895 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qk5x8\" (UniqueName: \"kubernetes.io/projected/f7235b7e-d6fe-45a2-93fd-66b0a77b4b76-kube-api-access-qk5x8\") pod \"redhat-operators-mtfgl\" (UID: \"f7235b7e-d6fe-45a2-93fd-66b0a77b4b76\") " pod="openshift-marketplace/redhat-operators-mtfgl" Nov 25 15:27:19 crc kubenswrapper[4890]: I1125 15:27:19.111879 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mtfgl" Nov 25 15:27:19 crc kubenswrapper[4890]: I1125 15:27:19.986067 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4fdm7"] Nov 25 15:27:19 crc kubenswrapper[4890]: I1125 15:27:19.988816 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4fdm7" Nov 25 15:27:20 crc kubenswrapper[4890]: I1125 15:27:20.003210 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4fdm7"] Nov 25 15:27:20 crc kubenswrapper[4890]: I1125 15:27:20.083373 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0a34ff8-8257-4f49-90d2-90576cbc40b7-catalog-content\") pod \"redhat-operators-4fdm7\" (UID: \"a0a34ff8-8257-4f49-90d2-90576cbc40b7\") " pod="openshift-marketplace/redhat-operators-4fdm7" Nov 25 15:27:20 crc kubenswrapper[4890]: I1125 15:27:20.083534 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0a34ff8-8257-4f49-90d2-90576cbc40b7-utilities\") pod \"redhat-operators-4fdm7\" (UID: \"a0a34ff8-8257-4f49-90d2-90576cbc40b7\") " pod="openshift-marketplace/redhat-operators-4fdm7" Nov 25 15:27:20 crc kubenswrapper[4890]: I1125 15:27:20.083616 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbzmx\" (UniqueName: \"kubernetes.io/projected/a0a34ff8-8257-4f49-90d2-90576cbc40b7-kube-api-access-vbzmx\") pod \"redhat-operators-4fdm7\" (UID: \"a0a34ff8-8257-4f49-90d2-90576cbc40b7\") " pod="openshift-marketplace/redhat-operators-4fdm7" Nov 25 15:27:20 crc kubenswrapper[4890]: I1125 15:27:20.185498 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbzmx\" (UniqueName: \"kubernetes.io/projected/a0a34ff8-8257-4f49-90d2-90576cbc40b7-kube-api-access-vbzmx\") pod \"redhat-operators-4fdm7\" (UID: \"a0a34ff8-8257-4f49-90d2-90576cbc40b7\") " pod="openshift-marketplace/redhat-operators-4fdm7" Nov 25 15:27:20 crc kubenswrapper[4890]: I1125 15:27:20.185620 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0a34ff8-8257-4f49-90d2-90576cbc40b7-catalog-content\") pod \"redhat-operators-4fdm7\" (UID: \"a0a34ff8-8257-4f49-90d2-90576cbc40b7\") " pod="openshift-marketplace/redhat-operators-4fdm7" Nov 25 15:27:20 crc kubenswrapper[4890]: I1125 15:27:20.185861 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0a34ff8-8257-4f49-90d2-90576cbc40b7-utilities\") pod \"redhat-operators-4fdm7\" (UID: \"a0a34ff8-8257-4f49-90d2-90576cbc40b7\") " pod="openshift-marketplace/redhat-operators-4fdm7" Nov 25 15:27:20 crc kubenswrapper[4890]: I1125 15:27:20.186375 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0a34ff8-8257-4f49-90d2-90576cbc40b7-catalog-content\") pod \"redhat-operators-4fdm7\" (UID: \"a0a34ff8-8257-4f49-90d2-90576cbc40b7\") " pod="openshift-marketplace/redhat-operators-4fdm7" Nov 25 15:27:20 crc kubenswrapper[4890]: I1125 15:27:20.186562 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0a34ff8-8257-4f49-90d2-90576cbc40b7-utilities\") pod \"redhat-operators-4fdm7\" (UID: \"a0a34ff8-8257-4f49-90d2-90576cbc40b7\") " pod="openshift-marketplace/redhat-operators-4fdm7" Nov 25 15:27:20 crc kubenswrapper[4890]: I1125 15:27:20.206513 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbzmx\" (UniqueName: \"kubernetes.io/projected/a0a34ff8-8257-4f49-90d2-90576cbc40b7-kube-api-access-vbzmx\") pod \"redhat-operators-4fdm7\" (UID: \"a0a34ff8-8257-4f49-90d2-90576cbc40b7\") " pod="openshift-marketplace/redhat-operators-4fdm7" Nov 25 15:27:20 crc kubenswrapper[4890]: I1125 15:27:20.309700 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4fdm7" Nov 25 15:27:21 crc kubenswrapper[4890]: I1125 15:27:21.178149 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-f24bd"] Nov 25 15:27:21 crc kubenswrapper[4890]: I1125 15:27:21.180599 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f24bd" Nov 25 15:27:21 crc kubenswrapper[4890]: I1125 15:27:21.196295 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f24bd"] Nov 25 15:27:21 crc kubenswrapper[4890]: I1125 15:27:21.211434 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45wdj\" (UniqueName: \"kubernetes.io/projected/0f05c16a-1239-42a7-9910-ca03ef42b0f7-kube-api-access-45wdj\") pod \"redhat-operators-f24bd\" (UID: \"0f05c16a-1239-42a7-9910-ca03ef42b0f7\") " pod="openshift-marketplace/redhat-operators-f24bd" Nov 25 15:27:21 crc kubenswrapper[4890]: I1125 15:27:21.211558 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f05c16a-1239-42a7-9910-ca03ef42b0f7-utilities\") pod \"redhat-operators-f24bd\" (UID: \"0f05c16a-1239-42a7-9910-ca03ef42b0f7\") " pod="openshift-marketplace/redhat-operators-f24bd" Nov 25 15:27:21 crc kubenswrapper[4890]: I1125 15:27:21.211718 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f05c16a-1239-42a7-9910-ca03ef42b0f7-catalog-content\") pod \"redhat-operators-f24bd\" (UID: \"0f05c16a-1239-42a7-9910-ca03ef42b0f7\") " pod="openshift-marketplace/redhat-operators-f24bd" Nov 25 15:27:21 crc kubenswrapper[4890]: I1125 15:27:21.313329 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45wdj\" (UniqueName: \"kubernetes.io/projected/0f05c16a-1239-42a7-9910-ca03ef42b0f7-kube-api-access-45wdj\") pod \"redhat-operators-f24bd\" (UID: \"0f05c16a-1239-42a7-9910-ca03ef42b0f7\") " pod="openshift-marketplace/redhat-operators-f24bd" Nov 25 15:27:21 crc kubenswrapper[4890]: I1125 15:27:21.313485 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f05c16a-1239-42a7-9910-ca03ef42b0f7-utilities\") pod \"redhat-operators-f24bd\" (UID: \"0f05c16a-1239-42a7-9910-ca03ef42b0f7\") " pod="openshift-marketplace/redhat-operators-f24bd" Nov 25 15:27:21 crc kubenswrapper[4890]: I1125 15:27:21.313624 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f05c16a-1239-42a7-9910-ca03ef42b0f7-catalog-content\") pod \"redhat-operators-f24bd\" (UID: \"0f05c16a-1239-42a7-9910-ca03ef42b0f7\") " pod="openshift-marketplace/redhat-operators-f24bd" Nov 25 15:27:21 crc kubenswrapper[4890]: I1125 15:27:21.314253 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f05c16a-1239-42a7-9910-ca03ef42b0f7-catalog-content\") pod \"redhat-operators-f24bd\" (UID: \"0f05c16a-1239-42a7-9910-ca03ef42b0f7\") " pod="openshift-marketplace/redhat-operators-f24bd" Nov 25 15:27:21 crc kubenswrapper[4890]: I1125 15:27:21.314440 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f05c16a-1239-42a7-9910-ca03ef42b0f7-utilities\") pod \"redhat-operators-f24bd\" (UID: \"0f05c16a-1239-42a7-9910-ca03ef42b0f7\") " pod="openshift-marketplace/redhat-operators-f24bd" Nov 25 15:27:21 crc kubenswrapper[4890]: I1125 15:27:21.335000 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45wdj\" (UniqueName: \"kubernetes.io/projected/0f05c16a-1239-42a7-9910-ca03ef42b0f7-kube-api-access-45wdj\") pod \"redhat-operators-f24bd\" (UID: \"0f05c16a-1239-42a7-9910-ca03ef42b0f7\") " pod="openshift-marketplace/redhat-operators-f24bd" Nov 25 15:27:21 crc kubenswrapper[4890]: I1125 15:27:21.500529 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f24bd" Nov 25 15:27:23 crc kubenswrapper[4890]: I1125 15:27:23.769470 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kqz9d"] Nov 25 15:27:23 crc kubenswrapper[4890]: I1125 15:27:23.770637 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kqz9d" podUID="56f8430f-3421-47db-b225-5c40fd14c8bd" containerName="registry-server" containerID="cri-o://41b14b144856e47ab31876e7cebe7e0578f5a19f3916527de1b9c742c0b771c1" gracePeriod=2 Nov 25 15:27:24 crc kubenswrapper[4890]: E1125 15:27:24.282973 4890 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="41b14b144856e47ab31876e7cebe7e0578f5a19f3916527de1b9c742c0b771c1" cmd=["grpc_health_probe","-addr=:50051"] Nov 25 15:27:24 crc kubenswrapper[4890]: E1125 15:27:24.297313 4890 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="41b14b144856e47ab31876e7cebe7e0578f5a19f3916527de1b9c742c0b771c1" cmd=["grpc_health_probe","-addr=:50051"] Nov 25 15:27:24 crc kubenswrapper[4890]: E1125 15:27:24.319604 4890 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="41b14b144856e47ab31876e7cebe7e0578f5a19f3916527de1b9c742c0b771c1" cmd=["grpc_health_probe","-addr=:50051"] Nov 25 15:27:24 crc kubenswrapper[4890]: E1125 15:27:24.319951 4890 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-marketplace/certified-operators-kqz9d" podUID="56f8430f-3421-47db-b225-5c40fd14c8bd" containerName="registry-server" Nov 25 15:27:26 crc kubenswrapper[4890]: I1125 15:27:26.185212 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-24vzw"] Nov 25 15:27:26 crc kubenswrapper[4890]: I1125 15:27:26.187700 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-24vzw" Nov 25 15:27:26 crc kubenswrapper[4890]: I1125 15:27:26.192108 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-24vzw"] Nov 25 15:27:26 crc kubenswrapper[4890]: I1125 15:27:26.227254 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f82a9b69-631a-430c-9ae8-e18fbcb439dc-catalog-content\") pod \"redhat-marketplace-24vzw\" (UID: \"f82a9b69-631a-430c-9ae8-e18fbcb439dc\") " pod="openshift-marketplace/redhat-marketplace-24vzw" Nov 25 15:27:26 crc kubenswrapper[4890]: I1125 15:27:26.227310 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgmd7\" (UniqueName: \"kubernetes.io/projected/f82a9b69-631a-430c-9ae8-e18fbcb439dc-kube-api-access-kgmd7\") pod \"redhat-marketplace-24vzw\" (UID: \"f82a9b69-631a-430c-9ae8-e18fbcb439dc\") " pod="openshift-marketplace/redhat-marketplace-24vzw" Nov 25 15:27:26 crc kubenswrapper[4890]: I1125 15:27:26.227393 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f82a9b69-631a-430c-9ae8-e18fbcb439dc-utilities\") pod \"redhat-marketplace-24vzw\" (UID: \"f82a9b69-631a-430c-9ae8-e18fbcb439dc\") " pod="openshift-marketplace/redhat-marketplace-24vzw" Nov 25 15:27:26 crc kubenswrapper[4890]: I1125 15:27:26.329029 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f82a9b69-631a-430c-9ae8-e18fbcb439dc-catalog-content\") pod \"redhat-marketplace-24vzw\" (UID: \"f82a9b69-631a-430c-9ae8-e18fbcb439dc\") " pod="openshift-marketplace/redhat-marketplace-24vzw" Nov 25 15:27:26 crc kubenswrapper[4890]: I1125 15:27:26.329110 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgmd7\" (UniqueName: \"kubernetes.io/projected/f82a9b69-631a-430c-9ae8-e18fbcb439dc-kube-api-access-kgmd7\") pod \"redhat-marketplace-24vzw\" (UID: \"f82a9b69-631a-430c-9ae8-e18fbcb439dc\") " pod="openshift-marketplace/redhat-marketplace-24vzw" Nov 25 15:27:26 crc kubenswrapper[4890]: I1125 15:27:26.329230 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f82a9b69-631a-430c-9ae8-e18fbcb439dc-utilities\") pod \"redhat-marketplace-24vzw\" (UID: \"f82a9b69-631a-430c-9ae8-e18fbcb439dc\") " pod="openshift-marketplace/redhat-marketplace-24vzw" Nov 25 15:27:26 crc kubenswrapper[4890]: I1125 15:27:26.329677 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f82a9b69-631a-430c-9ae8-e18fbcb439dc-catalog-content\") pod \"redhat-marketplace-24vzw\" (UID: \"f82a9b69-631a-430c-9ae8-e18fbcb439dc\") " pod="openshift-marketplace/redhat-marketplace-24vzw" Nov 25 15:27:26 crc kubenswrapper[4890]: I1125 15:27:26.329715 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f82a9b69-631a-430c-9ae8-e18fbcb439dc-utilities\") pod \"redhat-marketplace-24vzw\" (UID: \"f82a9b69-631a-430c-9ae8-e18fbcb439dc\") " pod="openshift-marketplace/redhat-marketplace-24vzw" Nov 25 15:27:26 crc kubenswrapper[4890]: I1125 15:27:26.351614 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgmd7\" (UniqueName: \"kubernetes.io/projected/f82a9b69-631a-430c-9ae8-e18fbcb439dc-kube-api-access-kgmd7\") pod \"redhat-marketplace-24vzw\" (UID: \"f82a9b69-631a-430c-9ae8-e18fbcb439dc\") " pod="openshift-marketplace/redhat-marketplace-24vzw" Nov 25 15:27:26 crc kubenswrapper[4890]: I1125 15:27:26.447796 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 15:27:26 crc kubenswrapper[4890]: I1125 15:27:26.447868 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 15:27:26 crc kubenswrapper[4890]: I1125 15:27:26.514826 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-24vzw" Nov 25 15:27:26 crc kubenswrapper[4890]: I1125 15:27:26.528997 4890 generic.go:334] "Generic (PLEG): container finished" podID="56f8430f-3421-47db-b225-5c40fd14c8bd" containerID="41b14b144856e47ab31876e7cebe7e0578f5a19f3916527de1b9c742c0b771c1" exitCode=0 Nov 25 15:27:26 crc kubenswrapper[4890]: I1125 15:27:26.529062 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kqz9d" event={"ID":"56f8430f-3421-47db-b225-5c40fd14c8bd","Type":"ContainerDied","Data":"41b14b144856e47ab31876e7cebe7e0578f5a19f3916527de1b9c742c0b771c1"} Nov 25 15:27:34 crc kubenswrapper[4890]: E1125 15:27:34.275572 4890 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 41b14b144856e47ab31876e7cebe7e0578f5a19f3916527de1b9c742c0b771c1 is running failed: container process not found" containerID="41b14b144856e47ab31876e7cebe7e0578f5a19f3916527de1b9c742c0b771c1" cmd=["grpc_health_probe","-addr=:50051"] Nov 25 15:27:34 crc kubenswrapper[4890]: E1125 15:27:34.278011 4890 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 41b14b144856e47ab31876e7cebe7e0578f5a19f3916527de1b9c742c0b771c1 is running failed: container process not found" containerID="41b14b144856e47ab31876e7cebe7e0578f5a19f3916527de1b9c742c0b771c1" cmd=["grpc_health_probe","-addr=:50051"] Nov 25 15:27:34 crc kubenswrapper[4890]: E1125 15:27:34.278463 4890 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 41b14b144856e47ab31876e7cebe7e0578f5a19f3916527de1b9c742c0b771c1 is running failed: container process not found" containerID="41b14b144856e47ab31876e7cebe7e0578f5a19f3916527de1b9c742c0b771c1" cmd=["grpc_health_probe","-addr=:50051"] Nov 25 15:27:34 crc kubenswrapper[4890]: E1125 15:27:34.278547 4890 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 41b14b144856e47ab31876e7cebe7e0578f5a19f3916527de1b9c742c0b771c1 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-kqz9d" podUID="56f8430f-3421-47db-b225-5c40fd14c8bd" containerName="registry-server" Nov 25 15:27:44 crc kubenswrapper[4890]: E1125 15:27:44.274294 4890 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 41b14b144856e47ab31876e7cebe7e0578f5a19f3916527de1b9c742c0b771c1 is running failed: container process not found" containerID="41b14b144856e47ab31876e7cebe7e0578f5a19f3916527de1b9c742c0b771c1" cmd=["grpc_health_probe","-addr=:50051"] Nov 25 15:27:44 crc kubenswrapper[4890]: E1125 15:27:44.276128 4890 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 41b14b144856e47ab31876e7cebe7e0578f5a19f3916527de1b9c742c0b771c1 is running failed: container process not found" containerID="41b14b144856e47ab31876e7cebe7e0578f5a19f3916527de1b9c742c0b771c1" cmd=["grpc_health_probe","-addr=:50051"] Nov 25 15:27:44 crc kubenswrapper[4890]: E1125 15:27:44.276438 4890 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 41b14b144856e47ab31876e7cebe7e0578f5a19f3916527de1b9c742c0b771c1 is running failed: container process not found" containerID="41b14b144856e47ab31876e7cebe7e0578f5a19f3916527de1b9c742c0b771c1" cmd=["grpc_health_probe","-addr=:50051"] Nov 25 15:27:44 crc kubenswrapper[4890]: E1125 15:27:44.276517 4890 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 41b14b144856e47ab31876e7cebe7e0578f5a19f3916527de1b9c742c0b771c1 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-kqz9d" podUID="56f8430f-3421-47db-b225-5c40fd14c8bd" containerName="registry-server" Nov 25 15:27:46 crc kubenswrapper[4890]: I1125 15:27:46.398767 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kqz9d" Nov 25 15:27:46 crc kubenswrapper[4890]: I1125 15:27:46.445404 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56f8430f-3421-47db-b225-5c40fd14c8bd-catalog-content\") pod \"56f8430f-3421-47db-b225-5c40fd14c8bd\" (UID: \"56f8430f-3421-47db-b225-5c40fd14c8bd\") " Nov 25 15:27:46 crc kubenswrapper[4890]: I1125 15:27:46.445515 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56f8430f-3421-47db-b225-5c40fd14c8bd-utilities\") pod \"56f8430f-3421-47db-b225-5c40fd14c8bd\" (UID: \"56f8430f-3421-47db-b225-5c40fd14c8bd\") " Nov 25 15:27:46 crc kubenswrapper[4890]: I1125 15:27:46.445574 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dgt4c\" (UniqueName: \"kubernetes.io/projected/56f8430f-3421-47db-b225-5c40fd14c8bd-kube-api-access-dgt4c\") pod \"56f8430f-3421-47db-b225-5c40fd14c8bd\" (UID: \"56f8430f-3421-47db-b225-5c40fd14c8bd\") " Nov 25 15:27:46 crc kubenswrapper[4890]: I1125 15:27:46.446295 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56f8430f-3421-47db-b225-5c40fd14c8bd-utilities" (OuterVolumeSpecName: "utilities") pod "56f8430f-3421-47db-b225-5c40fd14c8bd" (UID: "56f8430f-3421-47db-b225-5c40fd14c8bd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:27:46 crc kubenswrapper[4890]: I1125 15:27:46.457302 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56f8430f-3421-47db-b225-5c40fd14c8bd-kube-api-access-dgt4c" (OuterVolumeSpecName: "kube-api-access-dgt4c") pod "56f8430f-3421-47db-b225-5c40fd14c8bd" (UID: "56f8430f-3421-47db-b225-5c40fd14c8bd"). InnerVolumeSpecName "kube-api-access-dgt4c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:27:46 crc kubenswrapper[4890]: I1125 15:27:46.495990 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56f8430f-3421-47db-b225-5c40fd14c8bd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "56f8430f-3421-47db-b225-5c40fd14c8bd" (UID: "56f8430f-3421-47db-b225-5c40fd14c8bd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:27:46 crc kubenswrapper[4890]: I1125 15:27:46.547728 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56f8430f-3421-47db-b225-5c40fd14c8bd-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 15:27:46 crc kubenswrapper[4890]: I1125 15:27:46.547762 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56f8430f-3421-47db-b225-5c40fd14c8bd-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 15:27:46 crc kubenswrapper[4890]: I1125 15:27:46.547771 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dgt4c\" (UniqueName: \"kubernetes.io/projected/56f8430f-3421-47db-b225-5c40fd14c8bd-kube-api-access-dgt4c\") on node \"crc\" DevicePath \"\"" Nov 25 15:27:46 crc kubenswrapper[4890]: I1125 15:27:46.718122 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kqz9d" event={"ID":"56f8430f-3421-47db-b225-5c40fd14c8bd","Type":"ContainerDied","Data":"9cd842b7acaffe38a41935af190486c9f26e28ac77eecb45b683e2602bd9dd90"} Nov 25 15:27:46 crc kubenswrapper[4890]: I1125 15:27:46.718193 4890 scope.go:117] "RemoveContainer" containerID="41b14b144856e47ab31876e7cebe7e0578f5a19f3916527de1b9c742c0b771c1" Nov 25 15:27:46 crc kubenswrapper[4890]: I1125 15:27:46.718354 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kqz9d" Nov 25 15:27:46 crc kubenswrapper[4890]: I1125 15:27:46.772347 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kqz9d"] Nov 25 15:27:46 crc kubenswrapper[4890]: I1125 15:27:46.783622 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kqz9d"] Nov 25 15:27:47 crc kubenswrapper[4890]: I1125 15:27:47.098306 4890 scope.go:117] "RemoveContainer" containerID="bd2aecab4c72b6d266fd711876be5a43e59971cc71bfadb5b3dee019f5c715c7" Nov 25 15:27:47 crc kubenswrapper[4890]: I1125 15:27:47.328285 4890 scope.go:117] "RemoveContainer" containerID="7d1f31d85be0bb337c001fcc38d4c57c070b912c36cb0b5146698612c8a597d7" Nov 25 15:27:47 crc kubenswrapper[4890]: I1125 15:27:47.605612 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f24bd"] Nov 25 15:27:47 crc kubenswrapper[4890]: I1125 15:27:47.673652 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-24vzw"] Nov 25 15:27:47 crc kubenswrapper[4890]: W1125 15:27:47.716814 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0f05c16a_1239_42a7_9910_ca03ef42b0f7.slice/crio-f087c6a953c0f73a4c176ffff375a048156a59224760eb2c0e7ab2046ec36214 WatchSource:0}: Error finding container f087c6a953c0f73a4c176ffff375a048156a59224760eb2c0e7ab2046ec36214: Status 404 returned error can't find the container with id f087c6a953c0f73a4c176ffff375a048156a59224760eb2c0e7ab2046ec36214 Nov 25 15:27:47 crc kubenswrapper[4890]: I1125 15:27:47.740985 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mtfgl"] Nov 25 15:27:47 crc kubenswrapper[4890]: W1125 15:27:47.784879 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf7235b7e_d6fe_45a2_93fd_66b0a77b4b76.slice/crio-2cbf9bfcbac76d87873600247a7a229961b2ceb24529a93268d7f3093f8a659c WatchSource:0}: Error finding container 2cbf9bfcbac76d87873600247a7a229961b2ceb24529a93268d7f3093f8a659c: Status 404 returned error can't find the container with id 2cbf9bfcbac76d87873600247a7a229961b2ceb24529a93268d7f3093f8a659c Nov 25 15:27:47 crc kubenswrapper[4890]: I1125 15:27:47.832113 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-48z2b"] Nov 25 15:27:47 crc kubenswrapper[4890]: I1125 15:27:47.843683 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v8mj2" event={"ID":"00771e9e-5583-42ab-b834-4643d9bbdbec","Type":"ContainerStarted","Data":"583629b7aeb99a0f6d9e57cd9965500d687d0958872fcdd86b78e86046b11f5a"} Nov 25 15:27:47 crc kubenswrapper[4890]: W1125 15:27:47.864243 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeacc1f05_701e_49af_b2d7_9c049f6cddb9.slice/crio-a044d2b6a7d48ac7eced8baea8581590025de47ee6b4fdb424a32c6bde3fa2d5 WatchSource:0}: Error finding container a044d2b6a7d48ac7eced8baea8581590025de47ee6b4fdb424a32c6bde3fa2d5: Status 404 returned error can't find the container with id a044d2b6a7d48ac7eced8baea8581590025de47ee6b4fdb424a32c6bde3fa2d5 Nov 25 15:27:47 crc kubenswrapper[4890]: I1125 15:27:47.890971 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kr7td" event={"ID":"cb551926-2f33-4616-a423-caee14e79f74","Type":"ContainerStarted","Data":"2a0ad6eb90cb3283291e18b1291d23c330c5079988eebcdbc75da3ecbe05dfd5"} Nov 25 15:27:47 crc kubenswrapper[4890]: I1125 15:27:47.908681 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gxfqw" event={"ID":"fc5b6ee3-033f-4b8f-82b7-46e28498f625","Type":"ContainerStarted","Data":"4883051ee5c35761b8e164c1e0358647f8b49a22917fe479ae5a47818d0e10c7"} Nov 25 15:27:47 crc kubenswrapper[4890]: I1125 15:27:47.941997 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gxfqw" podStartSLOduration=4.853471687 podStartE2EDuration="57.941974588s" podCreationTimestamp="2025-11-25 15:26:50 +0000 UTC" firstStartedPulling="2025-11-25 15:26:52.897580966 +0000 UTC m=+1471.340043576" lastFinishedPulling="2025-11-25 15:27:45.986083847 +0000 UTC m=+1524.428546477" observedRunningTime="2025-11-25 15:27:47.930529452 +0000 UTC m=+1526.372992062" watchObservedRunningTime="2025-11-25 15:27:47.941974588 +0000 UTC m=+1526.384437198" Nov 25 15:27:47 crc kubenswrapper[4890]: I1125 15:27:47.978403 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cr58b"] Nov 25 15:27:47 crc kubenswrapper[4890]: W1125 15:27:47.991278 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6d15a3f2_bbd4_43ae_b620_93251f74369e.slice/crio-07f2effed740c16bd10e19722336b98bdae3b19aeb07410d535393788f73462a WatchSource:0}: Error finding container 07f2effed740c16bd10e19722336b98bdae3b19aeb07410d535393788f73462a: Status 404 returned error can't find the container with id 07f2effed740c16bd10e19722336b98bdae3b19aeb07410d535393788f73462a Nov 25 15:27:47 crc kubenswrapper[4890]: I1125 15:27:47.996198 4890 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 15:27:48 crc kubenswrapper[4890]: I1125 15:27:48.018721 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4fdm7"] Nov 25 15:27:48 crc kubenswrapper[4890]: W1125 15:27:48.074335 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda0a34ff8_8257_4f49_90d2_90576cbc40b7.slice/crio-5cddf18bc8e2fcde0308a92cc51919688ad6bd00799fe1044ac649392b8446ac WatchSource:0}: Error finding container 5cddf18bc8e2fcde0308a92cc51919688ad6bd00799fe1044ac649392b8446ac: Status 404 returned error can't find the container with id 5cddf18bc8e2fcde0308a92cc51919688ad6bd00799fe1044ac649392b8446ac Nov 25 15:27:48 crc kubenswrapper[4890]: I1125 15:27:48.190519 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56f8430f-3421-47db-b225-5c40fd14c8bd" path="/var/lib/kubelet/pods/56f8430f-3421-47db-b225-5c40fd14c8bd/volumes" Nov 25 15:27:48 crc kubenswrapper[4890]: I1125 15:27:48.924294 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ddmtt" event={"ID":"b9fbb88d-f5f9-41c6-9719-15be794dfb6c","Type":"ContainerStarted","Data":"bb0dcecc09d90d56e793929595f4f4c4631c9363c8d877b4b159f0abbbf8d5fa"} Nov 25 15:27:48 crc kubenswrapper[4890]: I1125 15:27:48.926888 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-492nv" event={"ID":"1fcc307a-b17f-46ab-aabe-0a0361656701","Type":"ContainerStarted","Data":"5e92fd7ea83b5fb5a96a5febb629ff7a0ac952ad0684f3a95450d43595eb7d22"} Nov 25 15:27:48 crc kubenswrapper[4890]: I1125 15:27:48.933798 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-48z2b" event={"ID":"eacc1f05-701e-49af-b2d7-9c049f6cddb9","Type":"ContainerStarted","Data":"ae8f77220f5a568aeb2874fdf50f8002339e492b5a510906dd1e42b2d91d63a6"} Nov 25 15:27:48 crc kubenswrapper[4890]: I1125 15:27:48.933838 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-48z2b" event={"ID":"eacc1f05-701e-49af-b2d7-9c049f6cddb9","Type":"ContainerStarted","Data":"a044d2b6a7d48ac7eced8baea8581590025de47ee6b4fdb424a32c6bde3fa2d5"} Nov 25 15:27:48 crc kubenswrapper[4890]: I1125 15:27:48.936072 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6g727" event={"ID":"b53e3920-fe5e-45fe-9b2d-495a5cb16b2c","Type":"ContainerStarted","Data":"8b8951542a05f91dd43fa2161ca5c59912852a6eb9dcdc79d470c54c1dc552b8"} Nov 25 15:27:48 crc kubenswrapper[4890]: I1125 15:27:48.955781 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f24bd" event={"ID":"0f05c16a-1239-42a7-9910-ca03ef42b0f7","Type":"ContainerStarted","Data":"caf1c4a1acd3f56a5ecc9d123caeebca960888f6685246e0bf27ec17d5482207"} Nov 25 15:27:48 crc kubenswrapper[4890]: I1125 15:27:48.955825 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f24bd" event={"ID":"0f05c16a-1239-42a7-9910-ca03ef42b0f7","Type":"ContainerStarted","Data":"f087c6a953c0f73a4c176ffff375a048156a59224760eb2c0e7ab2046ec36214"} Nov 25 15:27:48 crc kubenswrapper[4890]: I1125 15:27:48.968643 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-492nv" podStartSLOduration=9.305803934 podStartE2EDuration="1m2.968619571s" podCreationTimestamp="2025-11-25 15:26:46 +0000 UTC" firstStartedPulling="2025-11-25 15:26:52.908508129 +0000 UTC m=+1471.350970739" lastFinishedPulling="2025-11-25 15:27:46.571323766 +0000 UTC m=+1525.013786376" observedRunningTime="2025-11-25 15:27:48.953015221 +0000 UTC m=+1527.395477831" watchObservedRunningTime="2025-11-25 15:27:48.968619571 +0000 UTC m=+1527.411082181" Nov 25 15:27:49 crc kubenswrapper[4890]: I1125 15:27:48.997549 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h44zl" event={"ID":"08bbb8ce-17b9-4797-80a7-de3beca270f4","Type":"ContainerStarted","Data":"35098fd0841e4cedb8a48c9b6dd5cc6d32370f0a2cb7b03e08538da936e45d55"} Nov 25 15:27:49 crc kubenswrapper[4890]: I1125 15:27:49.000280 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rcbhc" event={"ID":"b18eb2e9-77b4-46ba-9e32-31c1e9b73963","Type":"ContainerStarted","Data":"626d9b201a0e32fd3d13e65ebe08857bf56a1aa7841578b259565881cdd9fbdf"} Nov 25 15:27:49 crc kubenswrapper[4890]: I1125 15:27:49.015194 4890 generic.go:334] "Generic (PLEG): container finished" podID="f82a9b69-631a-430c-9ae8-e18fbcb439dc" containerID="0ddf005ec47d82c4ef1a06c2449c8ac9568e0aec70337c3a73b1e30e904e49cb" exitCode=0 Nov 25 15:27:49 crc kubenswrapper[4890]: I1125 15:27:49.015336 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-24vzw" event={"ID":"f82a9b69-631a-430c-9ae8-e18fbcb439dc","Type":"ContainerDied","Data":"0ddf005ec47d82c4ef1a06c2449c8ac9568e0aec70337c3a73b1e30e904e49cb"} Nov 25 15:27:49 crc kubenswrapper[4890]: I1125 15:27:49.015442 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-24vzw" event={"ID":"f82a9b69-631a-430c-9ae8-e18fbcb439dc","Type":"ContainerStarted","Data":"0ce789c448c94c35fa0e244f31949790e3d70b79cccf8f74a660d6089dec4e2e"} Nov 25 15:27:49 crc kubenswrapper[4890]: I1125 15:27:49.029944 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mtfgl" event={"ID":"f7235b7e-d6fe-45a2-93fd-66b0a77b4b76","Type":"ContainerStarted","Data":"f19ba65caf7a2838ce342808f2d9139f050a92071798091cb624ad53d054f55b"} Nov 25 15:27:49 crc kubenswrapper[4890]: I1125 15:27:49.030016 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mtfgl" event={"ID":"f7235b7e-d6fe-45a2-93fd-66b0a77b4b76","Type":"ContainerStarted","Data":"2cbf9bfcbac76d87873600247a7a229961b2ceb24529a93268d7f3093f8a659c"} Nov 25 15:27:49 crc kubenswrapper[4890]: I1125 15:27:49.064078 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cr58b" event={"ID":"6d15a3f2-bbd4-43ae-b620-93251f74369e","Type":"ContainerStarted","Data":"07f2effed740c16bd10e19722336b98bdae3b19aeb07410d535393788f73462a"} Nov 25 15:27:49 crc kubenswrapper[4890]: I1125 15:27:49.072119 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4fdm7" event={"ID":"a0a34ff8-8257-4f49-90d2-90576cbc40b7","Type":"ContainerStarted","Data":"5cddf18bc8e2fcde0308a92cc51919688ad6bd00799fe1044ac649392b8446ac"} Nov 25 15:27:49 crc kubenswrapper[4890]: I1125 15:27:49.102386 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g8wbt" event={"ID":"7c070ef9-daba-4ae9-ac05-dff449baf01a","Type":"ContainerStarted","Data":"74f159210a58c58af518d6ce062a98d2955b3a8f42e39be962750a76214a2f7f"} Nov 25 15:27:49 crc kubenswrapper[4890]: I1125 15:27:49.148904 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-v8mj2" podStartSLOduration=6.264181344 podStartE2EDuration="57.148879057s" podCreationTimestamp="2025-11-25 15:26:52 +0000 UTC" firstStartedPulling="2025-11-25 15:26:55.713417063 +0000 UTC m=+1474.155879673" lastFinishedPulling="2025-11-25 15:27:46.598114776 +0000 UTC m=+1525.040577386" observedRunningTime="2025-11-25 15:27:49.136867447 +0000 UTC m=+1527.579330057" watchObservedRunningTime="2025-11-25 15:27:49.148879057 +0000 UTC m=+1527.591341657" Nov 25 15:27:49 crc kubenswrapper[4890]: I1125 15:27:49.170133 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-g8wbt" podStartSLOduration=5.440105934 podStartE2EDuration="1m0.170117518s" podCreationTimestamp="2025-11-25 15:26:49 +0000 UTC" firstStartedPulling="2025-11-25 15:26:51.868088591 +0000 UTC m=+1470.310551201" lastFinishedPulling="2025-11-25 15:27:46.598100175 +0000 UTC m=+1525.040562785" observedRunningTime="2025-11-25 15:27:49.169325438 +0000 UTC m=+1527.611788058" watchObservedRunningTime="2025-11-25 15:27:49.170117518 +0000 UTC m=+1527.612580128" Nov 25 15:27:49 crc kubenswrapper[4890]: I1125 15:27:49.194049 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-kr7td" podStartSLOduration=41.086612742 podStartE2EDuration="1m1.194029686s" podCreationTimestamp="2025-11-25 15:26:48 +0000 UTC" firstStartedPulling="2025-11-25 15:26:51.86122957 +0000 UTC m=+1470.303692180" lastFinishedPulling="2025-11-25 15:27:11.968646514 +0000 UTC m=+1490.411109124" observedRunningTime="2025-11-25 15:27:49.19342352 +0000 UTC m=+1527.635886130" watchObservedRunningTime="2025-11-25 15:27:49.194029686 +0000 UTC m=+1527.636492296" Nov 25 15:27:49 crc kubenswrapper[4890]: I1125 15:27:49.699232 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-g8wbt" Nov 25 15:27:49 crc kubenswrapper[4890]: I1125 15:27:49.699280 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-g8wbt" Nov 25 15:27:50 crc kubenswrapper[4890]: I1125 15:27:50.112374 4890 generic.go:334] "Generic (PLEG): container finished" podID="f7235b7e-d6fe-45a2-93fd-66b0a77b4b76" containerID="f19ba65caf7a2838ce342808f2d9139f050a92071798091cb624ad53d054f55b" exitCode=0 Nov 25 15:27:50 crc kubenswrapper[4890]: I1125 15:27:50.112461 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mtfgl" event={"ID":"f7235b7e-d6fe-45a2-93fd-66b0a77b4b76","Type":"ContainerDied","Data":"f19ba65caf7a2838ce342808f2d9139f050a92071798091cb624ad53d054f55b"} Nov 25 15:27:50 crc kubenswrapper[4890]: I1125 15:27:50.115706 4890 generic.go:334] "Generic (PLEG): container finished" podID="eacc1f05-701e-49af-b2d7-9c049f6cddb9" containerID="ae8f77220f5a568aeb2874fdf50f8002339e492b5a510906dd1e42b2d91d63a6" exitCode=0 Nov 25 15:27:50 crc kubenswrapper[4890]: I1125 15:27:50.115795 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-48z2b" event={"ID":"eacc1f05-701e-49af-b2d7-9c049f6cddb9","Type":"ContainerDied","Data":"ae8f77220f5a568aeb2874fdf50f8002339e492b5a510906dd1e42b2d91d63a6"} Nov 25 15:27:50 crc kubenswrapper[4890]: I1125 15:27:50.117841 4890 generic.go:334] "Generic (PLEG): container finished" podID="a0a34ff8-8257-4f49-90d2-90576cbc40b7" containerID="33598a45238600ec7547cd57326233b2a1d69ba08ec0953da513c856306a0bc8" exitCode=0 Nov 25 15:27:50 crc kubenswrapper[4890]: I1125 15:27:50.117910 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4fdm7" event={"ID":"a0a34ff8-8257-4f49-90d2-90576cbc40b7","Type":"ContainerDied","Data":"33598a45238600ec7547cd57326233b2a1d69ba08ec0953da513c856306a0bc8"} Nov 25 15:27:50 crc kubenswrapper[4890]: I1125 15:27:50.120076 4890 generic.go:334] "Generic (PLEG): container finished" podID="0f05c16a-1239-42a7-9910-ca03ef42b0f7" containerID="caf1c4a1acd3f56a5ecc9d123caeebca960888f6685246e0bf27ec17d5482207" exitCode=0 Nov 25 15:27:50 crc kubenswrapper[4890]: I1125 15:27:50.120116 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f24bd" event={"ID":"0f05c16a-1239-42a7-9910-ca03ef42b0f7","Type":"ContainerDied","Data":"caf1c4a1acd3f56a5ecc9d123caeebca960888f6685246e0bf27ec17d5482207"} Nov 25 15:27:50 crc kubenswrapper[4890]: I1125 15:27:50.123189 4890 generic.go:334] "Generic (PLEG): container finished" podID="b18eb2e9-77b4-46ba-9e32-31c1e9b73963" containerID="626d9b201a0e32fd3d13e65ebe08857bf56a1aa7841578b259565881cdd9fbdf" exitCode=0 Nov 25 15:27:50 crc kubenswrapper[4890]: I1125 15:27:50.123259 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rcbhc" event={"ID":"b18eb2e9-77b4-46ba-9e32-31c1e9b73963","Type":"ContainerDied","Data":"626d9b201a0e32fd3d13e65ebe08857bf56a1aa7841578b259565881cdd9fbdf"} Nov 25 15:27:50 crc kubenswrapper[4890]: I1125 15:27:50.126250 4890 generic.go:334] "Generic (PLEG): container finished" podID="08bbb8ce-17b9-4797-80a7-de3beca270f4" containerID="35098fd0841e4cedb8a48c9b6dd5cc6d32370f0a2cb7b03e08538da936e45d55" exitCode=0 Nov 25 15:27:50 crc kubenswrapper[4890]: I1125 15:27:50.127279 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h44zl" event={"ID":"08bbb8ce-17b9-4797-80a7-de3beca270f4","Type":"ContainerDied","Data":"35098fd0841e4cedb8a48c9b6dd5cc6d32370f0a2cb7b03e08538da936e45d55"} Nov 25 15:27:50 crc kubenswrapper[4890]: I1125 15:27:50.181297 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ddmtt" podStartSLOduration=6.267635064 podStartE2EDuration="59.181279964s" podCreationTimestamp="2025-11-25 15:26:51 +0000 UTC" firstStartedPulling="2025-11-25 15:26:53.972843834 +0000 UTC m=+1472.415306454" lastFinishedPulling="2025-11-25 15:27:46.886488744 +0000 UTC m=+1525.328951354" observedRunningTime="2025-11-25 15:27:50.178936556 +0000 UTC m=+1528.621399186" watchObservedRunningTime="2025-11-25 15:27:50.181279964 +0000 UTC m=+1528.623742574" Nov 25 15:27:50 crc kubenswrapper[4890]: I1125 15:27:50.278896 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6g727" podStartSLOduration=10.342221824 podStartE2EDuration="1m5.278874264s" podCreationTimestamp="2025-11-25 15:26:45 +0000 UTC" firstStartedPulling="2025-11-25 15:26:51.867845495 +0000 UTC m=+1470.310308095" lastFinishedPulling="2025-11-25 15:27:46.804497915 +0000 UTC m=+1525.246960535" observedRunningTime="2025-11-25 15:27:50.26231729 +0000 UTC m=+1528.704779900" watchObservedRunningTime="2025-11-25 15:27:50.278874264 +0000 UTC m=+1528.721336874" Nov 25 15:27:50 crc kubenswrapper[4890]: I1125 15:27:50.746030 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-g8wbt" podUID="7c070ef9-daba-4ae9-ac05-dff449baf01a" containerName="registry-server" probeResult="failure" output=< Nov 25 15:27:50 crc kubenswrapper[4890]: timeout: failed to connect service ":50051" within 1s Nov 25 15:27:50 crc kubenswrapper[4890]: > Nov 25 15:27:50 crc kubenswrapper[4890]: I1125 15:27:50.899257 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gxfqw" Nov 25 15:27:50 crc kubenswrapper[4890]: I1125 15:27:50.899319 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gxfqw" Nov 25 15:27:50 crc kubenswrapper[4890]: I1125 15:27:50.984191 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gxfqw" Nov 25 15:27:52 crc kubenswrapper[4890]: I1125 15:27:52.103129 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ddmtt" Nov 25 15:27:52 crc kubenswrapper[4890]: I1125 15:27:52.104706 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ddmtt" Nov 25 15:27:52 crc kubenswrapper[4890]: I1125 15:27:52.160603 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ddmtt" Nov 25 15:27:53 crc kubenswrapper[4890]: I1125 15:27:53.212711 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ddmtt" Nov 25 15:27:53 crc kubenswrapper[4890]: I1125 15:27:53.318012 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-v8mj2" Nov 25 15:27:53 crc kubenswrapper[4890]: I1125 15:27:53.318566 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-v8mj2" Nov 25 15:27:53 crc kubenswrapper[4890]: I1125 15:27:53.376848 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-v8mj2" Nov 25 15:27:54 crc kubenswrapper[4890]: I1125 15:27:54.228556 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-v8mj2" Nov 25 15:27:56 crc kubenswrapper[4890]: I1125 15:27:56.108821 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6g727" Nov 25 15:27:56 crc kubenswrapper[4890]: I1125 15:27:56.109150 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6g727" Nov 25 15:27:56 crc kubenswrapper[4890]: I1125 15:27:56.160970 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6g727" Nov 25 15:27:56 crc kubenswrapper[4890]: I1125 15:27:56.251135 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6g727" Nov 25 15:27:56 crc kubenswrapper[4890]: I1125 15:27:56.447789 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 15:27:56 crc kubenswrapper[4890]: I1125 15:27:56.447870 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 15:27:56 crc kubenswrapper[4890]: I1125 15:27:56.447929 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" Nov 25 15:27:56 crc kubenswrapper[4890]: I1125 15:27:56.449062 4890 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9467aa8edc4d8bfe461dc02826750091f343018e37e7f269d305c77d21e832da"} pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 15:27:56 crc kubenswrapper[4890]: I1125 15:27:56.449140 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" containerID="cri-o://9467aa8edc4d8bfe461dc02826750091f343018e37e7f269d305c77d21e832da" gracePeriod=600 Nov 25 15:27:57 crc kubenswrapper[4890]: I1125 15:27:57.455894 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-492nv" Nov 25 15:27:57 crc kubenswrapper[4890]: I1125 15:27:57.456308 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-492nv" Nov 25 15:27:57 crc kubenswrapper[4890]: I1125 15:27:57.523944 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-492nv" Nov 25 15:27:58 crc kubenswrapper[4890]: I1125 15:27:58.231835 4890 generic.go:334] "Generic (PLEG): container finished" podID="4e4f849d-f239-4727-a73e-18327856929a" containerID="9467aa8edc4d8bfe461dc02826750091f343018e37e7f269d305c77d21e832da" exitCode=0 Nov 25 15:27:58 crc kubenswrapper[4890]: I1125 15:27:58.231889 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" event={"ID":"4e4f849d-f239-4727-a73e-18327856929a","Type":"ContainerDied","Data":"9467aa8edc4d8bfe461dc02826750091f343018e37e7f269d305c77d21e832da"} Nov 25 15:27:58 crc kubenswrapper[4890]: I1125 15:27:58.232333 4890 scope.go:117] "RemoveContainer" containerID="325b4a907d6b6c6f59e5653baa1b3273925ff29fed8f729e95af24a885f5e532" Nov 25 15:27:58 crc kubenswrapper[4890]: I1125 15:27:58.288984 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-492nv" Nov 25 15:27:58 crc kubenswrapper[4890]: E1125 15:27:58.391798 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:27:58 crc kubenswrapper[4890]: I1125 15:27:58.581391 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-kr7td" Nov 25 15:27:58 crc kubenswrapper[4890]: I1125 15:27:58.581455 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-kr7td" Nov 25 15:27:58 crc kubenswrapper[4890]: I1125 15:27:58.637698 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-kr7td" Nov 25 15:27:59 crc kubenswrapper[4890]: I1125 15:27:59.251614 4890 scope.go:117] "RemoveContainer" containerID="9467aa8edc4d8bfe461dc02826750091f343018e37e7f269d305c77d21e832da" Nov 25 15:27:59 crc kubenswrapper[4890]: E1125 15:27:59.251887 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:27:59 crc kubenswrapper[4890]: I1125 15:27:59.308754 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-kr7td" Nov 25 15:27:59 crc kubenswrapper[4890]: I1125 15:27:59.750789 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-g8wbt" Nov 25 15:27:59 crc kubenswrapper[4890]: I1125 15:27:59.816879 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-g8wbt" Nov 25 15:28:00 crc kubenswrapper[4890]: I1125 15:28:00.947774 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gxfqw" Nov 25 15:28:14 crc kubenswrapper[4890]: I1125 15:28:14.173737 4890 scope.go:117] "RemoveContainer" containerID="9467aa8edc4d8bfe461dc02826750091f343018e37e7f269d305c77d21e832da" Nov 25 15:28:14 crc kubenswrapper[4890]: E1125 15:28:14.174633 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:28:21 crc kubenswrapper[4890]: E1125 15:28:21.211026 4890 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest" Nov 25 15:28:21 crc kubenswrapper[4890]: E1125 15:28:21.211936 4890 kuberuntime_manager.go:1274] "Unhandled Error" err=< Nov 25 15:28:21 crc kubenswrapper[4890]: container &Container{Name:repo-setup-edpm-deployment-openstack-edpm-ipam,Image:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,Command:[],Args:[ansible-runner run /runner -p playbook.yaml -i repo-setup-edpm-deployment-openstack-edpm-ipam],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ANSIBLE_VERBOSITY,Value:2,ValueFrom:nil,},EnvVar{Name:RUNNER_PLAYBOOK,Value: Nov 25 15:28:21 crc kubenswrapper[4890]: - hosts: all Nov 25 15:28:21 crc kubenswrapper[4890]: strategy: linear Nov 25 15:28:21 crc kubenswrapper[4890]: tasks: Nov 25 15:28:21 crc kubenswrapper[4890]: - name: Enable podified-repos Nov 25 15:28:21 crc kubenswrapper[4890]: become: true Nov 25 15:28:21 crc kubenswrapper[4890]: ansible.builtin.shell: | Nov 25 15:28:21 crc kubenswrapper[4890]: set -euxo pipefail Nov 25 15:28:21 crc kubenswrapper[4890]: pushd /var/tmp Nov 25 15:28:21 crc kubenswrapper[4890]: curl -sL https://github.com/openstack-k8s-operators/repo-setup/archive/refs/heads/main.tar.gz | tar -xz Nov 25 15:28:21 crc kubenswrapper[4890]: pushd repo-setup-main Nov 25 15:28:21 crc kubenswrapper[4890]: python3 -m venv ./venv Nov 25 15:28:21 crc kubenswrapper[4890]: PBR_VERSION=0.0.0 ./venv/bin/pip install ./ Nov 25 15:28:21 crc kubenswrapper[4890]: ./venv/bin/repo-setup current-podified -b antelope Nov 25 15:28:21 crc kubenswrapper[4890]: popd Nov 25 15:28:21 crc kubenswrapper[4890]: rm -rf repo-setup-main Nov 25 15:28:21 crc kubenswrapper[4890]: Nov 25 15:28:21 crc kubenswrapper[4890]: Nov 25 15:28:21 crc kubenswrapper[4890]: ,ValueFrom:nil,},EnvVar{Name:RUNNER_EXTRA_VARS,Value: Nov 25 15:28:21 crc kubenswrapper[4890]: edpm_override_hosts: openstack-edpm-ipam Nov 25 15:28:21 crc kubenswrapper[4890]: edpm_service_type: repo-setup Nov 25 15:28:21 crc kubenswrapper[4890]: Nov 25 15:28:21 crc kubenswrapper[4890]: Nov 25 15:28:21 crc kubenswrapper[4890]: ,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:repo-setup-combined-ca-bundle,ReadOnly:false,MountPath:/var/lib/openstack/cacerts/repo-setup,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/runner/env/ssh_key,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:inventory,ReadOnly:false,MountPath:/runner/inventory/hosts,SubPath:inventory,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xq2st,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:openstack-aee-default-env,},Optional:*true,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod repo-setup-edpm-deployment-openstack-edpm-ipam-cr58b_openstack(6d15a3f2-bbd4-43ae-b620-93251f74369e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled Nov 25 15:28:21 crc kubenswrapper[4890]: > logger="UnhandledError" Nov 25 15:28:21 crc kubenswrapper[4890]: E1125 15:28:21.213187 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"repo-setup-edpm-deployment-openstack-edpm-ipam\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cr58b" podUID="6d15a3f2-bbd4-43ae-b620-93251f74369e" Nov 25 15:28:21 crc kubenswrapper[4890]: E1125 15:28:21.506627 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"repo-setup-edpm-deployment-openstack-edpm-ipam\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest\\\"\"" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cr58b" podUID="6d15a3f2-bbd4-43ae-b620-93251f74369e" Nov 25 15:28:22 crc kubenswrapper[4890]: I1125 15:28:22.508637 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h44zl" event={"ID":"08bbb8ce-17b9-4797-80a7-de3beca270f4","Type":"ContainerStarted","Data":"8b53a07dcabfab68bb15bca0ff4a64607a430f8cd9580588a484ad20985c4ef8"} Nov 25 15:28:22 crc kubenswrapper[4890]: I1125 15:28:22.510616 4890 generic.go:334] "Generic (PLEG): container finished" podID="f82a9b69-631a-430c-9ae8-e18fbcb439dc" containerID="03d70e127d1fe7385f358d33412a07d81173cb16045fec27408bf5c7cb2bc37c" exitCode=0 Nov 25 15:28:22 crc kubenswrapper[4890]: I1125 15:28:22.510716 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-24vzw" event={"ID":"f82a9b69-631a-430c-9ae8-e18fbcb439dc","Type":"ContainerDied","Data":"03d70e127d1fe7385f358d33412a07d81173cb16045fec27408bf5c7cb2bc37c"} Nov 25 15:28:22 crc kubenswrapper[4890]: I1125 15:28:22.515125 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mtfgl" event={"ID":"f7235b7e-d6fe-45a2-93fd-66b0a77b4b76","Type":"ContainerStarted","Data":"a96c1915aa5d9e95c5c7db8e3f90567de5b88e0a43de65bbeaea988efa55a8ff"} Nov 25 15:28:22 crc kubenswrapper[4890]: I1125 15:28:22.519333 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-48z2b" event={"ID":"eacc1f05-701e-49af-b2d7-9c049f6cddb9","Type":"ContainerStarted","Data":"bb4a6481afe58d7d9473c83486d0675ebc198deaa106003b6fe70e2bde92c9e3"} Nov 25 15:28:22 crc kubenswrapper[4890]: I1125 15:28:22.521777 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4fdm7" event={"ID":"a0a34ff8-8257-4f49-90d2-90576cbc40b7","Type":"ContainerStarted","Data":"6dcaca4001e4147931cd54f2beb3005bc92b25e287ced515f60a19e8b2585d90"} Nov 25 15:28:22 crc kubenswrapper[4890]: I1125 15:28:22.524674 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f24bd" event={"ID":"0f05c16a-1239-42a7-9910-ca03ef42b0f7","Type":"ContainerStarted","Data":"f7bca35abf69e8fefa3964fcbac0d5a3d7d365f424da842801a0424a4412311c"} Nov 25 15:28:22 crc kubenswrapper[4890]: I1125 15:28:22.541664 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rcbhc" event={"ID":"b18eb2e9-77b4-46ba-9e32-31c1e9b73963","Type":"ContainerStarted","Data":"a4ccb66699200fcad030dec7f35c81f1826dbfa3d9aede9a2806ab3b57bd0a09"} Nov 25 15:28:22 crc kubenswrapper[4890]: I1125 15:28:22.556944 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-h44zl" podStartSLOduration=7.363424088 podStartE2EDuration="1m28.556916425s" podCreationTimestamp="2025-11-25 15:26:54 +0000 UTC" firstStartedPulling="2025-11-25 15:27:00.219580884 +0000 UTC m=+1478.662043494" lastFinishedPulling="2025-11-25 15:28:21.413073221 +0000 UTC m=+1559.855535831" observedRunningTime="2025-11-25 15:28:22.540906854 +0000 UTC m=+1560.983369464" watchObservedRunningTime="2025-11-25 15:28:22.556916425 +0000 UTC m=+1560.999379035" Nov 25 15:28:22 crc kubenswrapper[4890]: I1125 15:28:22.662968 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rcbhc" podStartSLOduration=8.582879456 podStartE2EDuration="1m27.662945305s" podCreationTimestamp="2025-11-25 15:26:55 +0000 UTC" firstStartedPulling="2025-11-25 15:27:00.200940658 +0000 UTC m=+1478.643403268" lastFinishedPulling="2025-11-25 15:28:19.281006507 +0000 UTC m=+1557.723469117" observedRunningTime="2025-11-25 15:28:22.661805997 +0000 UTC m=+1561.104268617" watchObservedRunningTime="2025-11-25 15:28:22.662945305 +0000 UTC m=+1561.105407925" Nov 25 15:28:24 crc kubenswrapper[4890]: I1125 15:28:24.630895 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-h44zl" Nov 25 15:28:24 crc kubenswrapper[4890]: I1125 15:28:24.631224 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-h44zl" Nov 25 15:28:25 crc kubenswrapper[4890]: I1125 15:28:25.174049 4890 scope.go:117] "RemoveContainer" containerID="9467aa8edc4d8bfe461dc02826750091f343018e37e7f269d305c77d21e832da" Nov 25 15:28:25 crc kubenswrapper[4890]: E1125 15:28:25.174477 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:28:25 crc kubenswrapper[4890]: I1125 15:28:25.572518 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-24vzw" event={"ID":"f82a9b69-631a-430c-9ae8-e18fbcb439dc","Type":"ContainerStarted","Data":"b290a94eec07b5c615e67abcc269e9f1a88bedff111e406e0264b42709033774"} Nov 25 15:28:25 crc kubenswrapper[4890]: I1125 15:28:25.677409 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-h44zl" podUID="08bbb8ce-17b9-4797-80a7-de3beca270f4" containerName="registry-server" probeResult="failure" output=< Nov 25 15:28:25 crc kubenswrapper[4890]: timeout: failed to connect service ":50051" within 1s Nov 25 15:28:25 crc kubenswrapper[4890]: > Nov 25 15:28:25 crc kubenswrapper[4890]: I1125 15:28:25.723009 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rcbhc" Nov 25 15:28:25 crc kubenswrapper[4890]: I1125 15:28:25.723206 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rcbhc" Nov 25 15:28:25 crc kubenswrapper[4890]: I1125 15:28:25.775855 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rcbhc" Nov 25 15:28:26 crc kubenswrapper[4890]: I1125 15:28:26.608515 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-24vzw" podStartSLOduration=24.603307244 podStartE2EDuration="1m0.608491182s" podCreationTimestamp="2025-11-25 15:27:26 +0000 UTC" firstStartedPulling="2025-11-25 15:27:49.027357149 +0000 UTC m=+1527.469819759" lastFinishedPulling="2025-11-25 15:28:25.032541087 +0000 UTC m=+1563.475003697" observedRunningTime="2025-11-25 15:28:26.599697542 +0000 UTC m=+1565.042160162" watchObservedRunningTime="2025-11-25 15:28:26.608491182 +0000 UTC m=+1565.050953792" Nov 25 15:28:26 crc kubenswrapper[4890]: I1125 15:28:26.633696 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rcbhc" Nov 25 15:28:35 crc kubenswrapper[4890]: I1125 15:28:35.697126 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-h44zl" podUID="08bbb8ce-17b9-4797-80a7-de3beca270f4" containerName="registry-server" probeResult="failure" output=< Nov 25 15:28:35 crc kubenswrapper[4890]: timeout: failed to connect service ":50051" within 1s Nov 25 15:28:35 crc kubenswrapper[4890]: > Nov 25 15:28:36 crc kubenswrapper[4890]: I1125 15:28:36.516234 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-24vzw" Nov 25 15:28:36 crc kubenswrapper[4890]: I1125 15:28:36.516307 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-24vzw" Nov 25 15:28:36 crc kubenswrapper[4890]: I1125 15:28:36.561496 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-24vzw" Nov 25 15:28:36 crc kubenswrapper[4890]: I1125 15:28:36.733766 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-24vzw" Nov 25 15:28:36 crc kubenswrapper[4890]: I1125 15:28:36.804223 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-24vzw"] Nov 25 15:28:38 crc kubenswrapper[4890]: I1125 15:28:38.178755 4890 scope.go:117] "RemoveContainer" containerID="9467aa8edc4d8bfe461dc02826750091f343018e37e7f269d305c77d21e832da" Nov 25 15:28:38 crc kubenswrapper[4890]: E1125 15:28:38.179036 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:28:38 crc kubenswrapper[4890]: I1125 15:28:38.730188 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-24vzw" podUID="f82a9b69-631a-430c-9ae8-e18fbcb439dc" containerName="registry-server" containerID="cri-o://b290a94eec07b5c615e67abcc269e9f1a88bedff111e406e0264b42709033774" gracePeriod=2 Nov 25 15:28:41 crc kubenswrapper[4890]: I1125 15:28:41.763714 4890 generic.go:334] "Generic (PLEG): container finished" podID="f82a9b69-631a-430c-9ae8-e18fbcb439dc" containerID="b290a94eec07b5c615e67abcc269e9f1a88bedff111e406e0264b42709033774" exitCode=0 Nov 25 15:28:41 crc kubenswrapper[4890]: I1125 15:28:41.764261 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-24vzw" event={"ID":"f82a9b69-631a-430c-9ae8-e18fbcb439dc","Type":"ContainerDied","Data":"b290a94eec07b5c615e67abcc269e9f1a88bedff111e406e0264b42709033774"} Nov 25 15:28:45 crc kubenswrapper[4890]: I1125 15:28:45.687933 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-h44zl" podUID="08bbb8ce-17b9-4797-80a7-de3beca270f4" containerName="registry-server" probeResult="failure" output=< Nov 25 15:28:45 crc kubenswrapper[4890]: timeout: failed to connect service ":50051" within 1s Nov 25 15:28:45 crc kubenswrapper[4890]: > Nov 25 15:28:46 crc kubenswrapper[4890]: E1125 15:28:46.516788 4890 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b290a94eec07b5c615e67abcc269e9f1a88bedff111e406e0264b42709033774 is running failed: container process not found" containerID="b290a94eec07b5c615e67abcc269e9f1a88bedff111e406e0264b42709033774" cmd=["grpc_health_probe","-addr=:50051"] Nov 25 15:28:46 crc kubenswrapper[4890]: E1125 15:28:46.517399 4890 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b290a94eec07b5c615e67abcc269e9f1a88bedff111e406e0264b42709033774 is running failed: container process not found" containerID="b290a94eec07b5c615e67abcc269e9f1a88bedff111e406e0264b42709033774" cmd=["grpc_health_probe","-addr=:50051"] Nov 25 15:28:46 crc kubenswrapper[4890]: E1125 15:28:46.517849 4890 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b290a94eec07b5c615e67abcc269e9f1a88bedff111e406e0264b42709033774 is running failed: container process not found" containerID="b290a94eec07b5c615e67abcc269e9f1a88bedff111e406e0264b42709033774" cmd=["grpc_health_probe","-addr=:50051"] Nov 25 15:28:46 crc kubenswrapper[4890]: E1125 15:28:46.517932 4890 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b290a94eec07b5c615e67abcc269e9f1a88bedff111e406e0264b42709033774 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-24vzw" podUID="f82a9b69-631a-430c-9ae8-e18fbcb439dc" containerName="registry-server" Nov 25 15:28:49 crc kubenswrapper[4890]: I1125 15:28:49.677525 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-24vzw" Nov 25 15:28:49 crc kubenswrapper[4890]: I1125 15:28:49.744888 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f82a9b69-631a-430c-9ae8-e18fbcb439dc-catalog-content\") pod \"f82a9b69-631a-430c-9ae8-e18fbcb439dc\" (UID: \"f82a9b69-631a-430c-9ae8-e18fbcb439dc\") " Nov 25 15:28:49 crc kubenswrapper[4890]: I1125 15:28:49.745151 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kgmd7\" (UniqueName: \"kubernetes.io/projected/f82a9b69-631a-430c-9ae8-e18fbcb439dc-kube-api-access-kgmd7\") pod \"f82a9b69-631a-430c-9ae8-e18fbcb439dc\" (UID: \"f82a9b69-631a-430c-9ae8-e18fbcb439dc\") " Nov 25 15:28:49 crc kubenswrapper[4890]: I1125 15:28:49.745309 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f82a9b69-631a-430c-9ae8-e18fbcb439dc-utilities\") pod \"f82a9b69-631a-430c-9ae8-e18fbcb439dc\" (UID: \"f82a9b69-631a-430c-9ae8-e18fbcb439dc\") " Nov 25 15:28:49 crc kubenswrapper[4890]: I1125 15:28:49.747093 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f82a9b69-631a-430c-9ae8-e18fbcb439dc-utilities" (OuterVolumeSpecName: "utilities") pod "f82a9b69-631a-430c-9ae8-e18fbcb439dc" (UID: "f82a9b69-631a-430c-9ae8-e18fbcb439dc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:28:49 crc kubenswrapper[4890]: I1125 15:28:49.754440 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f82a9b69-631a-430c-9ae8-e18fbcb439dc-kube-api-access-kgmd7" (OuterVolumeSpecName: "kube-api-access-kgmd7") pod "f82a9b69-631a-430c-9ae8-e18fbcb439dc" (UID: "f82a9b69-631a-430c-9ae8-e18fbcb439dc"). InnerVolumeSpecName "kube-api-access-kgmd7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:28:49 crc kubenswrapper[4890]: I1125 15:28:49.772447 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f82a9b69-631a-430c-9ae8-e18fbcb439dc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f82a9b69-631a-430c-9ae8-e18fbcb439dc" (UID: "f82a9b69-631a-430c-9ae8-e18fbcb439dc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:28:49 crc kubenswrapper[4890]: I1125 15:28:49.848360 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f82a9b69-631a-430c-9ae8-e18fbcb439dc-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 15:28:49 crc kubenswrapper[4890]: I1125 15:28:49.848398 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kgmd7\" (UniqueName: \"kubernetes.io/projected/f82a9b69-631a-430c-9ae8-e18fbcb439dc-kube-api-access-kgmd7\") on node \"crc\" DevicePath \"\"" Nov 25 15:28:49 crc kubenswrapper[4890]: I1125 15:28:49.848410 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f82a9b69-631a-430c-9ae8-e18fbcb439dc-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 15:28:49 crc kubenswrapper[4890]: I1125 15:28:49.869282 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-24vzw" event={"ID":"f82a9b69-631a-430c-9ae8-e18fbcb439dc","Type":"ContainerDied","Data":"0ce789c448c94c35fa0e244f31949790e3d70b79cccf8f74a660d6089dec4e2e"} Nov 25 15:28:49 crc kubenswrapper[4890]: I1125 15:28:49.869356 4890 scope.go:117] "RemoveContainer" containerID="b290a94eec07b5c615e67abcc269e9f1a88bedff111e406e0264b42709033774" Nov 25 15:28:49 crc kubenswrapper[4890]: I1125 15:28:49.869434 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-24vzw" Nov 25 15:28:49 crc kubenswrapper[4890]: I1125 15:28:49.936302 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-24vzw"] Nov 25 15:28:49 crc kubenswrapper[4890]: I1125 15:28:49.951074 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-24vzw"] Nov 25 15:28:50 crc kubenswrapper[4890]: I1125 15:28:50.185916 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f82a9b69-631a-430c-9ae8-e18fbcb439dc" path="/var/lib/kubelet/pods/f82a9b69-631a-430c-9ae8-e18fbcb439dc/volumes" Nov 25 15:28:53 crc kubenswrapper[4890]: I1125 15:28:53.074671 4890 scope.go:117] "RemoveContainer" containerID="03d70e127d1fe7385f358d33412a07d81173cb16045fec27408bf5c7cb2bc37c" Nov 25 15:28:53 crc kubenswrapper[4890]: I1125 15:28:53.079281 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 15:28:53 crc kubenswrapper[4890]: I1125 15:28:53.130412 4890 scope.go:117] "RemoveContainer" containerID="0ddf005ec47d82c4ef1a06c2449c8ac9568e0aec70337c3a73b1e30e904e49cb" Nov 25 15:28:53 crc kubenswrapper[4890]: I1125 15:28:53.173924 4890 scope.go:117] "RemoveContainer" containerID="9467aa8edc4d8bfe461dc02826750091f343018e37e7f269d305c77d21e832da" Nov 25 15:28:53 crc kubenswrapper[4890]: E1125 15:28:53.174411 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:28:54 crc kubenswrapper[4890]: I1125 15:28:54.937779 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cr58b" event={"ID":"6d15a3f2-bbd4-43ae-b620-93251f74369e","Type":"ContainerStarted","Data":"b79ae5ca12f341a420e7388f25e996104f0aeac321ce10a2b6fc205e67fe894a"} Nov 25 15:28:55 crc kubenswrapper[4890]: I1125 15:28:55.674173 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-h44zl" podUID="08bbb8ce-17b9-4797-80a7-de3beca270f4" containerName="registry-server" probeResult="failure" output=< Nov 25 15:28:55 crc kubenswrapper[4890]: timeout: failed to connect service ":50051" within 1s Nov 25 15:28:55 crc kubenswrapper[4890]: > Nov 25 15:28:55 crc kubenswrapper[4890]: I1125 15:28:55.981115 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cr58b" podStartSLOduration=36.902143843 podStartE2EDuration="1m41.981096817s" podCreationTimestamp="2025-11-25 15:27:14 +0000 UTC" firstStartedPulling="2025-11-25 15:27:47.995949637 +0000 UTC m=+1526.438412237" lastFinishedPulling="2025-11-25 15:28:53.074902611 +0000 UTC m=+1591.517365211" observedRunningTime="2025-11-25 15:28:55.964754478 +0000 UTC m=+1594.407217098" watchObservedRunningTime="2025-11-25 15:28:55.981096817 +0000 UTC m=+1594.423559427" Nov 25 15:29:05 crc kubenswrapper[4890]: I1125 15:29:05.674386 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-h44zl" podUID="08bbb8ce-17b9-4797-80a7-de3beca270f4" containerName="registry-server" probeResult="failure" output=< Nov 25 15:29:05 crc kubenswrapper[4890]: timeout: failed to connect service ":50051" within 1s Nov 25 15:29:05 crc kubenswrapper[4890]: > Nov 25 15:29:06 crc kubenswrapper[4890]: I1125 15:29:06.174311 4890 scope.go:117] "RemoveContainer" containerID="9467aa8edc4d8bfe461dc02826750091f343018e37e7f269d305c77d21e832da" Nov 25 15:29:06 crc kubenswrapper[4890]: E1125 15:29:06.174715 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:29:15 crc kubenswrapper[4890]: I1125 15:29:15.077021 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-4ppjx"] Nov 25 15:29:15 crc kubenswrapper[4890]: I1125 15:29:15.091537 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-aa80-account-create-hcvwx"] Nov 25 15:29:15 crc kubenswrapper[4890]: I1125 15:29:15.103062 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-w9c2z"] Nov 25 15:29:15 crc kubenswrapper[4890]: I1125 15:29:15.113063 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-0806-account-create-xr7r5"] Nov 25 15:29:15 crc kubenswrapper[4890]: I1125 15:29:15.121919 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-0806-account-create-xr7r5"] Nov 25 15:29:15 crc kubenswrapper[4890]: I1125 15:29:15.130499 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-aa80-account-create-hcvwx"] Nov 25 15:29:15 crc kubenswrapper[4890]: I1125 15:29:15.140290 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-4ppjx"] Nov 25 15:29:15 crc kubenswrapper[4890]: I1125 15:29:15.153785 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-w9c2z"] Nov 25 15:29:15 crc kubenswrapper[4890]: I1125 15:29:15.678316 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-h44zl" podUID="08bbb8ce-17b9-4797-80a7-de3beca270f4" containerName="registry-server" probeResult="failure" output=< Nov 25 15:29:15 crc kubenswrapper[4890]: timeout: failed to connect service ":50051" within 1s Nov 25 15:29:15 crc kubenswrapper[4890]: > Nov 25 15:29:16 crc kubenswrapper[4890]: I1125 15:29:16.184880 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3af7c86b-336f-48ce-b5bf-294c0b8bc5ee" path="/var/lib/kubelet/pods/3af7c86b-336f-48ce-b5bf-294c0b8bc5ee/volumes" Nov 25 15:29:16 crc kubenswrapper[4890]: I1125 15:29:16.185758 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9076492c-2832-41b8-966c-f0ca12a43792" path="/var/lib/kubelet/pods/9076492c-2832-41b8-966c-f0ca12a43792/volumes" Nov 25 15:29:16 crc kubenswrapper[4890]: I1125 15:29:16.186382 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b69893e0-19f6-425b-b36f-b39fd2fd80d5" path="/var/lib/kubelet/pods/b69893e0-19f6-425b-b36f-b39fd2fd80d5/volumes" Nov 25 15:29:16 crc kubenswrapper[4890]: I1125 15:29:16.187039 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd3ad34d-4081-4341-b103-20d6ba8da9c5" path="/var/lib/kubelet/pods/dd3ad34d-4081-4341-b103-20d6ba8da9c5/volumes" Nov 25 15:29:19 crc kubenswrapper[4890]: I1125 15:29:19.029756 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-sppf8"] Nov 25 15:29:19 crc kubenswrapper[4890]: I1125 15:29:19.039877 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-sppf8"] Nov 25 15:29:20 crc kubenswrapper[4890]: I1125 15:29:20.042543 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-aca8-account-create-gm6cf"] Nov 25 15:29:20 crc kubenswrapper[4890]: I1125 15:29:20.058716 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-cf3e-account-create-4p2kd"] Nov 25 15:29:20 crc kubenswrapper[4890]: I1125 15:29:20.075690 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-aca8-account-create-gm6cf"] Nov 25 15:29:20 crc kubenswrapper[4890]: I1125 15:29:20.087743 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-cf3e-account-create-4p2kd"] Nov 25 15:29:20 crc kubenswrapper[4890]: I1125 15:29:20.174379 4890 scope.go:117] "RemoveContainer" containerID="9467aa8edc4d8bfe461dc02826750091f343018e37e7f269d305c77d21e832da" Nov 25 15:29:20 crc kubenswrapper[4890]: E1125 15:29:20.174801 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:29:20 crc kubenswrapper[4890]: I1125 15:29:20.193194 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="290df6b3-1433-4137-87e7-5caeba2bbb59" path="/var/lib/kubelet/pods/290df6b3-1433-4137-87e7-5caeba2bbb59/volumes" Nov 25 15:29:20 crc kubenswrapper[4890]: I1125 15:29:20.194426 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b59f5530-3d09-4573-b6f0-beb315e71d79" path="/var/lib/kubelet/pods/b59f5530-3d09-4573-b6f0-beb315e71d79/volumes" Nov 25 15:29:20 crc kubenswrapper[4890]: I1125 15:29:20.195266 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5d95f8e-951b-402b-afe8-ec8b3d090a45" path="/var/lib/kubelet/pods/c5d95f8e-951b-402b-afe8-ec8b3d090a45/volumes" Nov 25 15:29:21 crc kubenswrapper[4890]: I1125 15:29:21.052661 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-90a3-account-create-qvq8c"] Nov 25 15:29:21 crc kubenswrapper[4890]: I1125 15:29:21.067176 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-9354-account-create-5fdrf"] Nov 25 15:29:21 crc kubenswrapper[4890]: I1125 15:29:21.084268 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-9r7gm"] Nov 25 15:29:21 crc kubenswrapper[4890]: I1125 15:29:21.096424 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-xtxvs"] Nov 25 15:29:21 crc kubenswrapper[4890]: I1125 15:29:21.107438 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-hc7pc"] Nov 25 15:29:21 crc kubenswrapper[4890]: I1125 15:29:21.118618 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-xtxvs"] Nov 25 15:29:21 crc kubenswrapper[4890]: I1125 15:29:21.126855 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-9354-account-create-5fdrf"] Nov 25 15:29:21 crc kubenswrapper[4890]: I1125 15:29:21.136121 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-90a3-account-create-qvq8c"] Nov 25 15:29:21 crc kubenswrapper[4890]: I1125 15:29:21.145916 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-9r7gm"] Nov 25 15:29:21 crc kubenswrapper[4890]: I1125 15:29:21.155404 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-hc7pc"] Nov 25 15:29:22 crc kubenswrapper[4890]: I1125 15:29:22.185423 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c14ab22-4e5a-43e6-be42-956608ab877b" path="/var/lib/kubelet/pods/2c14ab22-4e5a-43e6-be42-956608ab877b/volumes" Nov 25 15:29:22 crc kubenswrapper[4890]: I1125 15:29:22.186585 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d380d3e-4b4c-424b-b9af-fb34e959e81c" path="/var/lib/kubelet/pods/2d380d3e-4b4c-424b-b9af-fb34e959e81c/volumes" Nov 25 15:29:22 crc kubenswrapper[4890]: I1125 15:29:22.187380 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8cd74cd-df01-4e7b-a8a9-391d5afbe2f9" path="/var/lib/kubelet/pods/a8cd74cd-df01-4e7b-a8a9-391d5afbe2f9/volumes" Nov 25 15:29:22 crc kubenswrapper[4890]: I1125 15:29:22.188139 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aba9e84f-bda0-4c6f-883e-ea9106b2521f" path="/var/lib/kubelet/pods/aba9e84f-bda0-4c6f-883e-ea9106b2521f/volumes" Nov 25 15:29:22 crc kubenswrapper[4890]: I1125 15:29:22.189413 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="acfb093b-f3af-4856-94b8-64a43c9c0cc1" path="/var/lib/kubelet/pods/acfb093b-f3af-4856-94b8-64a43c9c0cc1/volumes" Nov 25 15:29:24 crc kubenswrapper[4890]: I1125 15:29:24.263872 4890 generic.go:334] "Generic (PLEG): container finished" podID="eacc1f05-701e-49af-b2d7-9c049f6cddb9" containerID="bb4a6481afe58d7d9473c83486d0675ebc198deaa106003b6fe70e2bde92c9e3" exitCode=0 Nov 25 15:29:24 crc kubenswrapper[4890]: I1125 15:29:24.263960 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-48z2b" event={"ID":"eacc1f05-701e-49af-b2d7-9c049f6cddb9","Type":"ContainerDied","Data":"bb4a6481afe58d7d9473c83486d0675ebc198deaa106003b6fe70e2bde92c9e3"} Nov 25 15:29:24 crc kubenswrapper[4890]: I1125 15:29:24.275474 4890 generic.go:334] "Generic (PLEG): container finished" podID="a0a34ff8-8257-4f49-90d2-90576cbc40b7" containerID="6dcaca4001e4147931cd54f2beb3005bc92b25e287ced515f60a19e8b2585d90" exitCode=0 Nov 25 15:29:24 crc kubenswrapper[4890]: I1125 15:29:24.275567 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4fdm7" event={"ID":"a0a34ff8-8257-4f49-90d2-90576cbc40b7","Type":"ContainerDied","Data":"6dcaca4001e4147931cd54f2beb3005bc92b25e287ced515f60a19e8b2585d90"} Nov 25 15:29:24 crc kubenswrapper[4890]: I1125 15:29:24.286862 4890 generic.go:334] "Generic (PLEG): container finished" podID="0f05c16a-1239-42a7-9910-ca03ef42b0f7" containerID="f7bca35abf69e8fefa3964fcbac0d5a3d7d365f424da842801a0424a4412311c" exitCode=0 Nov 25 15:29:24 crc kubenswrapper[4890]: I1125 15:29:24.286961 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f24bd" event={"ID":"0f05c16a-1239-42a7-9910-ca03ef42b0f7","Type":"ContainerDied","Data":"f7bca35abf69e8fefa3964fcbac0d5a3d7d365f424da842801a0424a4412311c"} Nov 25 15:29:24 crc kubenswrapper[4890]: I1125 15:29:24.302150 4890 generic.go:334] "Generic (PLEG): container finished" podID="f7235b7e-d6fe-45a2-93fd-66b0a77b4b76" containerID="a96c1915aa5d9e95c5c7db8e3f90567de5b88e0a43de65bbeaea988efa55a8ff" exitCode=0 Nov 25 15:29:24 crc kubenswrapper[4890]: I1125 15:29:24.302219 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mtfgl" event={"ID":"f7235b7e-d6fe-45a2-93fd-66b0a77b4b76","Type":"ContainerDied","Data":"a96c1915aa5d9e95c5c7db8e3f90567de5b88e0a43de65bbeaea988efa55a8ff"} Nov 25 15:29:25 crc kubenswrapper[4890]: I1125 15:29:25.323361 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-48z2b" event={"ID":"eacc1f05-701e-49af-b2d7-9c049f6cddb9","Type":"ContainerStarted","Data":"f0e93c436975a4beaa31027e8591a496e2b0dbd7c8eab3144c79f7faa3188393"} Nov 25 15:29:25 crc kubenswrapper[4890]: I1125 15:29:25.352333 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-48z2b" podStartSLOduration=33.685626442 podStartE2EDuration="2m8.352313477s" podCreationTimestamp="2025-11-25 15:27:17 +0000 UTC" firstStartedPulling="2025-11-25 15:27:50.116827943 +0000 UTC m=+1528.559290553" lastFinishedPulling="2025-11-25 15:29:24.783514978 +0000 UTC m=+1623.225977588" observedRunningTime="2025-11-25 15:29:25.342140362 +0000 UTC m=+1623.784602972" watchObservedRunningTime="2025-11-25 15:29:25.352313477 +0000 UTC m=+1623.794776087" Nov 25 15:29:25 crc kubenswrapper[4890]: I1125 15:29:25.680753 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-h44zl" podUID="08bbb8ce-17b9-4797-80a7-de3beca270f4" containerName="registry-server" probeResult="failure" output=< Nov 25 15:29:25 crc kubenswrapper[4890]: timeout: failed to connect service ":50051" within 1s Nov 25 15:29:25 crc kubenswrapper[4890]: > Nov 25 15:29:26 crc kubenswrapper[4890]: I1125 15:29:26.359069 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4fdm7" event={"ID":"a0a34ff8-8257-4f49-90d2-90576cbc40b7","Type":"ContainerStarted","Data":"d55b3ccd8fde323cdc0d0587c0d3ffccd8748e2255ed40d3230a9a5404369928"} Nov 25 15:29:26 crc kubenswrapper[4890]: I1125 15:29:26.364182 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f24bd" event={"ID":"0f05c16a-1239-42a7-9910-ca03ef42b0f7","Type":"ContainerStarted","Data":"e4d7e3c3f0c96c99da055898bc7727760b326d9a7521ca55b599f4f9f9321a47"} Nov 25 15:29:26 crc kubenswrapper[4890]: I1125 15:29:26.367248 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mtfgl" event={"ID":"f7235b7e-d6fe-45a2-93fd-66b0a77b4b76","Type":"ContainerStarted","Data":"d78cdc34e1dfac4d6751e111e5e2a02f35cd675570357a84f0a9b5d9892b9881"} Nov 25 15:29:26 crc kubenswrapper[4890]: I1125 15:29:26.388043 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4fdm7" podStartSLOduration=32.531412824 podStartE2EDuration="2m7.388023697s" podCreationTimestamp="2025-11-25 15:27:19 +0000 UTC" firstStartedPulling="2025-11-25 15:27:50.119389247 +0000 UTC m=+1528.561851857" lastFinishedPulling="2025-11-25 15:29:24.97600012 +0000 UTC m=+1623.418462730" observedRunningTime="2025-11-25 15:29:26.384758005 +0000 UTC m=+1624.827220615" watchObservedRunningTime="2025-11-25 15:29:26.388023697 +0000 UTC m=+1624.830486307" Nov 25 15:29:26 crc kubenswrapper[4890]: I1125 15:29:26.405073 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mtfgl" podStartSLOduration=33.654003428 podStartE2EDuration="2m8.405020371s" podCreationTimestamp="2025-11-25 15:27:18 +0000 UTC" firstStartedPulling="2025-11-25 15:27:50.114724251 +0000 UTC m=+1528.557186861" lastFinishedPulling="2025-11-25 15:29:24.865741194 +0000 UTC m=+1623.308203804" observedRunningTime="2025-11-25 15:29:26.40337329 +0000 UTC m=+1624.845835910" watchObservedRunningTime="2025-11-25 15:29:26.405020371 +0000 UTC m=+1624.847482981" Nov 25 15:29:26 crc kubenswrapper[4890]: I1125 15:29:26.425836 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-f24bd" podStartSLOduration=30.717643059 podStartE2EDuration="2m5.425812531s" podCreationTimestamp="2025-11-25 15:27:21 +0000 UTC" firstStartedPulling="2025-11-25 15:27:50.122182707 +0000 UTC m=+1528.564645327" lastFinishedPulling="2025-11-25 15:29:24.830352189 +0000 UTC m=+1623.272814799" observedRunningTime="2025-11-25 15:29:26.422294663 +0000 UTC m=+1624.864757283" watchObservedRunningTime="2025-11-25 15:29:26.425812531 +0000 UTC m=+1624.868275151" Nov 25 15:29:28 crc kubenswrapper[4890]: I1125 15:29:28.185876 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-48z2b" Nov 25 15:29:28 crc kubenswrapper[4890]: I1125 15:29:28.186347 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-48z2b" Nov 25 15:29:29 crc kubenswrapper[4890]: I1125 15:29:29.112880 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mtfgl" Nov 25 15:29:29 crc kubenswrapper[4890]: I1125 15:29:29.112926 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mtfgl" Nov 25 15:29:29 crc kubenswrapper[4890]: I1125 15:29:29.232984 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-48z2b" podUID="eacc1f05-701e-49af-b2d7-9c049f6cddb9" containerName="registry-server" probeResult="failure" output=< Nov 25 15:29:29 crc kubenswrapper[4890]: timeout: failed to connect service ":50051" within 1s Nov 25 15:29:29 crc kubenswrapper[4890]: > Nov 25 15:29:30 crc kubenswrapper[4890]: I1125 15:29:30.169808 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mtfgl" podUID="f7235b7e-d6fe-45a2-93fd-66b0a77b4b76" containerName="registry-server" probeResult="failure" output=< Nov 25 15:29:30 crc kubenswrapper[4890]: timeout: failed to connect service ":50051" within 1s Nov 25 15:29:30 crc kubenswrapper[4890]: > Nov 25 15:29:30 crc kubenswrapper[4890]: I1125 15:29:30.310020 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4fdm7" Nov 25 15:29:30 crc kubenswrapper[4890]: I1125 15:29:30.310071 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4fdm7" Nov 25 15:29:31 crc kubenswrapper[4890]: I1125 15:29:31.356312 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-4fdm7" podUID="a0a34ff8-8257-4f49-90d2-90576cbc40b7" containerName="registry-server" probeResult="failure" output=< Nov 25 15:29:31 crc kubenswrapper[4890]: timeout: failed to connect service ":50051" within 1s Nov 25 15:29:31 crc kubenswrapper[4890]: > Nov 25 15:29:31 crc kubenswrapper[4890]: I1125 15:29:31.501200 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-f24bd" Nov 25 15:29:31 crc kubenswrapper[4890]: I1125 15:29:31.501265 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-f24bd" Nov 25 15:29:32 crc kubenswrapper[4890]: I1125 15:29:32.547272 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-f24bd" podUID="0f05c16a-1239-42a7-9910-ca03ef42b0f7" containerName="registry-server" probeResult="failure" output=< Nov 25 15:29:32 crc kubenswrapper[4890]: timeout: failed to connect service ":50051" within 1s Nov 25 15:29:32 crc kubenswrapper[4890]: > Nov 25 15:29:33 crc kubenswrapper[4890]: I1125 15:29:33.173864 4890 scope.go:117] "RemoveContainer" containerID="9467aa8edc4d8bfe461dc02826750091f343018e37e7f269d305c77d21e832da" Nov 25 15:29:33 crc kubenswrapper[4890]: E1125 15:29:33.174174 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:29:35 crc kubenswrapper[4890]: I1125 15:29:35.685403 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-h44zl" podUID="08bbb8ce-17b9-4797-80a7-de3beca270f4" containerName="registry-server" probeResult="failure" output=< Nov 25 15:29:35 crc kubenswrapper[4890]: timeout: failed to connect service ":50051" within 1s Nov 25 15:29:35 crc kubenswrapper[4890]: > Nov 25 15:29:36 crc kubenswrapper[4890]: I1125 15:29:36.433800 4890 scope.go:117] "RemoveContainer" containerID="708d473f40456f828711ffe0bd476d3b270c68b158df27ec57944565e5c99d1f" Nov 25 15:29:36 crc kubenswrapper[4890]: I1125 15:29:36.491850 4890 scope.go:117] "RemoveContainer" containerID="3004fbf5a7a3bea35404dc2c53d85fe29655581adbfbcf65d2feead1f3d308ec" Nov 25 15:29:36 crc kubenswrapper[4890]: I1125 15:29:36.535137 4890 scope.go:117] "RemoveContainer" containerID="6788492c06d96c69fc86d50b4aa2c4f3788e2e10ff1a313014a37ac06c72e219" Nov 25 15:29:36 crc kubenswrapper[4890]: I1125 15:29:36.581091 4890 scope.go:117] "RemoveContainer" containerID="b0ef80b34e3b82d11514b90037a50d0155f4666970674afded415b4192f4e7bb" Nov 25 15:29:36 crc kubenswrapper[4890]: I1125 15:29:36.635042 4890 scope.go:117] "RemoveContainer" containerID="83d8df7f74e56f66d558ce883b578ee8d08a0a14374ce74454a0337e40f72975" Nov 25 15:29:36 crc kubenswrapper[4890]: I1125 15:29:36.683786 4890 scope.go:117] "RemoveContainer" containerID="3936a5f46f158adeb4f89a31e4321cabf8b85456daa4c610139731932f1bd706" Nov 25 15:29:36 crc kubenswrapper[4890]: I1125 15:29:36.761708 4890 scope.go:117] "RemoveContainer" containerID="5e84c82890d994cc8442e7f7ca92e33dce94970d4afefe933ecc2285c4373eb4" Nov 25 15:29:36 crc kubenswrapper[4890]: I1125 15:29:36.790365 4890 scope.go:117] "RemoveContainer" containerID="dd6d428a2188a9073530062c1a18cd685809e99a554254915e6b1d0f3fbdac6e" Nov 25 15:29:36 crc kubenswrapper[4890]: I1125 15:29:36.829090 4890 scope.go:117] "RemoveContainer" containerID="0e455866be4285b7152b5e4083233f26e58160d23d57cf58741003d8dd3e68ac" Nov 25 15:29:36 crc kubenswrapper[4890]: I1125 15:29:36.864937 4890 scope.go:117] "RemoveContainer" containerID="5a21a62b5c7af01f6ed9932f831944f2ebd39425c36497b2710c3ec98c04de47" Nov 25 15:29:36 crc kubenswrapper[4890]: I1125 15:29:36.898630 4890 scope.go:117] "RemoveContainer" containerID="e893380f6bc195a09503a2f355f95510799db8ab623f49a44eedfa4e6b040962" Nov 25 15:29:36 crc kubenswrapper[4890]: I1125 15:29:36.925438 4890 scope.go:117] "RemoveContainer" containerID="57d487aa9b7d58ac42a98c9c84936083cef357fa3420580c62c2769103dcd5b7" Nov 25 15:29:36 crc kubenswrapper[4890]: I1125 15:29:36.951192 4890 scope.go:117] "RemoveContainer" containerID="9a1c99eefeb5490ba9f1a7c457143c55e8a76e9cf439ecb289dfeec1c9681662" Nov 25 15:29:36 crc kubenswrapper[4890]: I1125 15:29:36.989911 4890 scope.go:117] "RemoveContainer" containerID="9329cc99c6bd7e2b61bd28adaf2eb65dd4b8abace7166cfffeb874f0197267b5" Nov 25 15:29:38 crc kubenswrapper[4890]: I1125 15:29:38.069546 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-vgq94"] Nov 25 15:29:38 crc kubenswrapper[4890]: I1125 15:29:38.087104 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-vgq94"] Nov 25 15:29:38 crc kubenswrapper[4890]: I1125 15:29:38.200000 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21963cd4-490e-4c7f-9a9b-d93fa2a90ebc" path="/var/lib/kubelet/pods/21963cd4-490e-4c7f-9a9b-d93fa2a90ebc/volumes" Nov 25 15:29:39 crc kubenswrapper[4890]: I1125 15:29:39.231599 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-48z2b" podUID="eacc1f05-701e-49af-b2d7-9c049f6cddb9" containerName="registry-server" probeResult="failure" output=< Nov 25 15:29:39 crc kubenswrapper[4890]: timeout: failed to connect service ":50051" within 1s Nov 25 15:29:39 crc kubenswrapper[4890]: > Nov 25 15:29:40 crc kubenswrapper[4890]: I1125 15:29:40.161214 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mtfgl" podUID="f7235b7e-d6fe-45a2-93fd-66b0a77b4b76" containerName="registry-server" probeResult="failure" output=< Nov 25 15:29:40 crc kubenswrapper[4890]: timeout: failed to connect service ":50051" within 1s Nov 25 15:29:40 crc kubenswrapper[4890]: > Nov 25 15:29:40 crc kubenswrapper[4890]: I1125 15:29:40.362000 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4fdm7" Nov 25 15:29:40 crc kubenswrapper[4890]: I1125 15:29:40.412065 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4fdm7" Nov 25 15:29:41 crc kubenswrapper[4890]: I1125 15:29:41.551601 4890 generic.go:334] "Generic (PLEG): container finished" podID="6d15a3f2-bbd4-43ae-b620-93251f74369e" containerID="b79ae5ca12f341a420e7388f25e996104f0aeac321ce10a2b6fc205e67fe894a" exitCode=0 Nov 25 15:29:41 crc kubenswrapper[4890]: I1125 15:29:41.551656 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cr58b" event={"ID":"6d15a3f2-bbd4-43ae-b620-93251f74369e","Type":"ContainerDied","Data":"b79ae5ca12f341a420e7388f25e996104f0aeac321ce10a2b6fc205e67fe894a"} Nov 25 15:29:42 crc kubenswrapper[4890]: I1125 15:29:42.572274 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-f24bd" podUID="0f05c16a-1239-42a7-9910-ca03ef42b0f7" containerName="registry-server" probeResult="failure" output=< Nov 25 15:29:42 crc kubenswrapper[4890]: timeout: failed to connect service ":50051" within 1s Nov 25 15:29:42 crc kubenswrapper[4890]: > Nov 25 15:29:43 crc kubenswrapper[4890]: I1125 15:29:43.554694 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cr58b" Nov 25 15:29:43 crc kubenswrapper[4890]: I1125 15:29:43.579273 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cr58b" event={"ID":"6d15a3f2-bbd4-43ae-b620-93251f74369e","Type":"ContainerDied","Data":"07f2effed740c16bd10e19722336b98bdae3b19aeb07410d535393788f73462a"} Nov 25 15:29:43 crc kubenswrapper[4890]: I1125 15:29:43.579316 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="07f2effed740c16bd10e19722336b98bdae3b19aeb07410d535393788f73462a" Nov 25 15:29:43 crc kubenswrapper[4890]: I1125 15:29:43.579378 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cr58b" Nov 25 15:29:43 crc kubenswrapper[4890]: I1125 15:29:43.648796 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d15a3f2-bbd4-43ae-b620-93251f74369e-repo-setup-combined-ca-bundle\") pod \"6d15a3f2-bbd4-43ae-b620-93251f74369e\" (UID: \"6d15a3f2-bbd4-43ae-b620-93251f74369e\") " Nov 25 15:29:43 crc kubenswrapper[4890]: I1125 15:29:43.648879 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6d15a3f2-bbd4-43ae-b620-93251f74369e-ssh-key\") pod \"6d15a3f2-bbd4-43ae-b620-93251f74369e\" (UID: \"6d15a3f2-bbd4-43ae-b620-93251f74369e\") " Nov 25 15:29:43 crc kubenswrapper[4890]: I1125 15:29:43.652201 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xq2st\" (UniqueName: \"kubernetes.io/projected/6d15a3f2-bbd4-43ae-b620-93251f74369e-kube-api-access-xq2st\") pod \"6d15a3f2-bbd4-43ae-b620-93251f74369e\" (UID: \"6d15a3f2-bbd4-43ae-b620-93251f74369e\") " Nov 25 15:29:43 crc kubenswrapper[4890]: I1125 15:29:43.652327 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6d15a3f2-bbd4-43ae-b620-93251f74369e-inventory\") pod \"6d15a3f2-bbd4-43ae-b620-93251f74369e\" (UID: \"6d15a3f2-bbd4-43ae-b620-93251f74369e\") " Nov 25 15:29:43 crc kubenswrapper[4890]: I1125 15:29:43.656307 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d15a3f2-bbd4-43ae-b620-93251f74369e-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "6d15a3f2-bbd4-43ae-b620-93251f74369e" (UID: "6d15a3f2-bbd4-43ae-b620-93251f74369e"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:29:43 crc kubenswrapper[4890]: I1125 15:29:43.662425 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d15a3f2-bbd4-43ae-b620-93251f74369e-kube-api-access-xq2st" (OuterVolumeSpecName: "kube-api-access-xq2st") pod "6d15a3f2-bbd4-43ae-b620-93251f74369e" (UID: "6d15a3f2-bbd4-43ae-b620-93251f74369e"). InnerVolumeSpecName "kube-api-access-xq2st". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:29:43 crc kubenswrapper[4890]: I1125 15:29:43.667907 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-w4tcq"] Nov 25 15:29:43 crc kubenswrapper[4890]: E1125 15:29:43.668349 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56f8430f-3421-47db-b225-5c40fd14c8bd" containerName="extract-content" Nov 25 15:29:43 crc kubenswrapper[4890]: I1125 15:29:43.668366 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="56f8430f-3421-47db-b225-5c40fd14c8bd" containerName="extract-content" Nov 25 15:29:43 crc kubenswrapper[4890]: E1125 15:29:43.668383 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d15a3f2-bbd4-43ae-b620-93251f74369e" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 25 15:29:43 crc kubenswrapper[4890]: I1125 15:29:43.668391 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d15a3f2-bbd4-43ae-b620-93251f74369e" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 25 15:29:43 crc kubenswrapper[4890]: E1125 15:29:43.668404 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f82a9b69-631a-430c-9ae8-e18fbcb439dc" containerName="extract-utilities" Nov 25 15:29:43 crc kubenswrapper[4890]: I1125 15:29:43.668410 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="f82a9b69-631a-430c-9ae8-e18fbcb439dc" containerName="extract-utilities" Nov 25 15:29:43 crc kubenswrapper[4890]: E1125 15:29:43.668426 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56f8430f-3421-47db-b225-5c40fd14c8bd" containerName="registry-server" Nov 25 15:29:43 crc kubenswrapper[4890]: I1125 15:29:43.668432 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="56f8430f-3421-47db-b225-5c40fd14c8bd" containerName="registry-server" Nov 25 15:29:43 crc kubenswrapper[4890]: E1125 15:29:43.668456 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56f8430f-3421-47db-b225-5c40fd14c8bd" containerName="extract-utilities" Nov 25 15:29:43 crc kubenswrapper[4890]: I1125 15:29:43.668462 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="56f8430f-3421-47db-b225-5c40fd14c8bd" containerName="extract-utilities" Nov 25 15:29:43 crc kubenswrapper[4890]: E1125 15:29:43.668475 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f82a9b69-631a-430c-9ae8-e18fbcb439dc" containerName="extract-content" Nov 25 15:29:43 crc kubenswrapper[4890]: I1125 15:29:43.668482 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="f82a9b69-631a-430c-9ae8-e18fbcb439dc" containerName="extract-content" Nov 25 15:29:43 crc kubenswrapper[4890]: E1125 15:29:43.668497 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f82a9b69-631a-430c-9ae8-e18fbcb439dc" containerName="registry-server" Nov 25 15:29:43 crc kubenswrapper[4890]: I1125 15:29:43.668502 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="f82a9b69-631a-430c-9ae8-e18fbcb439dc" containerName="registry-server" Nov 25 15:29:43 crc kubenswrapper[4890]: I1125 15:29:43.668709 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="56f8430f-3421-47db-b225-5c40fd14c8bd" containerName="registry-server" Nov 25 15:29:43 crc kubenswrapper[4890]: I1125 15:29:43.668739 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d15a3f2-bbd4-43ae-b620-93251f74369e" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 25 15:29:43 crc kubenswrapper[4890]: I1125 15:29:43.668757 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="f82a9b69-631a-430c-9ae8-e18fbcb439dc" containerName="registry-server" Nov 25 15:29:43 crc kubenswrapper[4890]: I1125 15:29:43.670179 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-w4tcq" Nov 25 15:29:43 crc kubenswrapper[4890]: I1125 15:29:43.683101 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-w4tcq"] Nov 25 15:29:43 crc kubenswrapper[4890]: I1125 15:29:43.693641 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d15a3f2-bbd4-43ae-b620-93251f74369e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6d15a3f2-bbd4-43ae-b620-93251f74369e" (UID: "6d15a3f2-bbd4-43ae-b620-93251f74369e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:29:43 crc kubenswrapper[4890]: I1125 15:29:43.694098 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d15a3f2-bbd4-43ae-b620-93251f74369e-inventory" (OuterVolumeSpecName: "inventory") pod "6d15a3f2-bbd4-43ae-b620-93251f74369e" (UID: "6d15a3f2-bbd4-43ae-b620-93251f74369e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:29:43 crc kubenswrapper[4890]: I1125 15:29:43.755059 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4f0ccf80-fad3-4920-83e3-1c68111ed1ac-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-w4tcq\" (UID: \"4f0ccf80-fad3-4920-83e3-1c68111ed1ac\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-w4tcq" Nov 25 15:29:43 crc kubenswrapper[4890]: I1125 15:29:43.755476 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8tdc\" (UniqueName: \"kubernetes.io/projected/4f0ccf80-fad3-4920-83e3-1c68111ed1ac-kube-api-access-p8tdc\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-w4tcq\" (UID: \"4f0ccf80-fad3-4920-83e3-1c68111ed1ac\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-w4tcq" Nov 25 15:29:43 crc kubenswrapper[4890]: I1125 15:29:43.755970 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4f0ccf80-fad3-4920-83e3-1c68111ed1ac-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-w4tcq\" (UID: \"4f0ccf80-fad3-4920-83e3-1c68111ed1ac\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-w4tcq" Nov 25 15:29:43 crc kubenswrapper[4890]: I1125 15:29:43.756237 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xq2st\" (UniqueName: \"kubernetes.io/projected/6d15a3f2-bbd4-43ae-b620-93251f74369e-kube-api-access-xq2st\") on node \"crc\" DevicePath \"\"" Nov 25 15:29:43 crc kubenswrapper[4890]: I1125 15:29:43.756318 4890 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6d15a3f2-bbd4-43ae-b620-93251f74369e-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 15:29:43 crc kubenswrapper[4890]: I1125 15:29:43.756379 4890 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d15a3f2-bbd4-43ae-b620-93251f74369e-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:29:43 crc kubenswrapper[4890]: I1125 15:29:43.756433 4890 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6d15a3f2-bbd4-43ae-b620-93251f74369e-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 15:29:43 crc kubenswrapper[4890]: I1125 15:29:43.857823 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4f0ccf80-fad3-4920-83e3-1c68111ed1ac-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-w4tcq\" (UID: \"4f0ccf80-fad3-4920-83e3-1c68111ed1ac\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-w4tcq" Nov 25 15:29:43 crc kubenswrapper[4890]: I1125 15:29:43.857899 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4f0ccf80-fad3-4920-83e3-1c68111ed1ac-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-w4tcq\" (UID: \"4f0ccf80-fad3-4920-83e3-1c68111ed1ac\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-w4tcq" Nov 25 15:29:43 crc kubenswrapper[4890]: I1125 15:29:43.857921 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8tdc\" (UniqueName: \"kubernetes.io/projected/4f0ccf80-fad3-4920-83e3-1c68111ed1ac-kube-api-access-p8tdc\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-w4tcq\" (UID: \"4f0ccf80-fad3-4920-83e3-1c68111ed1ac\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-w4tcq" Nov 25 15:29:43 crc kubenswrapper[4890]: I1125 15:29:43.862248 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4f0ccf80-fad3-4920-83e3-1c68111ed1ac-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-w4tcq\" (UID: \"4f0ccf80-fad3-4920-83e3-1c68111ed1ac\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-w4tcq" Nov 25 15:29:43 crc kubenswrapper[4890]: I1125 15:29:43.862309 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4f0ccf80-fad3-4920-83e3-1c68111ed1ac-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-w4tcq\" (UID: \"4f0ccf80-fad3-4920-83e3-1c68111ed1ac\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-w4tcq" Nov 25 15:29:43 crc kubenswrapper[4890]: I1125 15:29:43.876394 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8tdc\" (UniqueName: \"kubernetes.io/projected/4f0ccf80-fad3-4920-83e3-1c68111ed1ac-kube-api-access-p8tdc\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-w4tcq\" (UID: \"4f0ccf80-fad3-4920-83e3-1c68111ed1ac\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-w4tcq" Nov 25 15:29:44 crc kubenswrapper[4890]: I1125 15:29:44.083809 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-w4tcq" Nov 25 15:29:44 crc kubenswrapper[4890]: W1125 15:29:44.634913 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f0ccf80_fad3_4920_83e3_1c68111ed1ac.slice/crio-0df2e4436e3353b2c6f6b5e42ea7ce950b8a6770280b139c671dc3d3680aa9e6 WatchSource:0}: Error finding container 0df2e4436e3353b2c6f6b5e42ea7ce950b8a6770280b139c671dc3d3680aa9e6: Status 404 returned error can't find the container with id 0df2e4436e3353b2c6f6b5e42ea7ce950b8a6770280b139c671dc3d3680aa9e6 Nov 25 15:29:44 crc kubenswrapper[4890]: I1125 15:29:44.636039 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-w4tcq"] Nov 25 15:29:45 crc kubenswrapper[4890]: I1125 15:29:45.602151 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-w4tcq" event={"ID":"4f0ccf80-fad3-4920-83e3-1c68111ed1ac","Type":"ContainerStarted","Data":"0df2e4436e3353b2c6f6b5e42ea7ce950b8a6770280b139c671dc3d3680aa9e6"} Nov 25 15:29:45 crc kubenswrapper[4890]: I1125 15:29:45.673081 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-h44zl" podUID="08bbb8ce-17b9-4797-80a7-de3beca270f4" containerName="registry-server" probeResult="failure" output=< Nov 25 15:29:45 crc kubenswrapper[4890]: timeout: failed to connect service ":50051" within 1s Nov 25 15:29:45 crc kubenswrapper[4890]: > Nov 25 15:29:48 crc kubenswrapper[4890]: I1125 15:29:48.173340 4890 scope.go:117] "RemoveContainer" containerID="9467aa8edc4d8bfe461dc02826750091f343018e37e7f269d305c77d21e832da" Nov 25 15:29:48 crc kubenswrapper[4890]: E1125 15:29:48.174116 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:29:48 crc kubenswrapper[4890]: I1125 15:29:48.630354 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-w4tcq" event={"ID":"4f0ccf80-fad3-4920-83e3-1c68111ed1ac","Type":"ContainerStarted","Data":"91a4007d5ed13d51f933f96e5629a11df7a7393ba4d463df21451e8f59a8e3e9"} Nov 25 15:29:49 crc kubenswrapper[4890]: I1125 15:29:49.234345 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-48z2b" podUID="eacc1f05-701e-49af-b2d7-9c049f6cddb9" containerName="registry-server" probeResult="failure" output=< Nov 25 15:29:49 crc kubenswrapper[4890]: timeout: failed to connect service ":50051" within 1s Nov 25 15:29:49 crc kubenswrapper[4890]: > Nov 25 15:29:49 crc kubenswrapper[4890]: I1125 15:29:49.663270 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-w4tcq" podStartSLOduration=3.8017179580000002 podStartE2EDuration="6.663249898s" podCreationTimestamp="2025-11-25 15:29:43 +0000 UTC" firstStartedPulling="2025-11-25 15:29:44.638055183 +0000 UTC m=+1643.080517793" lastFinishedPulling="2025-11-25 15:29:47.499587133 +0000 UTC m=+1645.942049733" observedRunningTime="2025-11-25 15:29:49.654568631 +0000 UTC m=+1648.097031241" watchObservedRunningTime="2025-11-25 15:29:49.663249898 +0000 UTC m=+1648.105712508" Nov 25 15:29:50 crc kubenswrapper[4890]: I1125 15:29:50.165774 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mtfgl" podUID="f7235b7e-d6fe-45a2-93fd-66b0a77b4b76" containerName="registry-server" probeResult="failure" output=< Nov 25 15:29:50 crc kubenswrapper[4890]: timeout: failed to connect service ":50051" within 1s Nov 25 15:29:50 crc kubenswrapper[4890]: > Nov 25 15:29:51 crc kubenswrapper[4890]: I1125 15:29:51.665461 4890 generic.go:334] "Generic (PLEG): container finished" podID="4f0ccf80-fad3-4920-83e3-1c68111ed1ac" containerID="91a4007d5ed13d51f933f96e5629a11df7a7393ba4d463df21451e8f59a8e3e9" exitCode=0 Nov 25 15:29:51 crc kubenswrapper[4890]: I1125 15:29:51.665590 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-w4tcq" event={"ID":"4f0ccf80-fad3-4920-83e3-1c68111ed1ac","Type":"ContainerDied","Data":"91a4007d5ed13d51f933f96e5629a11df7a7393ba4d463df21451e8f59a8e3e9"} Nov 25 15:29:52 crc kubenswrapper[4890]: I1125 15:29:52.548881 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-f24bd" podUID="0f05c16a-1239-42a7-9910-ca03ef42b0f7" containerName="registry-server" probeResult="failure" output=< Nov 25 15:29:52 crc kubenswrapper[4890]: timeout: failed to connect service ":50051" within 1s Nov 25 15:29:52 crc kubenswrapper[4890]: > Nov 25 15:29:53 crc kubenswrapper[4890]: I1125 15:29:53.091683 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-w4tcq" Nov 25 15:29:53 crc kubenswrapper[4890]: I1125 15:29:53.178467 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p8tdc\" (UniqueName: \"kubernetes.io/projected/4f0ccf80-fad3-4920-83e3-1c68111ed1ac-kube-api-access-p8tdc\") pod \"4f0ccf80-fad3-4920-83e3-1c68111ed1ac\" (UID: \"4f0ccf80-fad3-4920-83e3-1c68111ed1ac\") " Nov 25 15:29:53 crc kubenswrapper[4890]: I1125 15:29:53.178558 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4f0ccf80-fad3-4920-83e3-1c68111ed1ac-inventory\") pod \"4f0ccf80-fad3-4920-83e3-1c68111ed1ac\" (UID: \"4f0ccf80-fad3-4920-83e3-1c68111ed1ac\") " Nov 25 15:29:53 crc kubenswrapper[4890]: I1125 15:29:53.178840 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4f0ccf80-fad3-4920-83e3-1c68111ed1ac-ssh-key\") pod \"4f0ccf80-fad3-4920-83e3-1c68111ed1ac\" (UID: \"4f0ccf80-fad3-4920-83e3-1c68111ed1ac\") " Nov 25 15:29:53 crc kubenswrapper[4890]: I1125 15:29:53.185452 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f0ccf80-fad3-4920-83e3-1c68111ed1ac-kube-api-access-p8tdc" (OuterVolumeSpecName: "kube-api-access-p8tdc") pod "4f0ccf80-fad3-4920-83e3-1c68111ed1ac" (UID: "4f0ccf80-fad3-4920-83e3-1c68111ed1ac"). InnerVolumeSpecName "kube-api-access-p8tdc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:29:53 crc kubenswrapper[4890]: I1125 15:29:53.210106 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f0ccf80-fad3-4920-83e3-1c68111ed1ac-inventory" (OuterVolumeSpecName: "inventory") pod "4f0ccf80-fad3-4920-83e3-1c68111ed1ac" (UID: "4f0ccf80-fad3-4920-83e3-1c68111ed1ac"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:29:53 crc kubenswrapper[4890]: I1125 15:29:53.211505 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f0ccf80-fad3-4920-83e3-1c68111ed1ac-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4f0ccf80-fad3-4920-83e3-1c68111ed1ac" (UID: "4f0ccf80-fad3-4920-83e3-1c68111ed1ac"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:29:53 crc kubenswrapper[4890]: I1125 15:29:53.280639 4890 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4f0ccf80-fad3-4920-83e3-1c68111ed1ac-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 15:29:53 crc kubenswrapper[4890]: I1125 15:29:53.280677 4890 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4f0ccf80-fad3-4920-83e3-1c68111ed1ac-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 15:29:53 crc kubenswrapper[4890]: I1125 15:29:53.280690 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p8tdc\" (UniqueName: \"kubernetes.io/projected/4f0ccf80-fad3-4920-83e3-1c68111ed1ac-kube-api-access-p8tdc\") on node \"crc\" DevicePath \"\"" Nov 25 15:29:53 crc kubenswrapper[4890]: I1125 15:29:53.691702 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-w4tcq" event={"ID":"4f0ccf80-fad3-4920-83e3-1c68111ed1ac","Type":"ContainerDied","Data":"0df2e4436e3353b2c6f6b5e42ea7ce950b8a6770280b139c671dc3d3680aa9e6"} Nov 25 15:29:53 crc kubenswrapper[4890]: I1125 15:29:53.692185 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0df2e4436e3353b2c6f6b5e42ea7ce950b8a6770280b139c671dc3d3680aa9e6" Nov 25 15:29:53 crc kubenswrapper[4890]: I1125 15:29:53.692121 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-w4tcq" Nov 25 15:29:53 crc kubenswrapper[4890]: I1125 15:29:53.758642 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ddcgt"] Nov 25 15:29:53 crc kubenswrapper[4890]: E1125 15:29:53.759063 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f0ccf80-fad3-4920-83e3-1c68111ed1ac" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 25 15:29:53 crc kubenswrapper[4890]: I1125 15:29:53.759081 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f0ccf80-fad3-4920-83e3-1c68111ed1ac" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 25 15:29:53 crc kubenswrapper[4890]: I1125 15:29:53.759351 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f0ccf80-fad3-4920-83e3-1c68111ed1ac" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 25 15:29:53 crc kubenswrapper[4890]: I1125 15:29:53.760744 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ddcgt" Nov 25 15:29:53 crc kubenswrapper[4890]: I1125 15:29:53.765715 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 15:29:53 crc kubenswrapper[4890]: I1125 15:29:53.765985 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 15:29:53 crc kubenswrapper[4890]: I1125 15:29:53.766224 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-22zzp" Nov 25 15:29:53 crc kubenswrapper[4890]: I1125 15:29:53.766295 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 15:29:53 crc kubenswrapper[4890]: I1125 15:29:53.776453 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ddcgt"] Nov 25 15:29:53 crc kubenswrapper[4890]: I1125 15:29:53.789509 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dbln\" (UniqueName: \"kubernetes.io/projected/a2e80fb3-23be-4b01-b663-554207d5d538-kube-api-access-6dbln\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ddcgt\" (UID: \"a2e80fb3-23be-4b01-b663-554207d5d538\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ddcgt" Nov 25 15:29:53 crc kubenswrapper[4890]: I1125 15:29:53.789609 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a2e80fb3-23be-4b01-b663-554207d5d538-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ddcgt\" (UID: \"a2e80fb3-23be-4b01-b663-554207d5d538\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ddcgt" Nov 25 15:29:53 crc kubenswrapper[4890]: I1125 15:29:53.789647 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2e80fb3-23be-4b01-b663-554207d5d538-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ddcgt\" (UID: \"a2e80fb3-23be-4b01-b663-554207d5d538\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ddcgt" Nov 25 15:29:53 crc kubenswrapper[4890]: I1125 15:29:53.789711 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a2e80fb3-23be-4b01-b663-554207d5d538-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ddcgt\" (UID: \"a2e80fb3-23be-4b01-b663-554207d5d538\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ddcgt" Nov 25 15:29:53 crc kubenswrapper[4890]: I1125 15:29:53.891601 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a2e80fb3-23be-4b01-b663-554207d5d538-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ddcgt\" (UID: \"a2e80fb3-23be-4b01-b663-554207d5d538\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ddcgt" Nov 25 15:29:53 crc kubenswrapper[4890]: I1125 15:29:53.891704 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dbln\" (UniqueName: \"kubernetes.io/projected/a2e80fb3-23be-4b01-b663-554207d5d538-kube-api-access-6dbln\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ddcgt\" (UID: \"a2e80fb3-23be-4b01-b663-554207d5d538\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ddcgt" Nov 25 15:29:53 crc kubenswrapper[4890]: I1125 15:29:53.891800 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a2e80fb3-23be-4b01-b663-554207d5d538-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ddcgt\" (UID: \"a2e80fb3-23be-4b01-b663-554207d5d538\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ddcgt" Nov 25 15:29:53 crc kubenswrapper[4890]: I1125 15:29:53.891858 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2e80fb3-23be-4b01-b663-554207d5d538-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ddcgt\" (UID: \"a2e80fb3-23be-4b01-b663-554207d5d538\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ddcgt" Nov 25 15:29:53 crc kubenswrapper[4890]: I1125 15:29:53.898325 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a2e80fb3-23be-4b01-b663-554207d5d538-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ddcgt\" (UID: \"a2e80fb3-23be-4b01-b663-554207d5d538\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ddcgt" Nov 25 15:29:53 crc kubenswrapper[4890]: I1125 15:29:53.898447 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2e80fb3-23be-4b01-b663-554207d5d538-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ddcgt\" (UID: \"a2e80fb3-23be-4b01-b663-554207d5d538\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ddcgt" Nov 25 15:29:53 crc kubenswrapper[4890]: I1125 15:29:53.903030 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a2e80fb3-23be-4b01-b663-554207d5d538-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ddcgt\" (UID: \"a2e80fb3-23be-4b01-b663-554207d5d538\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ddcgt" Nov 25 15:29:53 crc kubenswrapper[4890]: I1125 15:29:53.923122 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dbln\" (UniqueName: \"kubernetes.io/projected/a2e80fb3-23be-4b01-b663-554207d5d538-kube-api-access-6dbln\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ddcgt\" (UID: \"a2e80fb3-23be-4b01-b663-554207d5d538\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ddcgt" Nov 25 15:29:54 crc kubenswrapper[4890]: I1125 15:29:54.082049 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ddcgt" Nov 25 15:29:54 crc kubenswrapper[4890]: I1125 15:29:54.682416 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ddcgt"] Nov 25 15:29:54 crc kubenswrapper[4890]: I1125 15:29:54.700861 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ddcgt" event={"ID":"a2e80fb3-23be-4b01-b663-554207d5d538","Type":"ContainerStarted","Data":"46857d31f257187cade02b75c1cbd691288b2fb6c1644e81b34a915e28718df0"} Nov 25 15:29:55 crc kubenswrapper[4890]: I1125 15:29:55.678023 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-h44zl" podUID="08bbb8ce-17b9-4797-80a7-de3beca270f4" containerName="registry-server" probeResult="failure" output=< Nov 25 15:29:55 crc kubenswrapper[4890]: timeout: failed to connect service ":50051" within 1s Nov 25 15:29:55 crc kubenswrapper[4890]: > Nov 25 15:29:55 crc kubenswrapper[4890]: I1125 15:29:55.678468 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-h44zl" Nov 25 15:29:55 crc kubenswrapper[4890]: I1125 15:29:55.679367 4890 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="registry-server" containerStatusID={"Type":"cri-o","ID":"8b53a07dcabfab68bb15bca0ff4a64607a430f8cd9580588a484ad20985c4ef8"} pod="openshift-marketplace/community-operators-h44zl" containerMessage="Container registry-server failed startup probe, will be restarted" Nov 25 15:29:55 crc kubenswrapper[4890]: I1125 15:29:55.679416 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-h44zl" podUID="08bbb8ce-17b9-4797-80a7-de3beca270f4" containerName="registry-server" containerID="cri-o://8b53a07dcabfab68bb15bca0ff4a64607a430f8cd9580588a484ad20985c4ef8" gracePeriod=30 Nov 25 15:29:57 crc kubenswrapper[4890]: I1125 15:29:57.731603 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ddcgt" event={"ID":"a2e80fb3-23be-4b01-b663-554207d5d538","Type":"ContainerStarted","Data":"82d1d929bc5761143b76abb4dbac8ffe1f724b89064540f4914bf10674b0f5aa"} Nov 25 15:29:57 crc kubenswrapper[4890]: I1125 15:29:57.749929 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ddcgt" podStartSLOduration=2.843092712 podStartE2EDuration="4.749904217s" podCreationTimestamp="2025-11-25 15:29:53 +0000 UTC" firstStartedPulling="2025-11-25 15:29:54.682013709 +0000 UTC m=+1653.124476309" lastFinishedPulling="2025-11-25 15:29:56.588825204 +0000 UTC m=+1655.031287814" observedRunningTime="2025-11-25 15:29:57.748287176 +0000 UTC m=+1656.190749786" watchObservedRunningTime="2025-11-25 15:29:57.749904217 +0000 UTC m=+1656.192366827" Nov 25 15:29:59 crc kubenswrapper[4890]: I1125 15:29:59.231307 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-48z2b" podUID="eacc1f05-701e-49af-b2d7-9c049f6cddb9" containerName="registry-server" probeResult="failure" output=< Nov 25 15:29:59 crc kubenswrapper[4890]: timeout: failed to connect service ":50051" within 1s Nov 25 15:29:59 crc kubenswrapper[4890]: > Nov 25 15:29:59 crc kubenswrapper[4890]: I1125 15:29:59.771994 4890 generic.go:334] "Generic (PLEG): container finished" podID="08bbb8ce-17b9-4797-80a7-de3beca270f4" containerID="8b53a07dcabfab68bb15bca0ff4a64607a430f8cd9580588a484ad20985c4ef8" exitCode=0 Nov 25 15:29:59 crc kubenswrapper[4890]: I1125 15:29:59.772074 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h44zl" event={"ID":"08bbb8ce-17b9-4797-80a7-de3beca270f4","Type":"ContainerDied","Data":"8b53a07dcabfab68bb15bca0ff4a64607a430f8cd9580588a484ad20985c4ef8"} Nov 25 15:30:00 crc kubenswrapper[4890]: I1125 15:30:00.148722 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401410-ngdkg"] Nov 25 15:30:00 crc kubenswrapper[4890]: I1125 15:30:00.150802 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401410-ngdkg" Nov 25 15:30:00 crc kubenswrapper[4890]: I1125 15:30:00.157459 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mtfgl" podUID="f7235b7e-d6fe-45a2-93fd-66b0a77b4b76" containerName="registry-server" probeResult="failure" output=< Nov 25 15:30:00 crc kubenswrapper[4890]: timeout: failed to connect service ":50051" within 1s Nov 25 15:30:00 crc kubenswrapper[4890]: > Nov 25 15:30:00 crc kubenswrapper[4890]: I1125 15:30:00.157487 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 15:30:00 crc kubenswrapper[4890]: I1125 15:30:00.161731 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401410-ngdkg"] Nov 25 15:30:00 crc kubenswrapper[4890]: I1125 15:30:00.162728 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 15:30:00 crc kubenswrapper[4890]: I1125 15:30:00.173704 4890 scope.go:117] "RemoveContainer" containerID="9467aa8edc4d8bfe461dc02826750091f343018e37e7f269d305c77d21e832da" Nov 25 15:30:00 crc kubenswrapper[4890]: E1125 15:30:00.174061 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:30:00 crc kubenswrapper[4890]: I1125 15:30:00.329044 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/95656df6-ba37-41ef-a281-87240f41c544-secret-volume\") pod \"collect-profiles-29401410-ngdkg\" (UID: \"95656df6-ba37-41ef-a281-87240f41c544\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401410-ngdkg" Nov 25 15:30:00 crc kubenswrapper[4890]: I1125 15:30:00.329122 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jnlk\" (UniqueName: \"kubernetes.io/projected/95656df6-ba37-41ef-a281-87240f41c544-kube-api-access-9jnlk\") pod \"collect-profiles-29401410-ngdkg\" (UID: \"95656df6-ba37-41ef-a281-87240f41c544\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401410-ngdkg" Nov 25 15:30:00 crc kubenswrapper[4890]: I1125 15:30:00.329182 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/95656df6-ba37-41ef-a281-87240f41c544-config-volume\") pod \"collect-profiles-29401410-ngdkg\" (UID: \"95656df6-ba37-41ef-a281-87240f41c544\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401410-ngdkg" Nov 25 15:30:00 crc kubenswrapper[4890]: I1125 15:30:00.432022 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jnlk\" (UniqueName: \"kubernetes.io/projected/95656df6-ba37-41ef-a281-87240f41c544-kube-api-access-9jnlk\") pod \"collect-profiles-29401410-ngdkg\" (UID: \"95656df6-ba37-41ef-a281-87240f41c544\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401410-ngdkg" Nov 25 15:30:00 crc kubenswrapper[4890]: I1125 15:30:00.432150 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/95656df6-ba37-41ef-a281-87240f41c544-config-volume\") pod \"collect-profiles-29401410-ngdkg\" (UID: \"95656df6-ba37-41ef-a281-87240f41c544\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401410-ngdkg" Nov 25 15:30:00 crc kubenswrapper[4890]: I1125 15:30:00.432405 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/95656df6-ba37-41ef-a281-87240f41c544-secret-volume\") pod \"collect-profiles-29401410-ngdkg\" (UID: \"95656df6-ba37-41ef-a281-87240f41c544\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401410-ngdkg" Nov 25 15:30:00 crc kubenswrapper[4890]: I1125 15:30:00.433635 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/95656df6-ba37-41ef-a281-87240f41c544-config-volume\") pod \"collect-profiles-29401410-ngdkg\" (UID: \"95656df6-ba37-41ef-a281-87240f41c544\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401410-ngdkg" Nov 25 15:30:00 crc kubenswrapper[4890]: I1125 15:30:00.441695 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/95656df6-ba37-41ef-a281-87240f41c544-secret-volume\") pod \"collect-profiles-29401410-ngdkg\" (UID: \"95656df6-ba37-41ef-a281-87240f41c544\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401410-ngdkg" Nov 25 15:30:00 crc kubenswrapper[4890]: I1125 15:30:00.451726 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jnlk\" (UniqueName: \"kubernetes.io/projected/95656df6-ba37-41ef-a281-87240f41c544-kube-api-access-9jnlk\") pod \"collect-profiles-29401410-ngdkg\" (UID: \"95656df6-ba37-41ef-a281-87240f41c544\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401410-ngdkg" Nov 25 15:30:00 crc kubenswrapper[4890]: I1125 15:30:00.484736 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401410-ngdkg" Nov 25 15:30:00 crc kubenswrapper[4890]: I1125 15:30:00.959646 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401410-ngdkg"] Nov 25 15:30:00 crc kubenswrapper[4890]: W1125 15:30:00.972757 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod95656df6_ba37_41ef_a281_87240f41c544.slice/crio-64b1bd585fc8a6db91ba6bcb8e480c6bbf661d02465dc03192aa471ab0cbb7fc WatchSource:0}: Error finding container 64b1bd585fc8a6db91ba6bcb8e480c6bbf661d02465dc03192aa471ab0cbb7fc: Status 404 returned error can't find the container with id 64b1bd585fc8a6db91ba6bcb8e480c6bbf661d02465dc03192aa471ab0cbb7fc Nov 25 15:30:01 crc kubenswrapper[4890]: I1125 15:30:01.794735 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h44zl" event={"ID":"08bbb8ce-17b9-4797-80a7-de3beca270f4","Type":"ContainerStarted","Data":"7cbbc2af37308da187efc26ff75f2815686e7168fb9c0b836eee86b68b721c60"} Nov 25 15:30:01 crc kubenswrapper[4890]: I1125 15:30:01.798936 4890 generic.go:334] "Generic (PLEG): container finished" podID="95656df6-ba37-41ef-a281-87240f41c544" containerID="586f0cfcafc29ab77e85b46b9543b5054cba15609dc50e05be6900565ee49a9f" exitCode=0 Nov 25 15:30:01 crc kubenswrapper[4890]: I1125 15:30:01.798986 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401410-ngdkg" event={"ID":"95656df6-ba37-41ef-a281-87240f41c544","Type":"ContainerDied","Data":"586f0cfcafc29ab77e85b46b9543b5054cba15609dc50e05be6900565ee49a9f"} Nov 25 15:30:01 crc kubenswrapper[4890]: I1125 15:30:01.799015 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401410-ngdkg" event={"ID":"95656df6-ba37-41ef-a281-87240f41c544","Type":"ContainerStarted","Data":"64b1bd585fc8a6db91ba6bcb8e480c6bbf661d02465dc03192aa471ab0cbb7fc"} Nov 25 15:30:02 crc kubenswrapper[4890]: I1125 15:30:02.552881 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-f24bd" podUID="0f05c16a-1239-42a7-9910-ca03ef42b0f7" containerName="registry-server" probeResult="failure" output=< Nov 25 15:30:02 crc kubenswrapper[4890]: timeout: failed to connect service ":50051" within 1s Nov 25 15:30:02 crc kubenswrapper[4890]: > Nov 25 15:30:03 crc kubenswrapper[4890]: I1125 15:30:03.487004 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401410-ngdkg" Nov 25 15:30:03 crc kubenswrapper[4890]: I1125 15:30:03.600500 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/95656df6-ba37-41ef-a281-87240f41c544-secret-volume\") pod \"95656df6-ba37-41ef-a281-87240f41c544\" (UID: \"95656df6-ba37-41ef-a281-87240f41c544\") " Nov 25 15:30:03 crc kubenswrapper[4890]: I1125 15:30:03.600615 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/95656df6-ba37-41ef-a281-87240f41c544-config-volume\") pod \"95656df6-ba37-41ef-a281-87240f41c544\" (UID: \"95656df6-ba37-41ef-a281-87240f41c544\") " Nov 25 15:30:03 crc kubenswrapper[4890]: I1125 15:30:03.600777 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9jnlk\" (UniqueName: \"kubernetes.io/projected/95656df6-ba37-41ef-a281-87240f41c544-kube-api-access-9jnlk\") pod \"95656df6-ba37-41ef-a281-87240f41c544\" (UID: \"95656df6-ba37-41ef-a281-87240f41c544\") " Nov 25 15:30:03 crc kubenswrapper[4890]: I1125 15:30:03.601574 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95656df6-ba37-41ef-a281-87240f41c544-config-volume" (OuterVolumeSpecName: "config-volume") pod "95656df6-ba37-41ef-a281-87240f41c544" (UID: "95656df6-ba37-41ef-a281-87240f41c544"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:30:03 crc kubenswrapper[4890]: I1125 15:30:03.606905 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95656df6-ba37-41ef-a281-87240f41c544-kube-api-access-9jnlk" (OuterVolumeSpecName: "kube-api-access-9jnlk") pod "95656df6-ba37-41ef-a281-87240f41c544" (UID: "95656df6-ba37-41ef-a281-87240f41c544"). InnerVolumeSpecName "kube-api-access-9jnlk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:30:03 crc kubenswrapper[4890]: I1125 15:30:03.606949 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95656df6-ba37-41ef-a281-87240f41c544-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "95656df6-ba37-41ef-a281-87240f41c544" (UID: "95656df6-ba37-41ef-a281-87240f41c544"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:30:03 crc kubenswrapper[4890]: I1125 15:30:03.703414 4890 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/95656df6-ba37-41ef-a281-87240f41c544-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 15:30:03 crc kubenswrapper[4890]: I1125 15:30:03.703470 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9jnlk\" (UniqueName: \"kubernetes.io/projected/95656df6-ba37-41ef-a281-87240f41c544-kube-api-access-9jnlk\") on node \"crc\" DevicePath \"\"" Nov 25 15:30:03 crc kubenswrapper[4890]: I1125 15:30:03.703485 4890 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/95656df6-ba37-41ef-a281-87240f41c544-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 15:30:03 crc kubenswrapper[4890]: I1125 15:30:03.821582 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401410-ngdkg" event={"ID":"95656df6-ba37-41ef-a281-87240f41c544","Type":"ContainerDied","Data":"64b1bd585fc8a6db91ba6bcb8e480c6bbf661d02465dc03192aa471ab0cbb7fc"} Nov 25 15:30:03 crc kubenswrapper[4890]: I1125 15:30:03.821622 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="64b1bd585fc8a6db91ba6bcb8e480c6bbf661d02465dc03192aa471ab0cbb7fc" Nov 25 15:30:03 crc kubenswrapper[4890]: I1125 15:30:03.821638 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401410-ngdkg" Nov 25 15:30:04 crc kubenswrapper[4890]: I1125 15:30:04.624658 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-h44zl" Nov 25 15:30:04 crc kubenswrapper[4890]: I1125 15:30:04.625020 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-h44zl" Nov 25 15:30:05 crc kubenswrapper[4890]: I1125 15:30:05.675927 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-h44zl" podUID="08bbb8ce-17b9-4797-80a7-de3beca270f4" containerName="registry-server" probeResult="failure" output=< Nov 25 15:30:05 crc kubenswrapper[4890]: timeout: failed to connect service ":50051" within 1s Nov 25 15:30:05 crc kubenswrapper[4890]: > Nov 25 15:30:09 crc kubenswrapper[4890]: I1125 15:30:09.227916 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-48z2b" podUID="eacc1f05-701e-49af-b2d7-9c049f6cddb9" containerName="registry-server" probeResult="failure" output=< Nov 25 15:30:09 crc kubenswrapper[4890]: timeout: failed to connect service ":50051" within 1s Nov 25 15:30:09 crc kubenswrapper[4890]: > Nov 25 15:30:10 crc kubenswrapper[4890]: I1125 15:30:10.178983 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mtfgl" podUID="f7235b7e-d6fe-45a2-93fd-66b0a77b4b76" containerName="registry-server" probeResult="failure" output=< Nov 25 15:30:10 crc kubenswrapper[4890]: timeout: failed to connect service ":50051" within 1s Nov 25 15:30:10 crc kubenswrapper[4890]: > Nov 25 15:30:11 crc kubenswrapper[4890]: I1125 15:30:11.174594 4890 scope.go:117] "RemoveContainer" containerID="9467aa8edc4d8bfe461dc02826750091f343018e37e7f269d305c77d21e832da" Nov 25 15:30:11 crc kubenswrapper[4890]: E1125 15:30:11.175201 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:30:12 crc kubenswrapper[4890]: I1125 15:30:12.562915 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-f24bd" podUID="0f05c16a-1239-42a7-9910-ca03ef42b0f7" containerName="registry-server" probeResult="failure" output=< Nov 25 15:30:12 crc kubenswrapper[4890]: timeout: failed to connect service ":50051" within 1s Nov 25 15:30:12 crc kubenswrapper[4890]: > Nov 25 15:30:14 crc kubenswrapper[4890]: I1125 15:30:14.677348 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-h44zl" Nov 25 15:30:14 crc kubenswrapper[4890]: I1125 15:30:14.736235 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-h44zl" Nov 25 15:30:14 crc kubenswrapper[4890]: I1125 15:30:14.851953 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-492nv"] Nov 25 15:30:14 crc kubenswrapper[4890]: I1125 15:30:14.922690 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wv9pw"] Nov 25 15:30:14 crc kubenswrapper[4890]: I1125 15:30:14.923170 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-wv9pw" podUID="6febf84d-85dd-45d8-a32e-bc978311ed98" containerName="registry-server" containerID="cri-o://2b6d1256478ad7877a0259e89b82b8b3c6ecb95cb1a0b7f86c266cb04d479d08" gracePeriod=2 Nov 25 15:30:15 crc kubenswrapper[4890]: I1125 15:30:15.941107 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gxfqw"] Nov 25 15:30:16 crc kubenswrapper[4890]: I1125 15:30:16.016696 4890 generic.go:334] "Generic (PLEG): container finished" podID="6febf84d-85dd-45d8-a32e-bc978311ed98" containerID="2b6d1256478ad7877a0259e89b82b8b3c6ecb95cb1a0b7f86c266cb04d479d08" exitCode=0 Nov 25 15:30:16 crc kubenswrapper[4890]: I1125 15:30:16.016915 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wv9pw" event={"ID":"6febf84d-85dd-45d8-a32e-bc978311ed98","Type":"ContainerDied","Data":"2b6d1256478ad7877a0259e89b82b8b3c6ecb95cb1a0b7f86c266cb04d479d08"} Nov 25 15:30:16 crc kubenswrapper[4890]: I1125 15:30:16.076219 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wv9pw" Nov 25 15:30:16 crc kubenswrapper[4890]: I1125 15:30:16.128038 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-492nv"] Nov 25 15:30:16 crc kubenswrapper[4890]: I1125 15:30:16.128311 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-492nv" podUID="1fcc307a-b17f-46ab-aabe-0a0361656701" containerName="registry-server" containerID="cri-o://5e92fd7ea83b5fb5a96a5febb629ff7a0ac952ad0684f3a95450d43595eb7d22" gracePeriod=2 Nov 25 15:30:16 crc kubenswrapper[4890]: I1125 15:30:16.180954 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9f6v\" (UniqueName: \"kubernetes.io/projected/6febf84d-85dd-45d8-a32e-bc978311ed98-kube-api-access-j9f6v\") pod \"6febf84d-85dd-45d8-a32e-bc978311ed98\" (UID: \"6febf84d-85dd-45d8-a32e-bc978311ed98\") " Nov 25 15:30:16 crc kubenswrapper[4890]: I1125 15:30:16.181522 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6febf84d-85dd-45d8-a32e-bc978311ed98-catalog-content\") pod \"6febf84d-85dd-45d8-a32e-bc978311ed98\" (UID: \"6febf84d-85dd-45d8-a32e-bc978311ed98\") " Nov 25 15:30:16 crc kubenswrapper[4890]: I1125 15:30:16.181573 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6febf84d-85dd-45d8-a32e-bc978311ed98-utilities\") pod \"6febf84d-85dd-45d8-a32e-bc978311ed98\" (UID: \"6febf84d-85dd-45d8-a32e-bc978311ed98\") " Nov 25 15:30:16 crc kubenswrapper[4890]: I1125 15:30:16.182315 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6febf84d-85dd-45d8-a32e-bc978311ed98-utilities" (OuterVolumeSpecName: "utilities") pod "6febf84d-85dd-45d8-a32e-bc978311ed98" (UID: "6febf84d-85dd-45d8-a32e-bc978311ed98"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:30:16 crc kubenswrapper[4890]: I1125 15:30:16.192380 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6febf84d-85dd-45d8-a32e-bc978311ed98-kube-api-access-j9f6v" (OuterVolumeSpecName: "kube-api-access-j9f6v") pod "6febf84d-85dd-45d8-a32e-bc978311ed98" (UID: "6febf84d-85dd-45d8-a32e-bc978311ed98"). InnerVolumeSpecName "kube-api-access-j9f6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:30:16 crc kubenswrapper[4890]: I1125 15:30:16.243361 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6febf84d-85dd-45d8-a32e-bc978311ed98-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6febf84d-85dd-45d8-a32e-bc978311ed98" (UID: "6febf84d-85dd-45d8-a32e-bc978311ed98"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:30:16 crc kubenswrapper[4890]: I1125 15:30:16.284576 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6febf84d-85dd-45d8-a32e-bc978311ed98-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 15:30:16 crc kubenswrapper[4890]: I1125 15:30:16.284620 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6febf84d-85dd-45d8-a32e-bc978311ed98-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 15:30:16 crc kubenswrapper[4890]: I1125 15:30:16.284633 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9f6v\" (UniqueName: \"kubernetes.io/projected/6febf84d-85dd-45d8-a32e-bc978311ed98-kube-api-access-j9f6v\") on node \"crc\" DevicePath \"\"" Nov 25 15:30:16 crc kubenswrapper[4890]: I1125 15:30:16.628473 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-492nv" Nov 25 15:30:16 crc kubenswrapper[4890]: I1125 15:30:16.806377 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fcc307a-b17f-46ab-aabe-0a0361656701-utilities\") pod \"1fcc307a-b17f-46ab-aabe-0a0361656701\" (UID: \"1fcc307a-b17f-46ab-aabe-0a0361656701\") " Nov 25 15:30:16 crc kubenswrapper[4890]: I1125 15:30:16.806567 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fcc307a-b17f-46ab-aabe-0a0361656701-catalog-content\") pod \"1fcc307a-b17f-46ab-aabe-0a0361656701\" (UID: \"1fcc307a-b17f-46ab-aabe-0a0361656701\") " Nov 25 15:30:16 crc kubenswrapper[4890]: I1125 15:30:16.806655 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gxd9v\" (UniqueName: \"kubernetes.io/projected/1fcc307a-b17f-46ab-aabe-0a0361656701-kube-api-access-gxd9v\") pod \"1fcc307a-b17f-46ab-aabe-0a0361656701\" (UID: \"1fcc307a-b17f-46ab-aabe-0a0361656701\") " Nov 25 15:30:16 crc kubenswrapper[4890]: I1125 15:30:16.807220 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1fcc307a-b17f-46ab-aabe-0a0361656701-utilities" (OuterVolumeSpecName: "utilities") pod "1fcc307a-b17f-46ab-aabe-0a0361656701" (UID: "1fcc307a-b17f-46ab-aabe-0a0361656701"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:30:16 crc kubenswrapper[4890]: I1125 15:30:16.812246 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1fcc307a-b17f-46ab-aabe-0a0361656701-kube-api-access-gxd9v" (OuterVolumeSpecName: "kube-api-access-gxd9v") pod "1fcc307a-b17f-46ab-aabe-0a0361656701" (UID: "1fcc307a-b17f-46ab-aabe-0a0361656701"). InnerVolumeSpecName "kube-api-access-gxd9v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:30:16 crc kubenswrapper[4890]: I1125 15:30:16.861397 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1fcc307a-b17f-46ab-aabe-0a0361656701-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1fcc307a-b17f-46ab-aabe-0a0361656701" (UID: "1fcc307a-b17f-46ab-aabe-0a0361656701"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:30:16 crc kubenswrapper[4890]: I1125 15:30:16.908644 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fcc307a-b17f-46ab-aabe-0a0361656701-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 15:30:16 crc kubenswrapper[4890]: I1125 15:30:16.908680 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fcc307a-b17f-46ab-aabe-0a0361656701-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 15:30:16 crc kubenswrapper[4890]: I1125 15:30:16.908691 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gxd9v\" (UniqueName: \"kubernetes.io/projected/1fcc307a-b17f-46ab-aabe-0a0361656701-kube-api-access-gxd9v\") on node \"crc\" DevicePath \"\"" Nov 25 15:30:17 crc kubenswrapper[4890]: I1125 15:30:17.032073 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wv9pw" event={"ID":"6febf84d-85dd-45d8-a32e-bc978311ed98","Type":"ContainerDied","Data":"f7f3ac4b168bff400e533780771eea8c855e98984c4dc3dc6fdc76a899ae27e6"} Nov 25 15:30:17 crc kubenswrapper[4890]: I1125 15:30:17.032113 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wv9pw" Nov 25 15:30:17 crc kubenswrapper[4890]: I1125 15:30:17.032197 4890 scope.go:117] "RemoveContainer" containerID="2b6d1256478ad7877a0259e89b82b8b3c6ecb95cb1a0b7f86c266cb04d479d08" Nov 25 15:30:17 crc kubenswrapper[4890]: I1125 15:30:17.034841 4890 generic.go:334] "Generic (PLEG): container finished" podID="1fcc307a-b17f-46ab-aabe-0a0361656701" containerID="5e92fd7ea83b5fb5a96a5febb629ff7a0ac952ad0684f3a95450d43595eb7d22" exitCode=0 Nov 25 15:30:17 crc kubenswrapper[4890]: I1125 15:30:17.034887 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-492nv" event={"ID":"1fcc307a-b17f-46ab-aabe-0a0361656701","Type":"ContainerDied","Data":"5e92fd7ea83b5fb5a96a5febb629ff7a0ac952ad0684f3a95450d43595eb7d22"} Nov 25 15:30:17 crc kubenswrapper[4890]: I1125 15:30:17.034927 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-492nv" event={"ID":"1fcc307a-b17f-46ab-aabe-0a0361656701","Type":"ContainerDied","Data":"bc3d2402820722971a63507a69a3578ea47f648835295734093bfd05126b245b"} Nov 25 15:30:17 crc kubenswrapper[4890]: I1125 15:30:17.034973 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-492nv" Nov 25 15:30:17 crc kubenswrapper[4890]: I1125 15:30:17.066290 4890 scope.go:117] "RemoveContainer" containerID="2acfef7b9ba50e368fad34e67848e20758f2619222bcac74ca7cefa436c0ba87" Nov 25 15:30:17 crc kubenswrapper[4890]: I1125 15:30:17.088461 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wv9pw"] Nov 25 15:30:17 crc kubenswrapper[4890]: I1125 15:30:17.100845 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-wv9pw"] Nov 25 15:30:17 crc kubenswrapper[4890]: I1125 15:30:17.103891 4890 scope.go:117] "RemoveContainer" containerID="2fc8c56c15eaf81ef26fdcab5ab1bf53eb820dc3116e4e2336449c2d1654c39e" Nov 25 15:30:17 crc kubenswrapper[4890]: I1125 15:30:17.115004 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-492nv"] Nov 25 15:30:17 crc kubenswrapper[4890]: I1125 15:30:17.127278 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-492nv"] Nov 25 15:30:17 crc kubenswrapper[4890]: I1125 15:30:17.129447 4890 scope.go:117] "RemoveContainer" containerID="5e92fd7ea83b5fb5a96a5febb629ff7a0ac952ad0684f3a95450d43595eb7d22" Nov 25 15:30:17 crc kubenswrapper[4890]: I1125 15:30:17.150898 4890 scope.go:117] "RemoveContainer" containerID="60e586812f3d16610b2f2890383a9106ad815412ccc1d6c56e39851322fd6bb3" Nov 25 15:30:17 crc kubenswrapper[4890]: I1125 15:30:17.184932 4890 scope.go:117] "RemoveContainer" containerID="23a8a7f031d8d7fb3e4865029b2f6864c91388d8980dd59c503a7a2bc53430d6" Nov 25 15:30:17 crc kubenswrapper[4890]: I1125 15:30:17.288024 4890 scope.go:117] "RemoveContainer" containerID="5e92fd7ea83b5fb5a96a5febb629ff7a0ac952ad0684f3a95450d43595eb7d22" Nov 25 15:30:17 crc kubenswrapper[4890]: E1125 15:30:17.288667 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e92fd7ea83b5fb5a96a5febb629ff7a0ac952ad0684f3a95450d43595eb7d22\": container with ID starting with 5e92fd7ea83b5fb5a96a5febb629ff7a0ac952ad0684f3a95450d43595eb7d22 not found: ID does not exist" containerID="5e92fd7ea83b5fb5a96a5febb629ff7a0ac952ad0684f3a95450d43595eb7d22" Nov 25 15:30:17 crc kubenswrapper[4890]: I1125 15:30:17.288749 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e92fd7ea83b5fb5a96a5febb629ff7a0ac952ad0684f3a95450d43595eb7d22"} err="failed to get container status \"5e92fd7ea83b5fb5a96a5febb629ff7a0ac952ad0684f3a95450d43595eb7d22\": rpc error: code = NotFound desc = could not find container \"5e92fd7ea83b5fb5a96a5febb629ff7a0ac952ad0684f3a95450d43595eb7d22\": container with ID starting with 5e92fd7ea83b5fb5a96a5febb629ff7a0ac952ad0684f3a95450d43595eb7d22 not found: ID does not exist" Nov 25 15:30:17 crc kubenswrapper[4890]: I1125 15:30:17.288802 4890 scope.go:117] "RemoveContainer" containerID="60e586812f3d16610b2f2890383a9106ad815412ccc1d6c56e39851322fd6bb3" Nov 25 15:30:17 crc kubenswrapper[4890]: E1125 15:30:17.289351 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60e586812f3d16610b2f2890383a9106ad815412ccc1d6c56e39851322fd6bb3\": container with ID starting with 60e586812f3d16610b2f2890383a9106ad815412ccc1d6c56e39851322fd6bb3 not found: ID does not exist" containerID="60e586812f3d16610b2f2890383a9106ad815412ccc1d6c56e39851322fd6bb3" Nov 25 15:30:17 crc kubenswrapper[4890]: I1125 15:30:17.289389 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60e586812f3d16610b2f2890383a9106ad815412ccc1d6c56e39851322fd6bb3"} err="failed to get container status \"60e586812f3d16610b2f2890383a9106ad815412ccc1d6c56e39851322fd6bb3\": rpc error: code = NotFound desc = could not find container \"60e586812f3d16610b2f2890383a9106ad815412ccc1d6c56e39851322fd6bb3\": container with ID starting with 60e586812f3d16610b2f2890383a9106ad815412ccc1d6c56e39851322fd6bb3 not found: ID does not exist" Nov 25 15:30:17 crc kubenswrapper[4890]: I1125 15:30:17.289419 4890 scope.go:117] "RemoveContainer" containerID="23a8a7f031d8d7fb3e4865029b2f6864c91388d8980dd59c503a7a2bc53430d6" Nov 25 15:30:17 crc kubenswrapper[4890]: E1125 15:30:17.290016 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23a8a7f031d8d7fb3e4865029b2f6864c91388d8980dd59c503a7a2bc53430d6\": container with ID starting with 23a8a7f031d8d7fb3e4865029b2f6864c91388d8980dd59c503a7a2bc53430d6 not found: ID does not exist" containerID="23a8a7f031d8d7fb3e4865029b2f6864c91388d8980dd59c503a7a2bc53430d6" Nov 25 15:30:17 crc kubenswrapper[4890]: I1125 15:30:17.290079 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23a8a7f031d8d7fb3e4865029b2f6864c91388d8980dd59c503a7a2bc53430d6"} err="failed to get container status \"23a8a7f031d8d7fb3e4865029b2f6864c91388d8980dd59c503a7a2bc53430d6\": rpc error: code = NotFound desc = could not find container \"23a8a7f031d8d7fb3e4865029b2f6864c91388d8980dd59c503a7a2bc53430d6\": container with ID starting with 23a8a7f031d8d7fb3e4865029b2f6864c91388d8980dd59c503a7a2bc53430d6 not found: ID does not exist" Nov 25 15:30:17 crc kubenswrapper[4890]: I1125 15:30:17.524606 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-h44zl"] Nov 25 15:30:17 crc kubenswrapper[4890]: I1125 15:30:17.525117 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-h44zl" podUID="08bbb8ce-17b9-4797-80a7-de3beca270f4" containerName="registry-server" containerID="cri-o://7cbbc2af37308da187efc26ff75f2815686e7168fb9c0b836eee86b68b721c60" gracePeriod=2 Nov 25 15:30:17 crc kubenswrapper[4890]: I1125 15:30:17.726268 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rcbhc"] Nov 25 15:30:17 crc kubenswrapper[4890]: I1125 15:30:17.726783 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rcbhc" podUID="b18eb2e9-77b4-46ba-9e32-31c1e9b73963" containerName="registry-server" containerID="cri-o://a4ccb66699200fcad030dec7f35c81f1826dbfa3d9aede9a2806ab3b57bd0a09" gracePeriod=2 Nov 25 15:30:17 crc kubenswrapper[4890]: I1125 15:30:17.927300 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-v8mj2"] Nov 25 15:30:17 crc kubenswrapper[4890]: I1125 15:30:17.928311 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-v8mj2" podUID="00771e9e-5583-42ab-b834-4643d9bbdbec" containerName="registry-server" containerID="cri-o://583629b7aeb99a0f6d9e57cd9965500d687d0958872fcdd86b78e86046b11f5a" gracePeriod=2 Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.112865 4890 generic.go:334] "Generic (PLEG): container finished" podID="08bbb8ce-17b9-4797-80a7-de3beca270f4" containerID="7cbbc2af37308da187efc26ff75f2815686e7168fb9c0b836eee86b68b721c60" exitCode=0 Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.113649 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h44zl" event={"ID":"08bbb8ce-17b9-4797-80a7-de3beca270f4","Type":"ContainerDied","Data":"7cbbc2af37308da187efc26ff75f2815686e7168fb9c0b836eee86b68b721c60"} Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.113774 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h44zl" event={"ID":"08bbb8ce-17b9-4797-80a7-de3beca270f4","Type":"ContainerDied","Data":"32d6e83eadedd458f04dc0727d7bb04a7014d5a7650ee8ec9b4afbb8ce0f29b4"} Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.113843 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="32d6e83eadedd458f04dc0727d7bb04a7014d5a7650ee8ec9b4afbb8ce0f29b4" Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.113930 4890 scope.go:117] "RemoveContainer" containerID="8b53a07dcabfab68bb15bca0ff4a64607a430f8cd9580588a484ad20985c4ef8" Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.139372 4890 generic.go:334] "Generic (PLEG): container finished" podID="b18eb2e9-77b4-46ba-9e32-31c1e9b73963" containerID="a4ccb66699200fcad030dec7f35c81f1826dbfa3d9aede9a2806ab3b57bd0a09" exitCode=0 Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.139423 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rcbhc" event={"ID":"b18eb2e9-77b4-46ba-9e32-31c1e9b73963","Type":"ContainerDied","Data":"a4ccb66699200fcad030dec7f35c81f1826dbfa3d9aede9a2806ab3b57bd0a09"} Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.143881 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h44zl" Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.202505 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1fcc307a-b17f-46ab-aabe-0a0361656701" path="/var/lib/kubelet/pods/1fcc307a-b17f-46ab-aabe-0a0361656701/volumes" Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.214869 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6febf84d-85dd-45d8-a32e-bc978311ed98" path="/var/lib/kubelet/pods/6febf84d-85dd-45d8-a32e-bc978311ed98/volumes" Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.245306 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08bbb8ce-17b9-4797-80a7-de3beca270f4-utilities\") pod \"08bbb8ce-17b9-4797-80a7-de3beca270f4\" (UID: \"08bbb8ce-17b9-4797-80a7-de3beca270f4\") " Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.245475 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-49srl\" (UniqueName: \"kubernetes.io/projected/08bbb8ce-17b9-4797-80a7-de3beca270f4-kube-api-access-49srl\") pod \"08bbb8ce-17b9-4797-80a7-de3beca270f4\" (UID: \"08bbb8ce-17b9-4797-80a7-de3beca270f4\") " Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.245638 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08bbb8ce-17b9-4797-80a7-de3beca270f4-catalog-content\") pod \"08bbb8ce-17b9-4797-80a7-de3beca270f4\" (UID: \"08bbb8ce-17b9-4797-80a7-de3beca270f4\") " Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.249643 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08bbb8ce-17b9-4797-80a7-de3beca270f4-utilities" (OuterVolumeSpecName: "utilities") pod "08bbb8ce-17b9-4797-80a7-de3beca270f4" (UID: "08bbb8ce-17b9-4797-80a7-de3beca270f4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.251840 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gh9mr"] Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.252099 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gh9mr" podUID="1c9a571e-db77-4303-96cc-c111add7626d" containerName="registry-server" containerID="cri-o://d11854ae7e58b3739a5b1b33d4e46c1c1b48da778db1b7e3f005b331cb2434b6" gracePeriod=2 Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.273350 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08bbb8ce-17b9-4797-80a7-de3beca270f4-kube-api-access-49srl" (OuterVolumeSpecName: "kube-api-access-49srl") pod "08bbb8ce-17b9-4797-80a7-de3beca270f4" (UID: "08bbb8ce-17b9-4797-80a7-de3beca270f4"). InnerVolumeSpecName "kube-api-access-49srl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.337257 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kr7td"] Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.337553 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-kr7td" podUID="cb551926-2f33-4616-a423-caee14e79f74" containerName="registry-server" containerID="cri-o://2a0ad6eb90cb3283291e18b1291d23c330c5079988eebcdbc75da3ecbe05dfd5" gracePeriod=2 Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.350483 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08bbb8ce-17b9-4797-80a7-de3beca270f4-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.350551 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-49srl\" (UniqueName: \"kubernetes.io/projected/08bbb8ce-17b9-4797-80a7-de3beca270f4-kube-api-access-49srl\") on node \"crc\" DevicePath \"\"" Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.413391 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08bbb8ce-17b9-4797-80a7-de3beca270f4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "08bbb8ce-17b9-4797-80a7-de3beca270f4" (UID: "08bbb8ce-17b9-4797-80a7-de3beca270f4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.454519 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08bbb8ce-17b9-4797-80a7-de3beca270f4-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.525090 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6g727"] Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.525991 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6g727" podUID="b53e3920-fe5e-45fe-9b2d-495a5cb16b2c" containerName="registry-server" containerID="cri-o://8b8951542a05f91dd43fa2161ca5c59912852a6eb9dcdc79d470c54c1dc552b8" gracePeriod=2 Nov 25 15:30:18 crc kubenswrapper[4890]: E1125 15:30:18.583471 4890 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2a0ad6eb90cb3283291e18b1291d23c330c5079988eebcdbc75da3ecbe05dfd5 is running failed: container process not found" containerID="2a0ad6eb90cb3283291e18b1291d23c330c5079988eebcdbc75da3ecbe05dfd5" cmd=["grpc_health_probe","-addr=:50051"] Nov 25 15:30:18 crc kubenswrapper[4890]: E1125 15:30:18.586021 4890 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2a0ad6eb90cb3283291e18b1291d23c330c5079988eebcdbc75da3ecbe05dfd5 is running failed: container process not found" containerID="2a0ad6eb90cb3283291e18b1291d23c330c5079988eebcdbc75da3ecbe05dfd5" cmd=["grpc_health_probe","-addr=:50051"] Nov 25 15:30:18 crc kubenswrapper[4890]: E1125 15:30:18.586519 4890 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2a0ad6eb90cb3283291e18b1291d23c330c5079988eebcdbc75da3ecbe05dfd5 is running failed: container process not found" containerID="2a0ad6eb90cb3283291e18b1291d23c330c5079988eebcdbc75da3ecbe05dfd5" cmd=["grpc_health_probe","-addr=:50051"] Nov 25 15:30:18 crc kubenswrapper[4890]: E1125 15:30:18.586586 4890 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2a0ad6eb90cb3283291e18b1291d23c330c5079988eebcdbc75da3ecbe05dfd5 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-kr7td" podUID="cb551926-2f33-4616-a423-caee14e79f74" containerName="registry-server" Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.664370 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rcbhc" Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.682052 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v8mj2" Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.760768 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00771e9e-5583-42ab-b834-4643d9bbdbec-catalog-content\") pod \"00771e9e-5583-42ab-b834-4643d9bbdbec\" (UID: \"00771e9e-5583-42ab-b834-4643d9bbdbec\") " Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.761521 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hr2md\" (UniqueName: \"kubernetes.io/projected/b18eb2e9-77b4-46ba-9e32-31c1e9b73963-kube-api-access-hr2md\") pod \"b18eb2e9-77b4-46ba-9e32-31c1e9b73963\" (UID: \"b18eb2e9-77b4-46ba-9e32-31c1e9b73963\") " Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.761667 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b18eb2e9-77b4-46ba-9e32-31c1e9b73963-catalog-content\") pod \"b18eb2e9-77b4-46ba-9e32-31c1e9b73963\" (UID: \"b18eb2e9-77b4-46ba-9e32-31c1e9b73963\") " Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.761703 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b18eb2e9-77b4-46ba-9e32-31c1e9b73963-utilities\") pod \"b18eb2e9-77b4-46ba-9e32-31c1e9b73963\" (UID: \"b18eb2e9-77b4-46ba-9e32-31c1e9b73963\") " Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.760907 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gh9mr" Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.761761 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7x9w4\" (UniqueName: \"kubernetes.io/projected/00771e9e-5583-42ab-b834-4643d9bbdbec-kube-api-access-7x9w4\") pod \"00771e9e-5583-42ab-b834-4643d9bbdbec\" (UID: \"00771e9e-5583-42ab-b834-4643d9bbdbec\") " Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.761827 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00771e9e-5583-42ab-b834-4643d9bbdbec-utilities\") pod \"00771e9e-5583-42ab-b834-4643d9bbdbec\" (UID: \"00771e9e-5583-42ab-b834-4643d9bbdbec\") " Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.763022 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b18eb2e9-77b4-46ba-9e32-31c1e9b73963-utilities" (OuterVolumeSpecName: "utilities") pod "b18eb2e9-77b4-46ba-9e32-31c1e9b73963" (UID: "b18eb2e9-77b4-46ba-9e32-31c1e9b73963"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.765290 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00771e9e-5583-42ab-b834-4643d9bbdbec-utilities" (OuterVolumeSpecName: "utilities") pod "00771e9e-5583-42ab-b834-4643d9bbdbec" (UID: "00771e9e-5583-42ab-b834-4643d9bbdbec"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.765336 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ddmtt"] Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.765546 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ddmtt" podUID="b9fbb88d-f5f9-41c6-9719-15be794dfb6c" containerName="registry-server" containerID="cri-o://bb0dcecc09d90d56e793929595f4f4c4631c9363c8d877b4b159f0abbbf8d5fa" gracePeriod=2 Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.773775 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b18eb2e9-77b4-46ba-9e32-31c1e9b73963-kube-api-access-hr2md" (OuterVolumeSpecName: "kube-api-access-hr2md") pod "b18eb2e9-77b4-46ba-9e32-31c1e9b73963" (UID: "b18eb2e9-77b4-46ba-9e32-31c1e9b73963"). InnerVolumeSpecName "kube-api-access-hr2md". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.777130 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00771e9e-5583-42ab-b834-4643d9bbdbec-kube-api-access-7x9w4" (OuterVolumeSpecName: "kube-api-access-7x9w4") pod "00771e9e-5583-42ab-b834-4643d9bbdbec" (UID: "00771e9e-5583-42ab-b834-4643d9bbdbec"). InnerVolumeSpecName "kube-api-access-7x9w4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.834398 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b18eb2e9-77b4-46ba-9e32-31c1e9b73963-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b18eb2e9-77b4-46ba-9e32-31c1e9b73963" (UID: "b18eb2e9-77b4-46ba-9e32-31c1e9b73963"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.842259 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00771e9e-5583-42ab-b834-4643d9bbdbec-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "00771e9e-5583-42ab-b834-4643d9bbdbec" (UID: "00771e9e-5583-42ab-b834-4643d9bbdbec"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.867141 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c9a571e-db77-4303-96cc-c111add7626d-utilities\") pod \"1c9a571e-db77-4303-96cc-c111add7626d\" (UID: \"1c9a571e-db77-4303-96cc-c111add7626d\") " Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.867633 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c9a571e-db77-4303-96cc-c111add7626d-catalog-content\") pod \"1c9a571e-db77-4303-96cc-c111add7626d\" (UID: \"1c9a571e-db77-4303-96cc-c111add7626d\") " Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.873313 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-76jx8\" (UniqueName: \"kubernetes.io/projected/1c9a571e-db77-4303-96cc-c111add7626d-kube-api-access-76jx8\") pod \"1c9a571e-db77-4303-96cc-c111add7626d\" (UID: \"1c9a571e-db77-4303-96cc-c111add7626d\") " Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.874375 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00771e9e-5583-42ab-b834-4643d9bbdbec-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.874475 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hr2md\" (UniqueName: \"kubernetes.io/projected/b18eb2e9-77b4-46ba-9e32-31c1e9b73963-kube-api-access-hr2md\") on node \"crc\" DevicePath \"\"" Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.874548 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b18eb2e9-77b4-46ba-9e32-31c1e9b73963-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.887468 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b18eb2e9-77b4-46ba-9e32-31c1e9b73963-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.887498 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7x9w4\" (UniqueName: \"kubernetes.io/projected/00771e9e-5583-42ab-b834-4643d9bbdbec-kube-api-access-7x9w4\") on node \"crc\" DevicePath \"\"" Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.887511 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00771e9e-5583-42ab-b834-4643d9bbdbec-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.868876 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c9a571e-db77-4303-96cc-c111add7626d-utilities" (OuterVolumeSpecName: "utilities") pod "1c9a571e-db77-4303-96cc-c111add7626d" (UID: "1c9a571e-db77-4303-96cc-c111add7626d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.883810 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c9a571e-db77-4303-96cc-c111add7626d-kube-api-access-76jx8" (OuterVolumeSpecName: "kube-api-access-76jx8") pod "1c9a571e-db77-4303-96cc-c111add7626d" (UID: "1c9a571e-db77-4303-96cc-c111add7626d"). InnerVolumeSpecName "kube-api-access-76jx8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.914364 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c9a571e-db77-4303-96cc-c111add7626d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1c9a571e-db77-4303-96cc-c111add7626d" (UID: "1c9a571e-db77-4303-96cc-c111add7626d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.923937 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-g8wbt"] Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.924522 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-g8wbt" podUID="7c070ef9-daba-4ae9-ac05-dff449baf01a" containerName="registry-server" containerID="cri-o://74f159210a58c58af518d6ce062a98d2955b3a8f42e39be962750a76214a2f7f" gracePeriod=2 Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.969955 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kr7td" Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.994855 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-76jx8\" (UniqueName: \"kubernetes.io/projected/1c9a571e-db77-4303-96cc-c111add7626d-kube-api-access-76jx8\") on node \"crc\" DevicePath \"\"" Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.994938 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c9a571e-db77-4303-96cc-c111add7626d-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 15:30:18 crc kubenswrapper[4890]: I1125 15:30:18.994954 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c9a571e-db77-4303-96cc-c111add7626d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.059041 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6g727" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.097062 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb551926-2f33-4616-a423-caee14e79f74-utilities\") pod \"cb551926-2f33-4616-a423-caee14e79f74\" (UID: \"cb551926-2f33-4616-a423-caee14e79f74\") " Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.097143 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2plp\" (UniqueName: \"kubernetes.io/projected/cb551926-2f33-4616-a423-caee14e79f74-kube-api-access-c2plp\") pod \"cb551926-2f33-4616-a423-caee14e79f74\" (UID: \"cb551926-2f33-4616-a423-caee14e79f74\") " Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.097204 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb551926-2f33-4616-a423-caee14e79f74-catalog-content\") pod \"cb551926-2f33-4616-a423-caee14e79f74\" (UID: \"cb551926-2f33-4616-a423-caee14e79f74\") " Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.097561 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb551926-2f33-4616-a423-caee14e79f74-utilities" (OuterVolumeSpecName: "utilities") pod "cb551926-2f33-4616-a423-caee14e79f74" (UID: "cb551926-2f33-4616-a423-caee14e79f74"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.099221 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb551926-2f33-4616-a423-caee14e79f74-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.112423 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb551926-2f33-4616-a423-caee14e79f74-kube-api-access-c2plp" (OuterVolumeSpecName: "kube-api-access-c2plp") pod "cb551926-2f33-4616-a423-caee14e79f74" (UID: "cb551926-2f33-4616-a423-caee14e79f74"). InnerVolumeSpecName "kube-api-access-c2plp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.163940 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb551926-2f33-4616-a423-caee14e79f74-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cb551926-2f33-4616-a423-caee14e79f74" (UID: "cb551926-2f33-4616-a423-caee14e79f74"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.170293 4890 generic.go:334] "Generic (PLEG): container finished" podID="7c070ef9-daba-4ae9-ac05-dff449baf01a" containerID="74f159210a58c58af518d6ce062a98d2955b3a8f42e39be962750a76214a2f7f" exitCode=0 Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.170328 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g8wbt" event={"ID":"7c070ef9-daba-4ae9-ac05-dff449baf01a","Type":"ContainerDied","Data":"74f159210a58c58af518d6ce062a98d2955b3a8f42e39be962750a76214a2f7f"} Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.173841 4890 generic.go:334] "Generic (PLEG): container finished" podID="00771e9e-5583-42ab-b834-4643d9bbdbec" containerID="583629b7aeb99a0f6d9e57cd9965500d687d0958872fcdd86b78e86046b11f5a" exitCode=0 Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.173879 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v8mj2" event={"ID":"00771e9e-5583-42ab-b834-4643d9bbdbec","Type":"ContainerDied","Data":"583629b7aeb99a0f6d9e57cd9965500d687d0958872fcdd86b78e86046b11f5a"} Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.173903 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v8mj2" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.173915 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v8mj2" event={"ID":"00771e9e-5583-42ab-b834-4643d9bbdbec","Type":"ContainerDied","Data":"49174bc1bc2916c23aad34ea827d1d0da668cc5ce08d9282197851218bf6fc0e"} Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.173944 4890 scope.go:117] "RemoveContainer" containerID="583629b7aeb99a0f6d9e57cd9965500d687d0958872fcdd86b78e86046b11f5a" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.174118 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mtfgl" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.181928 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rcbhc" event={"ID":"b18eb2e9-77b4-46ba-9e32-31c1e9b73963","Type":"ContainerDied","Data":"a758e8482e3d183aa59e2885cbf5a7e5cd809f48c007a4712f6979acf2a9337d"} Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.182052 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rcbhc" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.186012 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h44zl" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.207133 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b53e3920-fe5e-45fe-9b2d-495a5cb16b2c-utilities\") pod \"b53e3920-fe5e-45fe-9b2d-495a5cb16b2c\" (UID: \"b53e3920-fe5e-45fe-9b2d-495a5cb16b2c\") " Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.207561 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b53e3920-fe5e-45fe-9b2d-495a5cb16b2c-catalog-content\") pod \"b53e3920-fe5e-45fe-9b2d-495a5cb16b2c\" (UID: \"b53e3920-fe5e-45fe-9b2d-495a5cb16b2c\") " Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.207672 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pv9s7\" (UniqueName: \"kubernetes.io/projected/b53e3920-fe5e-45fe-9b2d-495a5cb16b2c-kube-api-access-pv9s7\") pod \"b53e3920-fe5e-45fe-9b2d-495a5cb16b2c\" (UID: \"b53e3920-fe5e-45fe-9b2d-495a5cb16b2c\") " Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.208385 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2plp\" (UniqueName: \"kubernetes.io/projected/cb551926-2f33-4616-a423-caee14e79f74-kube-api-access-c2plp\") on node \"crc\" DevicePath \"\"" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.208404 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb551926-2f33-4616-a423-caee14e79f74-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.208895 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b53e3920-fe5e-45fe-9b2d-495a5cb16b2c-utilities" (OuterVolumeSpecName: "utilities") pod "b53e3920-fe5e-45fe-9b2d-495a5cb16b2c" (UID: "b53e3920-fe5e-45fe-9b2d-495a5cb16b2c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.214090 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b53e3920-fe5e-45fe-9b2d-495a5cb16b2c-kube-api-access-pv9s7" (OuterVolumeSpecName: "kube-api-access-pv9s7") pod "b53e3920-fe5e-45fe-9b2d-495a5cb16b2c" (UID: "b53e3920-fe5e-45fe-9b2d-495a5cb16b2c"). InnerVolumeSpecName "kube-api-access-pv9s7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.215210 4890 scope.go:117] "RemoveContainer" containerID="25383107f01eedb261016e63aaa8d7e88ddff63b3170d5bdc523e921f50e8778" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.215886 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gh9mr" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.215730 4890 generic.go:334] "Generic (PLEG): container finished" podID="1c9a571e-db77-4303-96cc-c111add7626d" containerID="d11854ae7e58b3739a5b1b33d4e46c1c1b48da778db1b7e3f005b331cb2434b6" exitCode=0 Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.220423 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gh9mr" event={"ID":"1c9a571e-db77-4303-96cc-c111add7626d","Type":"ContainerDied","Data":"d11854ae7e58b3739a5b1b33d4e46c1c1b48da778db1b7e3f005b331cb2434b6"} Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.243256 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gh9mr" event={"ID":"1c9a571e-db77-4303-96cc-c111add7626d","Type":"ContainerDied","Data":"88fe258dfd5b32f740f82794e6235d890a94e9f5a650318106758fa7930b9a8b"} Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.255781 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-48z2b" podUID="eacc1f05-701e-49af-b2d7-9c049f6cddb9" containerName="registry-server" probeResult="failure" output=< Nov 25 15:30:19 crc kubenswrapper[4890]: timeout: failed to connect service ":50051" within 1s Nov 25 15:30:19 crc kubenswrapper[4890]: > Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.274780 4890 generic.go:334] "Generic (PLEG): container finished" podID="cb551926-2f33-4616-a423-caee14e79f74" containerID="2a0ad6eb90cb3283291e18b1291d23c330c5079988eebcdbc75da3ecbe05dfd5" exitCode=0 Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.274806 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-v8mj2"] Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.275190 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kr7td" event={"ID":"cb551926-2f33-4616-a423-caee14e79f74","Type":"ContainerDied","Data":"2a0ad6eb90cb3283291e18b1291d23c330c5079988eebcdbc75da3ecbe05dfd5"} Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.275297 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kr7td" event={"ID":"cb551926-2f33-4616-a423-caee14e79f74","Type":"ContainerDied","Data":"50abe642979368e933bd97e0e71a88affb9cd7aad1c209c05f9aa209991764ef"} Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.274892 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kr7td" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.279372 4890 generic.go:334] "Generic (PLEG): container finished" podID="b9fbb88d-f5f9-41c6-9719-15be794dfb6c" containerID="bb0dcecc09d90d56e793929595f4f4c4631c9363c8d877b4b159f0abbbf8d5fa" exitCode=0 Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.279992 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ddmtt" event={"ID":"b9fbb88d-f5f9-41c6-9719-15be794dfb6c","Type":"ContainerDied","Data":"bb0dcecc09d90d56e793929595f4f4c4631c9363c8d877b4b159f0abbbf8d5fa"} Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.285838 4890 generic.go:334] "Generic (PLEG): container finished" podID="b53e3920-fe5e-45fe-9b2d-495a5cb16b2c" containerID="8b8951542a05f91dd43fa2161ca5c59912852a6eb9dcdc79d470c54c1dc552b8" exitCode=0 Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.285877 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6g727" event={"ID":"b53e3920-fe5e-45fe-9b2d-495a5cb16b2c","Type":"ContainerDied","Data":"8b8951542a05f91dd43fa2161ca5c59912852a6eb9dcdc79d470c54c1dc552b8"} Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.285931 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6g727" event={"ID":"b53e3920-fe5e-45fe-9b2d-495a5cb16b2c","Type":"ContainerDied","Data":"ad3647df1b566cc7c46012aacf4ae79d207d5a05e18409b5af4d99ed9799d7ec"} Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.286054 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6g727" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.291964 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-v8mj2"] Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.325375 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b53e3920-fe5e-45fe-9b2d-495a5cb16b2c-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.325504 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pv9s7\" (UniqueName: \"kubernetes.io/projected/b53e3920-fe5e-45fe-9b2d-495a5cb16b2c-kube-api-access-pv9s7\") on node \"crc\" DevicePath \"\"" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.325913 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mtfgl" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.357592 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b53e3920-fe5e-45fe-9b2d-495a5cb16b2c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b53e3920-fe5e-45fe-9b2d-495a5cb16b2c" (UID: "b53e3920-fe5e-45fe-9b2d-495a5cb16b2c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.381631 4890 scope.go:117] "RemoveContainer" containerID="aba18652c1832a9d5351cea01a0e87328ddbaca9fafe0ce1be9bb66182cfebae" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.424950 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ddmtt" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.427180 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b53e3920-fe5e-45fe-9b2d-495a5cb16b2c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.445553 4890 scope.go:117] "RemoveContainer" containerID="583629b7aeb99a0f6d9e57cd9965500d687d0958872fcdd86b78e86046b11f5a" Nov 25 15:30:19 crc kubenswrapper[4890]: E1125 15:30:19.446369 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"583629b7aeb99a0f6d9e57cd9965500d687d0958872fcdd86b78e86046b11f5a\": container with ID starting with 583629b7aeb99a0f6d9e57cd9965500d687d0958872fcdd86b78e86046b11f5a not found: ID does not exist" containerID="583629b7aeb99a0f6d9e57cd9965500d687d0958872fcdd86b78e86046b11f5a" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.446418 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"583629b7aeb99a0f6d9e57cd9965500d687d0958872fcdd86b78e86046b11f5a"} err="failed to get container status \"583629b7aeb99a0f6d9e57cd9965500d687d0958872fcdd86b78e86046b11f5a\": rpc error: code = NotFound desc = could not find container \"583629b7aeb99a0f6d9e57cd9965500d687d0958872fcdd86b78e86046b11f5a\": container with ID starting with 583629b7aeb99a0f6d9e57cd9965500d687d0958872fcdd86b78e86046b11f5a not found: ID does not exist" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.446451 4890 scope.go:117] "RemoveContainer" containerID="25383107f01eedb261016e63aaa8d7e88ddff63b3170d5bdc523e921f50e8778" Nov 25 15:30:19 crc kubenswrapper[4890]: E1125 15:30:19.446911 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25383107f01eedb261016e63aaa8d7e88ddff63b3170d5bdc523e921f50e8778\": container with ID starting with 25383107f01eedb261016e63aaa8d7e88ddff63b3170d5bdc523e921f50e8778 not found: ID does not exist" containerID="25383107f01eedb261016e63aaa8d7e88ddff63b3170d5bdc523e921f50e8778" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.446946 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25383107f01eedb261016e63aaa8d7e88ddff63b3170d5bdc523e921f50e8778"} err="failed to get container status \"25383107f01eedb261016e63aaa8d7e88ddff63b3170d5bdc523e921f50e8778\": rpc error: code = NotFound desc = could not find container \"25383107f01eedb261016e63aaa8d7e88ddff63b3170d5bdc523e921f50e8778\": container with ID starting with 25383107f01eedb261016e63aaa8d7e88ddff63b3170d5bdc523e921f50e8778 not found: ID does not exist" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.447006 4890 scope.go:117] "RemoveContainer" containerID="aba18652c1832a9d5351cea01a0e87328ddbaca9fafe0ce1be9bb66182cfebae" Nov 25 15:30:19 crc kubenswrapper[4890]: E1125 15:30:19.447344 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aba18652c1832a9d5351cea01a0e87328ddbaca9fafe0ce1be9bb66182cfebae\": container with ID starting with aba18652c1832a9d5351cea01a0e87328ddbaca9fafe0ce1be9bb66182cfebae not found: ID does not exist" containerID="aba18652c1832a9d5351cea01a0e87328ddbaca9fafe0ce1be9bb66182cfebae" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.447371 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aba18652c1832a9d5351cea01a0e87328ddbaca9fafe0ce1be9bb66182cfebae"} err="failed to get container status \"aba18652c1832a9d5351cea01a0e87328ddbaca9fafe0ce1be9bb66182cfebae\": rpc error: code = NotFound desc = could not find container \"aba18652c1832a9d5351cea01a0e87328ddbaca9fafe0ce1be9bb66182cfebae\": container with ID starting with aba18652c1832a9d5351cea01a0e87328ddbaca9fafe0ce1be9bb66182cfebae not found: ID does not exist" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.447386 4890 scope.go:117] "RemoveContainer" containerID="a4ccb66699200fcad030dec7f35c81f1826dbfa3d9aede9a2806ab3b57bd0a09" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.529278 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9fbb88d-f5f9-41c6-9719-15be794dfb6c-utilities\") pod \"b9fbb88d-f5f9-41c6-9719-15be794dfb6c\" (UID: \"b9fbb88d-f5f9-41c6-9719-15be794dfb6c\") " Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.529470 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9fbb88d-f5f9-41c6-9719-15be794dfb6c-catalog-content\") pod \"b9fbb88d-f5f9-41c6-9719-15be794dfb6c\" (UID: \"b9fbb88d-f5f9-41c6-9719-15be794dfb6c\") " Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.529500 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9tvq\" (UniqueName: \"kubernetes.io/projected/b9fbb88d-f5f9-41c6-9719-15be794dfb6c-kube-api-access-j9tvq\") pod \"b9fbb88d-f5f9-41c6-9719-15be794dfb6c\" (UID: \"b9fbb88d-f5f9-41c6-9719-15be794dfb6c\") " Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.530365 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9fbb88d-f5f9-41c6-9719-15be794dfb6c-utilities" (OuterVolumeSpecName: "utilities") pod "b9fbb88d-f5f9-41c6-9719-15be794dfb6c" (UID: "b9fbb88d-f5f9-41c6-9719-15be794dfb6c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.535358 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9fbb88d-f5f9-41c6-9719-15be794dfb6c-kube-api-access-j9tvq" (OuterVolumeSpecName: "kube-api-access-j9tvq") pod "b9fbb88d-f5f9-41c6-9719-15be794dfb6c" (UID: "b9fbb88d-f5f9-41c6-9719-15be794dfb6c"). InnerVolumeSpecName "kube-api-access-j9tvq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.586553 4890 scope.go:117] "RemoveContainer" containerID="626d9b201a0e32fd3d13e65ebe08857bf56a1aa7841578b259565881cdd9fbdf" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.587807 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-h44zl"] Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.595580 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g8wbt" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.601991 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9fbb88d-f5f9-41c6-9719-15be794dfb6c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b9fbb88d-f5f9-41c6-9719-15be794dfb6c" (UID: "b9fbb88d-f5f9-41c6-9719-15be794dfb6c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.602761 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-h44zl"] Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.620833 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gh9mr"] Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.627194 4890 scope.go:117] "RemoveContainer" containerID="3b02fe5096c64727128fcb100ea22d7cfae1d8624fbb52019491522ac365aa2b" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.636256 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9fbb88d-f5f9-41c6-9719-15be794dfb6c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.636630 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9tvq\" (UniqueName: \"kubernetes.io/projected/b9fbb88d-f5f9-41c6-9719-15be794dfb6c-kube-api-access-j9tvq\") on node \"crc\" DevicePath \"\"" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.636742 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9fbb88d-f5f9-41c6-9719-15be794dfb6c-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.640963 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gh9mr"] Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.661640 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rcbhc"] Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.672482 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rcbhc"] Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.675695 4890 scope.go:117] "RemoveContainer" containerID="d11854ae7e58b3739a5b1b33d4e46c1c1b48da778db1b7e3f005b331cb2434b6" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.687511 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kr7td"] Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.702444 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-kr7td"] Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.710666 4890 scope.go:117] "RemoveContainer" containerID="bd5877a3900425fe22977396b122dbfc9504e980f0f73e0be858e810a14aab51" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.730881 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6g727"] Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.731558 4890 scope.go:117] "RemoveContainer" containerID="8c20e9b751cfadd82d3cf4b5f40aaf1d3b45c26081dd57e7fcf8fe32e1140f99" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.739196 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c070ef9-daba-4ae9-ac05-dff449baf01a-catalog-content\") pod \"7c070ef9-daba-4ae9-ac05-dff449baf01a\" (UID: \"7c070ef9-daba-4ae9-ac05-dff449baf01a\") " Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.739251 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rzqtm\" (UniqueName: \"kubernetes.io/projected/7c070ef9-daba-4ae9-ac05-dff449baf01a-kube-api-access-rzqtm\") pod \"7c070ef9-daba-4ae9-ac05-dff449baf01a\" (UID: \"7c070ef9-daba-4ae9-ac05-dff449baf01a\") " Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.739420 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c070ef9-daba-4ae9-ac05-dff449baf01a-utilities\") pod \"7c070ef9-daba-4ae9-ac05-dff449baf01a\" (UID: \"7c070ef9-daba-4ae9-ac05-dff449baf01a\") " Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.740235 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c070ef9-daba-4ae9-ac05-dff449baf01a-utilities" (OuterVolumeSpecName: "utilities") pod "7c070ef9-daba-4ae9-ac05-dff449baf01a" (UID: "7c070ef9-daba-4ae9-ac05-dff449baf01a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.742713 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c070ef9-daba-4ae9-ac05-dff449baf01a-kube-api-access-rzqtm" (OuterVolumeSpecName: "kube-api-access-rzqtm") pod "7c070ef9-daba-4ae9-ac05-dff449baf01a" (UID: "7c070ef9-daba-4ae9-ac05-dff449baf01a"). InnerVolumeSpecName "kube-api-access-rzqtm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.743187 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6g727"] Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.752816 4890 scope.go:117] "RemoveContainer" containerID="d11854ae7e58b3739a5b1b33d4e46c1c1b48da778db1b7e3f005b331cb2434b6" Nov 25 15:30:19 crc kubenswrapper[4890]: E1125 15:30:19.753413 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d11854ae7e58b3739a5b1b33d4e46c1c1b48da778db1b7e3f005b331cb2434b6\": container with ID starting with d11854ae7e58b3739a5b1b33d4e46c1c1b48da778db1b7e3f005b331cb2434b6 not found: ID does not exist" containerID="d11854ae7e58b3739a5b1b33d4e46c1c1b48da778db1b7e3f005b331cb2434b6" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.753457 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d11854ae7e58b3739a5b1b33d4e46c1c1b48da778db1b7e3f005b331cb2434b6"} err="failed to get container status \"d11854ae7e58b3739a5b1b33d4e46c1c1b48da778db1b7e3f005b331cb2434b6\": rpc error: code = NotFound desc = could not find container \"d11854ae7e58b3739a5b1b33d4e46c1c1b48da778db1b7e3f005b331cb2434b6\": container with ID starting with d11854ae7e58b3739a5b1b33d4e46c1c1b48da778db1b7e3f005b331cb2434b6 not found: ID does not exist" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.753492 4890 scope.go:117] "RemoveContainer" containerID="bd5877a3900425fe22977396b122dbfc9504e980f0f73e0be858e810a14aab51" Nov 25 15:30:19 crc kubenswrapper[4890]: E1125 15:30:19.754090 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd5877a3900425fe22977396b122dbfc9504e980f0f73e0be858e810a14aab51\": container with ID starting with bd5877a3900425fe22977396b122dbfc9504e980f0f73e0be858e810a14aab51 not found: ID does not exist" containerID="bd5877a3900425fe22977396b122dbfc9504e980f0f73e0be858e810a14aab51" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.754214 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd5877a3900425fe22977396b122dbfc9504e980f0f73e0be858e810a14aab51"} err="failed to get container status \"bd5877a3900425fe22977396b122dbfc9504e980f0f73e0be858e810a14aab51\": rpc error: code = NotFound desc = could not find container \"bd5877a3900425fe22977396b122dbfc9504e980f0f73e0be858e810a14aab51\": container with ID starting with bd5877a3900425fe22977396b122dbfc9504e980f0f73e0be858e810a14aab51 not found: ID does not exist" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.754256 4890 scope.go:117] "RemoveContainer" containerID="8c20e9b751cfadd82d3cf4b5f40aaf1d3b45c26081dd57e7fcf8fe32e1140f99" Nov 25 15:30:19 crc kubenswrapper[4890]: E1125 15:30:19.754669 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c20e9b751cfadd82d3cf4b5f40aaf1d3b45c26081dd57e7fcf8fe32e1140f99\": container with ID starting with 8c20e9b751cfadd82d3cf4b5f40aaf1d3b45c26081dd57e7fcf8fe32e1140f99 not found: ID does not exist" containerID="8c20e9b751cfadd82d3cf4b5f40aaf1d3b45c26081dd57e7fcf8fe32e1140f99" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.754697 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c20e9b751cfadd82d3cf4b5f40aaf1d3b45c26081dd57e7fcf8fe32e1140f99"} err="failed to get container status \"8c20e9b751cfadd82d3cf4b5f40aaf1d3b45c26081dd57e7fcf8fe32e1140f99\": rpc error: code = NotFound desc = could not find container \"8c20e9b751cfadd82d3cf4b5f40aaf1d3b45c26081dd57e7fcf8fe32e1140f99\": container with ID starting with 8c20e9b751cfadd82d3cf4b5f40aaf1d3b45c26081dd57e7fcf8fe32e1140f99 not found: ID does not exist" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.754713 4890 scope.go:117] "RemoveContainer" containerID="2a0ad6eb90cb3283291e18b1291d23c330c5079988eebcdbc75da3ecbe05dfd5" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.779781 4890 scope.go:117] "RemoveContainer" containerID="b2e5c93300d37e63edce4624180363409c7c85e5902027f073ae462124ddb757" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.798117 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c070ef9-daba-4ae9-ac05-dff449baf01a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7c070ef9-daba-4ae9-ac05-dff449baf01a" (UID: "7c070ef9-daba-4ae9-ac05-dff449baf01a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.805776 4890 scope.go:117] "RemoveContainer" containerID="324d3b471e383e8d75a04f41b4d279d80982950aa20b6522732f1ccd7ba394a8" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.828829 4890 scope.go:117] "RemoveContainer" containerID="2a0ad6eb90cb3283291e18b1291d23c330c5079988eebcdbc75da3ecbe05dfd5" Nov 25 15:30:19 crc kubenswrapper[4890]: E1125 15:30:19.829483 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a0ad6eb90cb3283291e18b1291d23c330c5079988eebcdbc75da3ecbe05dfd5\": container with ID starting with 2a0ad6eb90cb3283291e18b1291d23c330c5079988eebcdbc75da3ecbe05dfd5 not found: ID does not exist" containerID="2a0ad6eb90cb3283291e18b1291d23c330c5079988eebcdbc75da3ecbe05dfd5" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.829556 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a0ad6eb90cb3283291e18b1291d23c330c5079988eebcdbc75da3ecbe05dfd5"} err="failed to get container status \"2a0ad6eb90cb3283291e18b1291d23c330c5079988eebcdbc75da3ecbe05dfd5\": rpc error: code = NotFound desc = could not find container \"2a0ad6eb90cb3283291e18b1291d23c330c5079988eebcdbc75da3ecbe05dfd5\": container with ID starting with 2a0ad6eb90cb3283291e18b1291d23c330c5079988eebcdbc75da3ecbe05dfd5 not found: ID does not exist" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.829586 4890 scope.go:117] "RemoveContainer" containerID="b2e5c93300d37e63edce4624180363409c7c85e5902027f073ae462124ddb757" Nov 25 15:30:19 crc kubenswrapper[4890]: E1125 15:30:19.830495 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2e5c93300d37e63edce4624180363409c7c85e5902027f073ae462124ddb757\": container with ID starting with b2e5c93300d37e63edce4624180363409c7c85e5902027f073ae462124ddb757 not found: ID does not exist" containerID="b2e5c93300d37e63edce4624180363409c7c85e5902027f073ae462124ddb757" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.830539 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2e5c93300d37e63edce4624180363409c7c85e5902027f073ae462124ddb757"} err="failed to get container status \"b2e5c93300d37e63edce4624180363409c7c85e5902027f073ae462124ddb757\": rpc error: code = NotFound desc = could not find container \"b2e5c93300d37e63edce4624180363409c7c85e5902027f073ae462124ddb757\": container with ID starting with b2e5c93300d37e63edce4624180363409c7c85e5902027f073ae462124ddb757 not found: ID does not exist" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.830575 4890 scope.go:117] "RemoveContainer" containerID="324d3b471e383e8d75a04f41b4d279d80982950aa20b6522732f1ccd7ba394a8" Nov 25 15:30:19 crc kubenswrapper[4890]: E1125 15:30:19.830892 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"324d3b471e383e8d75a04f41b4d279d80982950aa20b6522732f1ccd7ba394a8\": container with ID starting with 324d3b471e383e8d75a04f41b4d279d80982950aa20b6522732f1ccd7ba394a8 not found: ID does not exist" containerID="324d3b471e383e8d75a04f41b4d279d80982950aa20b6522732f1ccd7ba394a8" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.830920 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"324d3b471e383e8d75a04f41b4d279d80982950aa20b6522732f1ccd7ba394a8"} err="failed to get container status \"324d3b471e383e8d75a04f41b4d279d80982950aa20b6522732f1ccd7ba394a8\": rpc error: code = NotFound desc = could not find container \"324d3b471e383e8d75a04f41b4d279d80982950aa20b6522732f1ccd7ba394a8\": container with ID starting with 324d3b471e383e8d75a04f41b4d279d80982950aa20b6522732f1ccd7ba394a8 not found: ID does not exist" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.830937 4890 scope.go:117] "RemoveContainer" containerID="8b8951542a05f91dd43fa2161ca5c59912852a6eb9dcdc79d470c54c1dc552b8" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.841628 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c070ef9-daba-4ae9-ac05-dff449baf01a-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.841673 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c070ef9-daba-4ae9-ac05-dff449baf01a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.841688 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rzqtm\" (UniqueName: \"kubernetes.io/projected/7c070ef9-daba-4ae9-ac05-dff449baf01a-kube-api-access-rzqtm\") on node \"crc\" DevicePath \"\"" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.852992 4890 scope.go:117] "RemoveContainer" containerID="32ed94c9ed63590d671f1a4e6483de3d71759ff85d32e388abb90a0e81bd77fe" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.875950 4890 scope.go:117] "RemoveContainer" containerID="82e3cdda2919b11d5bf605afb47f59af1d0184f90723c6570cf19eb01d175be1" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.898357 4890 scope.go:117] "RemoveContainer" containerID="8b8951542a05f91dd43fa2161ca5c59912852a6eb9dcdc79d470c54c1dc552b8" Nov 25 15:30:19 crc kubenswrapper[4890]: E1125 15:30:19.898884 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b8951542a05f91dd43fa2161ca5c59912852a6eb9dcdc79d470c54c1dc552b8\": container with ID starting with 8b8951542a05f91dd43fa2161ca5c59912852a6eb9dcdc79d470c54c1dc552b8 not found: ID does not exist" containerID="8b8951542a05f91dd43fa2161ca5c59912852a6eb9dcdc79d470c54c1dc552b8" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.898924 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b8951542a05f91dd43fa2161ca5c59912852a6eb9dcdc79d470c54c1dc552b8"} err="failed to get container status \"8b8951542a05f91dd43fa2161ca5c59912852a6eb9dcdc79d470c54c1dc552b8\": rpc error: code = NotFound desc = could not find container \"8b8951542a05f91dd43fa2161ca5c59912852a6eb9dcdc79d470c54c1dc552b8\": container with ID starting with 8b8951542a05f91dd43fa2161ca5c59912852a6eb9dcdc79d470c54c1dc552b8 not found: ID does not exist" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.898954 4890 scope.go:117] "RemoveContainer" containerID="32ed94c9ed63590d671f1a4e6483de3d71759ff85d32e388abb90a0e81bd77fe" Nov 25 15:30:19 crc kubenswrapper[4890]: E1125 15:30:19.899360 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32ed94c9ed63590d671f1a4e6483de3d71759ff85d32e388abb90a0e81bd77fe\": container with ID starting with 32ed94c9ed63590d671f1a4e6483de3d71759ff85d32e388abb90a0e81bd77fe not found: ID does not exist" containerID="32ed94c9ed63590d671f1a4e6483de3d71759ff85d32e388abb90a0e81bd77fe" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.899399 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32ed94c9ed63590d671f1a4e6483de3d71759ff85d32e388abb90a0e81bd77fe"} err="failed to get container status \"32ed94c9ed63590d671f1a4e6483de3d71759ff85d32e388abb90a0e81bd77fe\": rpc error: code = NotFound desc = could not find container \"32ed94c9ed63590d671f1a4e6483de3d71759ff85d32e388abb90a0e81bd77fe\": container with ID starting with 32ed94c9ed63590d671f1a4e6483de3d71759ff85d32e388abb90a0e81bd77fe not found: ID does not exist" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.899429 4890 scope.go:117] "RemoveContainer" containerID="82e3cdda2919b11d5bf605afb47f59af1d0184f90723c6570cf19eb01d175be1" Nov 25 15:30:19 crc kubenswrapper[4890]: E1125 15:30:19.899985 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82e3cdda2919b11d5bf605afb47f59af1d0184f90723c6570cf19eb01d175be1\": container with ID starting with 82e3cdda2919b11d5bf605afb47f59af1d0184f90723c6570cf19eb01d175be1 not found: ID does not exist" containerID="82e3cdda2919b11d5bf605afb47f59af1d0184f90723c6570cf19eb01d175be1" Nov 25 15:30:19 crc kubenswrapper[4890]: I1125 15:30:19.900022 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82e3cdda2919b11d5bf605afb47f59af1d0184f90723c6570cf19eb01d175be1"} err="failed to get container status \"82e3cdda2919b11d5bf605afb47f59af1d0184f90723c6570cf19eb01d175be1\": rpc error: code = NotFound desc = could not find container \"82e3cdda2919b11d5bf605afb47f59af1d0184f90723c6570cf19eb01d175be1\": container with ID starting with 82e3cdda2919b11d5bf605afb47f59af1d0184f90723c6570cf19eb01d175be1 not found: ID does not exist" Nov 25 15:30:20 crc kubenswrapper[4890]: I1125 15:30:20.186485 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00771e9e-5583-42ab-b834-4643d9bbdbec" path="/var/lib/kubelet/pods/00771e9e-5583-42ab-b834-4643d9bbdbec/volumes" Nov 25 15:30:20 crc kubenswrapper[4890]: I1125 15:30:20.187216 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08bbb8ce-17b9-4797-80a7-de3beca270f4" path="/var/lib/kubelet/pods/08bbb8ce-17b9-4797-80a7-de3beca270f4/volumes" Nov 25 15:30:20 crc kubenswrapper[4890]: I1125 15:30:20.188006 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c9a571e-db77-4303-96cc-c111add7626d" path="/var/lib/kubelet/pods/1c9a571e-db77-4303-96cc-c111add7626d/volumes" Nov 25 15:30:20 crc kubenswrapper[4890]: I1125 15:30:20.189179 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b18eb2e9-77b4-46ba-9e32-31c1e9b73963" path="/var/lib/kubelet/pods/b18eb2e9-77b4-46ba-9e32-31c1e9b73963/volumes" Nov 25 15:30:20 crc kubenswrapper[4890]: I1125 15:30:20.189954 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b53e3920-fe5e-45fe-9b2d-495a5cb16b2c" path="/var/lib/kubelet/pods/b53e3920-fe5e-45fe-9b2d-495a5cb16b2c/volumes" Nov 25 15:30:20 crc kubenswrapper[4890]: I1125 15:30:20.191066 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb551926-2f33-4616-a423-caee14e79f74" path="/var/lib/kubelet/pods/cb551926-2f33-4616-a423-caee14e79f74/volumes" Nov 25 15:30:20 crc kubenswrapper[4890]: I1125 15:30:20.303817 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g8wbt" event={"ID":"7c070ef9-daba-4ae9-ac05-dff449baf01a","Type":"ContainerDied","Data":"811aa1444ea14e66014986dfab6ee4501d180834bd5711daad15c99b109dc58c"} Nov 25 15:30:20 crc kubenswrapper[4890]: I1125 15:30:20.303873 4890 scope.go:117] "RemoveContainer" containerID="74f159210a58c58af518d6ce062a98d2955b3a8f42e39be962750a76214a2f7f" Nov 25 15:30:20 crc kubenswrapper[4890]: I1125 15:30:20.304149 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g8wbt" Nov 25 15:30:20 crc kubenswrapper[4890]: I1125 15:30:20.316006 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ddmtt" event={"ID":"b9fbb88d-f5f9-41c6-9719-15be794dfb6c","Type":"ContainerDied","Data":"98c4050c253d51989b1845008d34976e3e890d2b80a459e1ba7166dbd9890e80"} Nov 25 15:30:20 crc kubenswrapper[4890]: I1125 15:30:20.316062 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ddmtt" Nov 25 15:30:20 crc kubenswrapper[4890]: I1125 15:30:20.340050 4890 scope.go:117] "RemoveContainer" containerID="46a904cda6ea2d958239b096888450e7894e0315371384e6fb9a6ae541dc625b" Nov 25 15:30:20 crc kubenswrapper[4890]: I1125 15:30:20.342474 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-g8wbt"] Nov 25 15:30:20 crc kubenswrapper[4890]: I1125 15:30:20.355526 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-g8wbt"] Nov 25 15:30:20 crc kubenswrapper[4890]: I1125 15:30:20.372150 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ddmtt"] Nov 25 15:30:20 crc kubenswrapper[4890]: I1125 15:30:20.381632 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ddmtt"] Nov 25 15:30:20 crc kubenswrapper[4890]: I1125 15:30:20.389491 4890 scope.go:117] "RemoveContainer" containerID="27512991ab22e87d4234e5b73728a5adba06ff97504b8aea926363411d19c00f" Nov 25 15:30:20 crc kubenswrapper[4890]: I1125 15:30:20.415353 4890 scope.go:117] "RemoveContainer" containerID="bb0dcecc09d90d56e793929595f4f4c4631c9363c8d877b4b159f0abbbf8d5fa" Nov 25 15:30:20 crc kubenswrapper[4890]: I1125 15:30:20.461225 4890 scope.go:117] "RemoveContainer" containerID="5cb9b523a30d43bc28da23a27df37463d35f8bac93e2c391c98b08c3b39391b6" Nov 25 15:30:20 crc kubenswrapper[4890]: I1125 15:30:20.505406 4890 scope.go:117] "RemoveContainer" containerID="50d3fcd65f895ac4b64138cb2a54b5f4ab83b31085f788a228b32f5683880aff" Nov 25 15:30:22 crc kubenswrapper[4890]: I1125 15:30:22.187111 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c070ef9-daba-4ae9-ac05-dff449baf01a" path="/var/lib/kubelet/pods/7c070ef9-daba-4ae9-ac05-dff449baf01a/volumes" Nov 25 15:30:22 crc kubenswrapper[4890]: I1125 15:30:22.190475 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9fbb88d-f5f9-41c6-9719-15be794dfb6c" path="/var/lib/kubelet/pods/b9fbb88d-f5f9-41c6-9719-15be794dfb6c/volumes" Nov 25 15:30:22 crc kubenswrapper[4890]: I1125 15:30:22.554945 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-f24bd" podUID="0f05c16a-1239-42a7-9910-ca03ef42b0f7" containerName="registry-server" probeResult="failure" output=< Nov 25 15:30:22 crc kubenswrapper[4890]: timeout: failed to connect service ":50051" within 1s Nov 25 15:30:22 crc kubenswrapper[4890]: > Nov 25 15:30:24 crc kubenswrapper[4890]: I1125 15:30:24.173539 4890 scope.go:117] "RemoveContainer" containerID="9467aa8edc4d8bfe461dc02826750091f343018e37e7f269d305c77d21e832da" Nov 25 15:30:24 crc kubenswrapper[4890]: E1125 15:30:24.174418 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:30:28 crc kubenswrapper[4890]: I1125 15:30:28.228572 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-48z2b" Nov 25 15:30:28 crc kubenswrapper[4890]: I1125 15:30:28.284916 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-48z2b" Nov 25 15:30:32 crc kubenswrapper[4890]: I1125 15:30:32.548558 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-f24bd" podUID="0f05c16a-1239-42a7-9910-ca03ef42b0f7" containerName="registry-server" probeResult="failure" output=< Nov 25 15:30:32 crc kubenswrapper[4890]: timeout: failed to connect service ":50051" within 1s Nov 25 15:30:32 crc kubenswrapper[4890]: > Nov 25 15:30:37 crc kubenswrapper[4890]: I1125 15:30:37.376293 4890 scope.go:117] "RemoveContainer" containerID="14a4299c5a0c5640a64d516b2a7b493115f14487442b9308ee75e07c09e6b8ab" Nov 25 15:30:38 crc kubenswrapper[4890]: I1125 15:30:38.173298 4890 scope.go:117] "RemoveContainer" containerID="9467aa8edc4d8bfe461dc02826750091f343018e37e7f269d305c77d21e832da" Nov 25 15:30:38 crc kubenswrapper[4890]: E1125 15:30:38.173617 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:30:40 crc kubenswrapper[4890]: I1125 15:30:40.060503 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-2dtqm"] Nov 25 15:30:40 crc kubenswrapper[4890]: I1125 15:30:40.074099 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-2dtqm"] Nov 25 15:30:40 crc kubenswrapper[4890]: I1125 15:30:40.185997 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7354c28-0b1f-4dda-9b1b-42dcd2657a7a" path="/var/lib/kubelet/pods/c7354c28-0b1f-4dda-9b1b-42dcd2657a7a/volumes" Nov 25 15:30:42 crc kubenswrapper[4890]: I1125 15:30:42.610015 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-f24bd" podUID="0f05c16a-1239-42a7-9910-ca03ef42b0f7" containerName="registry-server" probeResult="failure" output=< Nov 25 15:30:42 crc kubenswrapper[4890]: timeout: failed to connect service ":50051" within 1s Nov 25 15:30:42 crc kubenswrapper[4890]: > Nov 25 15:30:51 crc kubenswrapper[4890]: I1125 15:30:51.173293 4890 scope.go:117] "RemoveContainer" containerID="9467aa8edc4d8bfe461dc02826750091f343018e37e7f269d305c77d21e832da" Nov 25 15:30:51 crc kubenswrapper[4890]: E1125 15:30:51.174296 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:30:52 crc kubenswrapper[4890]: I1125 15:30:52.544768 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-f24bd" podUID="0f05c16a-1239-42a7-9910-ca03ef42b0f7" containerName="registry-server" probeResult="failure" output=< Nov 25 15:30:52 crc kubenswrapper[4890]: timeout: failed to connect service ":50051" within 1s Nov 25 15:30:52 crc kubenswrapper[4890]: > Nov 25 15:31:02 crc kubenswrapper[4890]: I1125 15:31:02.560102 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-f24bd" podUID="0f05c16a-1239-42a7-9910-ca03ef42b0f7" containerName="registry-server" probeResult="failure" output=< Nov 25 15:31:02 crc kubenswrapper[4890]: timeout: failed to connect service ":50051" within 1s Nov 25 15:31:02 crc kubenswrapper[4890]: > Nov 25 15:31:02 crc kubenswrapper[4890]: I1125 15:31:02.560820 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-f24bd" Nov 25 15:31:02 crc kubenswrapper[4890]: I1125 15:31:02.561882 4890 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="registry-server" containerStatusID={"Type":"cri-o","ID":"e4d7e3c3f0c96c99da055898bc7727760b326d9a7521ca55b599f4f9f9321a47"} pod="openshift-marketplace/redhat-operators-f24bd" containerMessage="Container registry-server failed startup probe, will be restarted" Nov 25 15:31:02 crc kubenswrapper[4890]: I1125 15:31:02.561929 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-f24bd" podUID="0f05c16a-1239-42a7-9910-ca03ef42b0f7" containerName="registry-server" containerID="cri-o://e4d7e3c3f0c96c99da055898bc7727760b326d9a7521ca55b599f4f9f9321a47" gracePeriod=30 Nov 25 15:31:05 crc kubenswrapper[4890]: I1125 15:31:05.173896 4890 scope.go:117] "RemoveContainer" containerID="9467aa8edc4d8bfe461dc02826750091f343018e37e7f269d305c77d21e832da" Nov 25 15:31:05 crc kubenswrapper[4890]: E1125 15:31:05.174452 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:31:07 crc kubenswrapper[4890]: I1125 15:31:07.914770 4890 generic.go:334] "Generic (PLEG): container finished" podID="0f05c16a-1239-42a7-9910-ca03ef42b0f7" containerID="e4d7e3c3f0c96c99da055898bc7727760b326d9a7521ca55b599f4f9f9321a47" exitCode=0 Nov 25 15:31:07 crc kubenswrapper[4890]: I1125 15:31:07.914861 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f24bd" event={"ID":"0f05c16a-1239-42a7-9910-ca03ef42b0f7","Type":"ContainerDied","Data":"e4d7e3c3f0c96c99da055898bc7727760b326d9a7521ca55b599f4f9f9321a47"} Nov 25 15:31:09 crc kubenswrapper[4890]: I1125 15:31:09.045889 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-j7w9k"] Nov 25 15:31:09 crc kubenswrapper[4890]: I1125 15:31:09.055267 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-j7w9k"] Nov 25 15:31:09 crc kubenswrapper[4890]: I1125 15:31:09.952869 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f24bd" event={"ID":"0f05c16a-1239-42a7-9910-ca03ef42b0f7","Type":"ContainerStarted","Data":"66ba6427376d36dfc038095fda6afb925dd2f34d3a37925f2c39cefe166fd692"} Nov 25 15:31:10 crc kubenswrapper[4890]: I1125 15:31:10.184604 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="997a078a-52bd-4eb2-8fde-f80d05d9b18c" path="/var/lib/kubelet/pods/997a078a-52bd-4eb2-8fde-f80d05d9b18c/volumes" Nov 25 15:31:11 crc kubenswrapper[4890]: I1125 15:31:11.501615 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-f24bd" Nov 25 15:31:11 crc kubenswrapper[4890]: I1125 15:31:11.501885 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-f24bd" Nov 25 15:31:12 crc kubenswrapper[4890]: I1125 15:31:12.547864 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-f24bd" podUID="0f05c16a-1239-42a7-9910-ca03ef42b0f7" containerName="registry-server" probeResult="failure" output=< Nov 25 15:31:12 crc kubenswrapper[4890]: timeout: failed to connect service ":50051" within 1s Nov 25 15:31:12 crc kubenswrapper[4890]: > Nov 25 15:31:16 crc kubenswrapper[4890]: I1125 15:31:16.173318 4890 scope.go:117] "RemoveContainer" containerID="9467aa8edc4d8bfe461dc02826750091f343018e37e7f269d305c77d21e832da" Nov 25 15:31:16 crc kubenswrapper[4890]: E1125 15:31:16.174128 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:31:18 crc kubenswrapper[4890]: I1125 15:31:18.045519 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-f8ctf"] Nov 25 15:31:18 crc kubenswrapper[4890]: I1125 15:31:18.071620 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-f8ctf"] Nov 25 15:31:18 crc kubenswrapper[4890]: I1125 15:31:18.193410 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fdcacd06-b857-46cf-8516-345e271cf882" path="/var/lib/kubelet/pods/fdcacd06-b857-46cf-8516-345e271cf882/volumes" Nov 25 15:31:21 crc kubenswrapper[4890]: I1125 15:31:21.553580 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-f24bd" Nov 25 15:31:21 crc kubenswrapper[4890]: I1125 15:31:21.622089 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-f24bd" Nov 25 15:31:21 crc kubenswrapper[4890]: I1125 15:31:21.797528 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-48z2b"] Nov 25 15:31:21 crc kubenswrapper[4890]: I1125 15:31:21.798200 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-48z2b" podUID="eacc1f05-701e-49af-b2d7-9c049f6cddb9" containerName="registry-server" containerID="cri-o://f0e93c436975a4beaa31027e8591a496e2b0dbd7c8eab3144c79f7faa3188393" gracePeriod=2 Nov 25 15:31:21 crc kubenswrapper[4890]: I1125 15:31:21.997908 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4fdm7"] Nov 25 15:31:21 crc kubenswrapper[4890]: I1125 15:31:21.998134 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4fdm7" podUID="a0a34ff8-8257-4f49-90d2-90576cbc40b7" containerName="registry-server" containerID="cri-o://d55b3ccd8fde323cdc0d0587c0d3ffccd8748e2255ed40d3230a9a5404369928" gracePeriod=2 Nov 25 15:31:22 crc kubenswrapper[4890]: I1125 15:31:22.130727 4890 generic.go:334] "Generic (PLEG): container finished" podID="eacc1f05-701e-49af-b2d7-9c049f6cddb9" containerID="f0e93c436975a4beaa31027e8591a496e2b0dbd7c8eab3144c79f7faa3188393" exitCode=0 Nov 25 15:31:22 crc kubenswrapper[4890]: I1125 15:31:22.130862 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-48z2b" event={"ID":"eacc1f05-701e-49af-b2d7-9c049f6cddb9","Type":"ContainerDied","Data":"f0e93c436975a4beaa31027e8591a496e2b0dbd7c8eab3144c79f7faa3188393"} Nov 25 15:31:22 crc kubenswrapper[4890]: I1125 15:31:22.205679 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mtfgl"] Nov 25 15:31:22 crc kubenswrapper[4890]: I1125 15:31:22.206133 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-mtfgl" podUID="f7235b7e-d6fe-45a2-93fd-66b0a77b4b76" containerName="registry-server" containerID="cri-o://d78cdc34e1dfac4d6751e111e5e2a02f35cd675570357a84f0a9b5d9892b9881" gracePeriod=2 Nov 25 15:31:22 crc kubenswrapper[4890]: I1125 15:31:22.399194 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f24bd"] Nov 25 15:31:22 crc kubenswrapper[4890]: I1125 15:31:22.898419 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-48z2b" Nov 25 15:31:22 crc kubenswrapper[4890]: I1125 15:31:22.961030 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eacc1f05-701e-49af-b2d7-9c049f6cddb9-catalog-content\") pod \"eacc1f05-701e-49af-b2d7-9c049f6cddb9\" (UID: \"eacc1f05-701e-49af-b2d7-9c049f6cddb9\") " Nov 25 15:31:22 crc kubenswrapper[4890]: I1125 15:31:22.961144 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eacc1f05-701e-49af-b2d7-9c049f6cddb9-utilities\") pod \"eacc1f05-701e-49af-b2d7-9c049f6cddb9\" (UID: \"eacc1f05-701e-49af-b2d7-9c049f6cddb9\") " Nov 25 15:31:22 crc kubenswrapper[4890]: I1125 15:31:22.961254 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2gvv6\" (UniqueName: \"kubernetes.io/projected/eacc1f05-701e-49af-b2d7-9c049f6cddb9-kube-api-access-2gvv6\") pod \"eacc1f05-701e-49af-b2d7-9c049f6cddb9\" (UID: \"eacc1f05-701e-49af-b2d7-9c049f6cddb9\") " Nov 25 15:31:22 crc kubenswrapper[4890]: I1125 15:31:22.963021 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eacc1f05-701e-49af-b2d7-9c049f6cddb9-utilities" (OuterVolumeSpecName: "utilities") pod "eacc1f05-701e-49af-b2d7-9c049f6cddb9" (UID: "eacc1f05-701e-49af-b2d7-9c049f6cddb9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:31:22 crc kubenswrapper[4890]: I1125 15:31:22.973024 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eacc1f05-701e-49af-b2d7-9c049f6cddb9-kube-api-access-2gvv6" (OuterVolumeSpecName: "kube-api-access-2gvv6") pod "eacc1f05-701e-49af-b2d7-9c049f6cddb9" (UID: "eacc1f05-701e-49af-b2d7-9c049f6cddb9"). InnerVolumeSpecName "kube-api-access-2gvv6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:31:23 crc kubenswrapper[4890]: I1125 15:31:23.064434 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eacc1f05-701e-49af-b2d7-9c049f6cddb9-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 15:31:23 crc kubenswrapper[4890]: I1125 15:31:23.064477 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2gvv6\" (UniqueName: \"kubernetes.io/projected/eacc1f05-701e-49af-b2d7-9c049f6cddb9-kube-api-access-2gvv6\") on node \"crc\" DevicePath \"\"" Nov 25 15:31:23 crc kubenswrapper[4890]: I1125 15:31:23.094889 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eacc1f05-701e-49af-b2d7-9c049f6cddb9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eacc1f05-701e-49af-b2d7-9c049f6cddb9" (UID: "eacc1f05-701e-49af-b2d7-9c049f6cddb9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:31:23 crc kubenswrapper[4890]: I1125 15:31:23.126833 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4fdm7" Nov 25 15:31:23 crc kubenswrapper[4890]: I1125 15:31:23.167011 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vbzmx\" (UniqueName: \"kubernetes.io/projected/a0a34ff8-8257-4f49-90d2-90576cbc40b7-kube-api-access-vbzmx\") pod \"a0a34ff8-8257-4f49-90d2-90576cbc40b7\" (UID: \"a0a34ff8-8257-4f49-90d2-90576cbc40b7\") " Nov 25 15:31:23 crc kubenswrapper[4890]: I1125 15:31:23.167060 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0a34ff8-8257-4f49-90d2-90576cbc40b7-catalog-content\") pod \"a0a34ff8-8257-4f49-90d2-90576cbc40b7\" (UID: \"a0a34ff8-8257-4f49-90d2-90576cbc40b7\") " Nov 25 15:31:23 crc kubenswrapper[4890]: I1125 15:31:23.167350 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0a34ff8-8257-4f49-90d2-90576cbc40b7-utilities\") pod \"a0a34ff8-8257-4f49-90d2-90576cbc40b7\" (UID: \"a0a34ff8-8257-4f49-90d2-90576cbc40b7\") " Nov 25 15:31:23 crc kubenswrapper[4890]: I1125 15:31:23.168192 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0a34ff8-8257-4f49-90d2-90576cbc40b7-utilities" (OuterVolumeSpecName: "utilities") pod "a0a34ff8-8257-4f49-90d2-90576cbc40b7" (UID: "a0a34ff8-8257-4f49-90d2-90576cbc40b7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:31:23 crc kubenswrapper[4890]: I1125 15:31:23.169642 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0a34ff8-8257-4f49-90d2-90576cbc40b7-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 15:31:23 crc kubenswrapper[4890]: I1125 15:31:23.169691 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eacc1f05-701e-49af-b2d7-9c049f6cddb9-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 15:31:23 crc kubenswrapper[4890]: I1125 15:31:23.172355 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0a34ff8-8257-4f49-90d2-90576cbc40b7-kube-api-access-vbzmx" (OuterVolumeSpecName: "kube-api-access-vbzmx") pod "a0a34ff8-8257-4f49-90d2-90576cbc40b7" (UID: "a0a34ff8-8257-4f49-90d2-90576cbc40b7"). InnerVolumeSpecName "kube-api-access-vbzmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:31:23 crc kubenswrapper[4890]: I1125 15:31:23.179096 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-48z2b" Nov 25 15:31:23 crc kubenswrapper[4890]: I1125 15:31:23.184603 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-48z2b" event={"ID":"eacc1f05-701e-49af-b2d7-9c049f6cddb9","Type":"ContainerDied","Data":"a044d2b6a7d48ac7eced8baea8581590025de47ee6b4fdb424a32c6bde3fa2d5"} Nov 25 15:31:23 crc kubenswrapper[4890]: I1125 15:31:23.185137 4890 scope.go:117] "RemoveContainer" containerID="f0e93c436975a4beaa31027e8591a496e2b0dbd7c8eab3144c79f7faa3188393" Nov 25 15:31:23 crc kubenswrapper[4890]: I1125 15:31:23.187682 4890 generic.go:334] "Generic (PLEG): container finished" podID="a0a34ff8-8257-4f49-90d2-90576cbc40b7" containerID="d55b3ccd8fde323cdc0d0587c0d3ffccd8748e2255ed40d3230a9a5404369928" exitCode=0 Nov 25 15:31:23 crc kubenswrapper[4890]: I1125 15:31:23.187785 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4fdm7" event={"ID":"a0a34ff8-8257-4f49-90d2-90576cbc40b7","Type":"ContainerDied","Data":"d55b3ccd8fde323cdc0d0587c0d3ffccd8748e2255ed40d3230a9a5404369928"} Nov 25 15:31:23 crc kubenswrapper[4890]: I1125 15:31:23.187816 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4fdm7" event={"ID":"a0a34ff8-8257-4f49-90d2-90576cbc40b7","Type":"ContainerDied","Data":"5cddf18bc8e2fcde0308a92cc51919688ad6bd00799fe1044ac649392b8446ac"} Nov 25 15:31:23 crc kubenswrapper[4890]: I1125 15:31:23.188171 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4fdm7" Nov 25 15:31:23 crc kubenswrapper[4890]: I1125 15:31:23.193749 4890 generic.go:334] "Generic (PLEG): container finished" podID="f7235b7e-d6fe-45a2-93fd-66b0a77b4b76" containerID="d78cdc34e1dfac4d6751e111e5e2a02f35cd675570357a84f0a9b5d9892b9881" exitCode=0 Nov 25 15:31:23 crc kubenswrapper[4890]: I1125 15:31:23.194072 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mtfgl" event={"ID":"f7235b7e-d6fe-45a2-93fd-66b0a77b4b76","Type":"ContainerDied","Data":"d78cdc34e1dfac4d6751e111e5e2a02f35cd675570357a84f0a9b5d9892b9881"} Nov 25 15:31:23 crc kubenswrapper[4890]: I1125 15:31:23.194318 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-f24bd" podUID="0f05c16a-1239-42a7-9910-ca03ef42b0f7" containerName="registry-server" containerID="cri-o://66ba6427376d36dfc038095fda6afb925dd2f34d3a37925f2c39cefe166fd692" gracePeriod=2 Nov 25 15:31:23 crc kubenswrapper[4890]: I1125 15:31:23.274859 4890 scope.go:117] "RemoveContainer" containerID="bb4a6481afe58d7d9473c83486d0675ebc198deaa106003b6fe70e2bde92c9e3" Nov 25 15:31:23 crc kubenswrapper[4890]: I1125 15:31:23.278692 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vbzmx\" (UniqueName: \"kubernetes.io/projected/a0a34ff8-8257-4f49-90d2-90576cbc40b7-kube-api-access-vbzmx\") on node \"crc\" DevicePath \"\"" Nov 25 15:31:23 crc kubenswrapper[4890]: I1125 15:31:23.279690 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-48z2b"] Nov 25 15:31:23 crc kubenswrapper[4890]: I1125 15:31:23.297520 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0a34ff8-8257-4f49-90d2-90576cbc40b7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a0a34ff8-8257-4f49-90d2-90576cbc40b7" (UID: "a0a34ff8-8257-4f49-90d2-90576cbc40b7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:31:23 crc kubenswrapper[4890]: I1125 15:31:23.299849 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-48z2b"] Nov 25 15:31:23 crc kubenswrapper[4890]: I1125 15:31:23.382241 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0a34ff8-8257-4f49-90d2-90576cbc40b7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 15:31:23 crc kubenswrapper[4890]: I1125 15:31:23.398292 4890 scope.go:117] "RemoveContainer" containerID="ae8f77220f5a568aeb2874fdf50f8002339e492b5a510906dd1e42b2d91d63a6" Nov 25 15:31:23 crc kubenswrapper[4890]: I1125 15:31:23.488108 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mtfgl" Nov 25 15:31:23 crc kubenswrapper[4890]: I1125 15:31:23.551760 4890 scope.go:117] "RemoveContainer" containerID="d55b3ccd8fde323cdc0d0587c0d3ffccd8748e2255ed40d3230a9a5404369928" Nov 25 15:31:23 crc kubenswrapper[4890]: I1125 15:31:23.562747 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4fdm7"] Nov 25 15:31:23 crc kubenswrapper[4890]: I1125 15:31:23.574834 4890 scope.go:117] "RemoveContainer" containerID="6dcaca4001e4147931cd54f2beb3005bc92b25e287ced515f60a19e8b2585d90" Nov 25 15:31:23 crc kubenswrapper[4890]: I1125 15:31:23.575099 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4fdm7"] Nov 25 15:31:23 crc kubenswrapper[4890]: I1125 15:31:23.586520 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7235b7e-d6fe-45a2-93fd-66b0a77b4b76-utilities\") pod \"f7235b7e-d6fe-45a2-93fd-66b0a77b4b76\" (UID: \"f7235b7e-d6fe-45a2-93fd-66b0a77b4b76\") " Nov 25 15:31:23 crc kubenswrapper[4890]: I1125 15:31:23.586667 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7235b7e-d6fe-45a2-93fd-66b0a77b4b76-catalog-content\") pod \"f7235b7e-d6fe-45a2-93fd-66b0a77b4b76\" (UID: \"f7235b7e-d6fe-45a2-93fd-66b0a77b4b76\") " Nov 25 15:31:23 crc kubenswrapper[4890]: I1125 15:31:23.586711 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qk5x8\" (UniqueName: \"kubernetes.io/projected/f7235b7e-d6fe-45a2-93fd-66b0a77b4b76-kube-api-access-qk5x8\") pod \"f7235b7e-d6fe-45a2-93fd-66b0a77b4b76\" (UID: \"f7235b7e-d6fe-45a2-93fd-66b0a77b4b76\") " Nov 25 15:31:23 crc kubenswrapper[4890]: I1125 15:31:23.587747 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7235b7e-d6fe-45a2-93fd-66b0a77b4b76-utilities" (OuterVolumeSpecName: "utilities") pod "f7235b7e-d6fe-45a2-93fd-66b0a77b4b76" (UID: "f7235b7e-d6fe-45a2-93fd-66b0a77b4b76"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:31:23 crc kubenswrapper[4890]: I1125 15:31:23.591562 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7235b7e-d6fe-45a2-93fd-66b0a77b4b76-kube-api-access-qk5x8" (OuterVolumeSpecName: "kube-api-access-qk5x8") pod "f7235b7e-d6fe-45a2-93fd-66b0a77b4b76" (UID: "f7235b7e-d6fe-45a2-93fd-66b0a77b4b76"). InnerVolumeSpecName "kube-api-access-qk5x8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:31:23 crc kubenswrapper[4890]: I1125 15:31:23.634580 4890 scope.go:117] "RemoveContainer" containerID="33598a45238600ec7547cd57326233b2a1d69ba08ec0953da513c856306a0bc8" Nov 25 15:31:23 crc kubenswrapper[4890]: I1125 15:31:23.678470 4890 scope.go:117] "RemoveContainer" containerID="d55b3ccd8fde323cdc0d0587c0d3ffccd8748e2255ed40d3230a9a5404369928" Nov 25 15:31:23 crc kubenswrapper[4890]: E1125 15:31:23.683689 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d55b3ccd8fde323cdc0d0587c0d3ffccd8748e2255ed40d3230a9a5404369928\": container with ID starting with d55b3ccd8fde323cdc0d0587c0d3ffccd8748e2255ed40d3230a9a5404369928 not found: ID does not exist" containerID="d55b3ccd8fde323cdc0d0587c0d3ffccd8748e2255ed40d3230a9a5404369928" Nov 25 15:31:23 crc kubenswrapper[4890]: I1125 15:31:23.683746 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d55b3ccd8fde323cdc0d0587c0d3ffccd8748e2255ed40d3230a9a5404369928"} err="failed to get container status \"d55b3ccd8fde323cdc0d0587c0d3ffccd8748e2255ed40d3230a9a5404369928\": rpc error: code = NotFound desc = could not find container \"d55b3ccd8fde323cdc0d0587c0d3ffccd8748e2255ed40d3230a9a5404369928\": container with ID starting with d55b3ccd8fde323cdc0d0587c0d3ffccd8748e2255ed40d3230a9a5404369928 not found: ID does not exist" Nov 25 15:31:23 crc kubenswrapper[4890]: I1125 15:31:23.683788 4890 scope.go:117] "RemoveContainer" containerID="6dcaca4001e4147931cd54f2beb3005bc92b25e287ced515f60a19e8b2585d90" Nov 25 15:31:23 crc kubenswrapper[4890]: E1125 15:31:23.684475 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6dcaca4001e4147931cd54f2beb3005bc92b25e287ced515f60a19e8b2585d90\": container with ID starting with 6dcaca4001e4147931cd54f2beb3005bc92b25e287ced515f60a19e8b2585d90 not found: ID does not exist" containerID="6dcaca4001e4147931cd54f2beb3005bc92b25e287ced515f60a19e8b2585d90" Nov 25 15:31:23 crc kubenswrapper[4890]: I1125 15:31:23.684544 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6dcaca4001e4147931cd54f2beb3005bc92b25e287ced515f60a19e8b2585d90"} err="failed to get container status \"6dcaca4001e4147931cd54f2beb3005bc92b25e287ced515f60a19e8b2585d90\": rpc error: code = NotFound desc = could not find container \"6dcaca4001e4147931cd54f2beb3005bc92b25e287ced515f60a19e8b2585d90\": container with ID starting with 6dcaca4001e4147931cd54f2beb3005bc92b25e287ced515f60a19e8b2585d90 not found: ID does not exist" Nov 25 15:31:23 crc kubenswrapper[4890]: I1125 15:31:23.684597 4890 scope.go:117] "RemoveContainer" containerID="33598a45238600ec7547cd57326233b2a1d69ba08ec0953da513c856306a0bc8" Nov 25 15:31:23 crc kubenswrapper[4890]: E1125 15:31:23.685147 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33598a45238600ec7547cd57326233b2a1d69ba08ec0953da513c856306a0bc8\": container with ID starting with 33598a45238600ec7547cd57326233b2a1d69ba08ec0953da513c856306a0bc8 not found: ID does not exist" containerID="33598a45238600ec7547cd57326233b2a1d69ba08ec0953da513c856306a0bc8" Nov 25 15:31:23 crc kubenswrapper[4890]: I1125 15:31:23.685225 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33598a45238600ec7547cd57326233b2a1d69ba08ec0953da513c856306a0bc8"} err="failed to get container status \"33598a45238600ec7547cd57326233b2a1d69ba08ec0953da513c856306a0bc8\": rpc error: code = NotFound desc = could not find container \"33598a45238600ec7547cd57326233b2a1d69ba08ec0953da513c856306a0bc8\": container with ID starting with 33598a45238600ec7547cd57326233b2a1d69ba08ec0953da513c856306a0bc8 not found: ID does not exist" Nov 25 15:31:23 crc kubenswrapper[4890]: I1125 15:31:23.690880 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7235b7e-d6fe-45a2-93fd-66b0a77b4b76-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 15:31:23 crc kubenswrapper[4890]: I1125 15:31:23.690929 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qk5x8\" (UniqueName: \"kubernetes.io/projected/f7235b7e-d6fe-45a2-93fd-66b0a77b4b76-kube-api-access-qk5x8\") on node \"crc\" DevicePath \"\"" Nov 25 15:31:23 crc kubenswrapper[4890]: I1125 15:31:23.707485 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7235b7e-d6fe-45a2-93fd-66b0a77b4b76-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f7235b7e-d6fe-45a2-93fd-66b0a77b4b76" (UID: "f7235b7e-d6fe-45a2-93fd-66b0a77b4b76"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:31:23 crc kubenswrapper[4890]: I1125 15:31:23.794151 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7235b7e-d6fe-45a2-93fd-66b0a77b4b76-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 15:31:24 crc kubenswrapper[4890]: I1125 15:31:24.187742 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0a34ff8-8257-4f49-90d2-90576cbc40b7" path="/var/lib/kubelet/pods/a0a34ff8-8257-4f49-90d2-90576cbc40b7/volumes" Nov 25 15:31:24 crc kubenswrapper[4890]: I1125 15:31:24.188864 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eacc1f05-701e-49af-b2d7-9c049f6cddb9" path="/var/lib/kubelet/pods/eacc1f05-701e-49af-b2d7-9c049f6cddb9/volumes" Nov 25 15:31:24 crc kubenswrapper[4890]: I1125 15:31:24.214771 4890 generic.go:334] "Generic (PLEG): container finished" podID="0f05c16a-1239-42a7-9910-ca03ef42b0f7" containerID="66ba6427376d36dfc038095fda6afb925dd2f34d3a37925f2c39cefe166fd692" exitCode=0 Nov 25 15:31:24 crc kubenswrapper[4890]: I1125 15:31:24.214859 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f24bd" event={"ID":"0f05c16a-1239-42a7-9910-ca03ef42b0f7","Type":"ContainerDied","Data":"66ba6427376d36dfc038095fda6afb925dd2f34d3a37925f2c39cefe166fd692"} Nov 25 15:31:24 crc kubenswrapper[4890]: I1125 15:31:24.214908 4890 scope.go:117] "RemoveContainer" containerID="e4d7e3c3f0c96c99da055898bc7727760b326d9a7521ca55b599f4f9f9321a47" Nov 25 15:31:24 crc kubenswrapper[4890]: I1125 15:31:24.223298 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mtfgl" event={"ID":"f7235b7e-d6fe-45a2-93fd-66b0a77b4b76","Type":"ContainerDied","Data":"2cbf9bfcbac76d87873600247a7a229961b2ceb24529a93268d7f3093f8a659c"} Nov 25 15:31:24 crc kubenswrapper[4890]: I1125 15:31:24.223599 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mtfgl" Nov 25 15:31:24 crc kubenswrapper[4890]: I1125 15:31:24.295433 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mtfgl"] Nov 25 15:31:24 crc kubenswrapper[4890]: I1125 15:31:24.304133 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-mtfgl"] Nov 25 15:31:24 crc kubenswrapper[4890]: I1125 15:31:24.437194 4890 scope.go:117] "RemoveContainer" containerID="d78cdc34e1dfac4d6751e111e5e2a02f35cd675570357a84f0a9b5d9892b9881" Nov 25 15:31:24 crc kubenswrapper[4890]: I1125 15:31:24.474119 4890 scope.go:117] "RemoveContainer" containerID="a96c1915aa5d9e95c5c7db8e3f90567de5b88e0a43de65bbeaea988efa55a8ff" Nov 25 15:31:24 crc kubenswrapper[4890]: I1125 15:31:24.612226 4890 scope.go:117] "RemoveContainer" containerID="f19ba65caf7a2838ce342808f2d9139f050a92071798091cb624ad53d054f55b" Nov 25 15:31:24 crc kubenswrapper[4890]: I1125 15:31:24.639344 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f24bd" Nov 25 15:31:24 crc kubenswrapper[4890]: I1125 15:31:24.716951 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f05c16a-1239-42a7-9910-ca03ef42b0f7-utilities\") pod \"0f05c16a-1239-42a7-9910-ca03ef42b0f7\" (UID: \"0f05c16a-1239-42a7-9910-ca03ef42b0f7\") " Nov 25 15:31:24 crc kubenswrapper[4890]: I1125 15:31:24.717243 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f05c16a-1239-42a7-9910-ca03ef42b0f7-catalog-content\") pod \"0f05c16a-1239-42a7-9910-ca03ef42b0f7\" (UID: \"0f05c16a-1239-42a7-9910-ca03ef42b0f7\") " Nov 25 15:31:24 crc kubenswrapper[4890]: I1125 15:31:24.717660 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-45wdj\" (UniqueName: \"kubernetes.io/projected/0f05c16a-1239-42a7-9910-ca03ef42b0f7-kube-api-access-45wdj\") pod \"0f05c16a-1239-42a7-9910-ca03ef42b0f7\" (UID: \"0f05c16a-1239-42a7-9910-ca03ef42b0f7\") " Nov 25 15:31:24 crc kubenswrapper[4890]: I1125 15:31:24.718746 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f05c16a-1239-42a7-9910-ca03ef42b0f7-utilities" (OuterVolumeSpecName: "utilities") pod "0f05c16a-1239-42a7-9910-ca03ef42b0f7" (UID: "0f05c16a-1239-42a7-9910-ca03ef42b0f7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:31:24 crc kubenswrapper[4890]: I1125 15:31:24.719333 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f05c16a-1239-42a7-9910-ca03ef42b0f7-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 15:31:24 crc kubenswrapper[4890]: I1125 15:31:24.756760 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f05c16a-1239-42a7-9910-ca03ef42b0f7-kube-api-access-45wdj" (OuterVolumeSpecName: "kube-api-access-45wdj") pod "0f05c16a-1239-42a7-9910-ca03ef42b0f7" (UID: "0f05c16a-1239-42a7-9910-ca03ef42b0f7"). InnerVolumeSpecName "kube-api-access-45wdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:31:24 crc kubenswrapper[4890]: I1125 15:31:24.820969 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-45wdj\" (UniqueName: \"kubernetes.io/projected/0f05c16a-1239-42a7-9910-ca03ef42b0f7-kube-api-access-45wdj\") on node \"crc\" DevicePath \"\"" Nov 25 15:31:24 crc kubenswrapper[4890]: I1125 15:31:24.838475 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f05c16a-1239-42a7-9910-ca03ef42b0f7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0f05c16a-1239-42a7-9910-ca03ef42b0f7" (UID: "0f05c16a-1239-42a7-9910-ca03ef42b0f7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:31:24 crc kubenswrapper[4890]: I1125 15:31:24.923716 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f05c16a-1239-42a7-9910-ca03ef42b0f7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 15:31:25 crc kubenswrapper[4890]: I1125 15:31:25.248484 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f24bd" event={"ID":"0f05c16a-1239-42a7-9910-ca03ef42b0f7","Type":"ContainerDied","Data":"f087c6a953c0f73a4c176ffff375a048156a59224760eb2c0e7ab2046ec36214"} Nov 25 15:31:25 crc kubenswrapper[4890]: I1125 15:31:25.249337 4890 scope.go:117] "RemoveContainer" containerID="66ba6427376d36dfc038095fda6afb925dd2f34d3a37925f2c39cefe166fd692" Nov 25 15:31:25 crc kubenswrapper[4890]: I1125 15:31:25.248605 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f24bd" Nov 25 15:31:25 crc kubenswrapper[4890]: I1125 15:31:25.277701 4890 scope.go:117] "RemoveContainer" containerID="f7bca35abf69e8fefa3964fcbac0d5a3d7d365f424da842801a0424a4412311c" Nov 25 15:31:25 crc kubenswrapper[4890]: I1125 15:31:25.322043 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f24bd"] Nov 25 15:31:25 crc kubenswrapper[4890]: I1125 15:31:25.333621 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-f24bd"] Nov 25 15:31:25 crc kubenswrapper[4890]: I1125 15:31:25.367754 4890 scope.go:117] "RemoveContainer" containerID="caf1c4a1acd3f56a5ecc9d123caeebca960888f6685246e0bf27ec17d5482207" Nov 25 15:31:26 crc kubenswrapper[4890]: I1125 15:31:26.191311 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f05c16a-1239-42a7-9910-ca03ef42b0f7" path="/var/lib/kubelet/pods/0f05c16a-1239-42a7-9910-ca03ef42b0f7/volumes" Nov 25 15:31:26 crc kubenswrapper[4890]: I1125 15:31:26.192485 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7235b7e-d6fe-45a2-93fd-66b0a77b4b76" path="/var/lib/kubelet/pods/f7235b7e-d6fe-45a2-93fd-66b0a77b4b76/volumes" Nov 25 15:31:31 crc kubenswrapper[4890]: I1125 15:31:31.173580 4890 scope.go:117] "RemoveContainer" containerID="9467aa8edc4d8bfe461dc02826750091f343018e37e7f269d305c77d21e832da" Nov 25 15:31:31 crc kubenswrapper[4890]: E1125 15:31:31.175477 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:31:35 crc kubenswrapper[4890]: I1125 15:31:35.067727 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-wlvxc"] Nov 25 15:31:35 crc kubenswrapper[4890]: I1125 15:31:35.080553 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-c5b26"] Nov 25 15:31:35 crc kubenswrapper[4890]: I1125 15:31:35.096700 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-wlvxc"] Nov 25 15:31:35 crc kubenswrapper[4890]: I1125 15:31:35.106567 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-c5b26"] Nov 25 15:31:36 crc kubenswrapper[4890]: I1125 15:31:36.189884 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="449c7dd2-e3c8-48c2-8770-c1d6a77fce7b" path="/var/lib/kubelet/pods/449c7dd2-e3c8-48c2-8770-c1d6a77fce7b/volumes" Nov 25 15:31:36 crc kubenswrapper[4890]: I1125 15:31:36.193619 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dfc6fde5-9e4c-4224-8200-0f230e127f7e" path="/var/lib/kubelet/pods/dfc6fde5-9e4c-4224-8200-0f230e127f7e/volumes" Nov 25 15:31:37 crc kubenswrapper[4890]: I1125 15:31:37.698250 4890 scope.go:117] "RemoveContainer" containerID="30fb995c1452e7fb88486c70e38cdf87240b7d5488b0d1aaa70d5a725f84c6bf" Nov 25 15:31:37 crc kubenswrapper[4890]: I1125 15:31:37.748030 4890 scope.go:117] "RemoveContainer" containerID="ee352510e5cc498c734f0e0cdd27fdfe9c81cc7a997ed487be227c10e1a3569c" Nov 25 15:31:37 crc kubenswrapper[4890]: I1125 15:31:37.806546 4890 scope.go:117] "RemoveContainer" containerID="92f0c8a918d324a2aabee016770d39884c7dbbe53d97c21e28654daf6eef0a8a" Nov 25 15:31:37 crc kubenswrapper[4890]: I1125 15:31:37.888389 4890 scope.go:117] "RemoveContainer" containerID="f8010cc46ab44d816173f84c66043f945180f00900ca208dc599fb56bcf061b2" Nov 25 15:31:37 crc kubenswrapper[4890]: I1125 15:31:37.921050 4890 scope.go:117] "RemoveContainer" containerID="cc5632f90b85f3b1e3a9f74ec6b9c2852942cfe570e5db273119a852e7086cb4" Nov 25 15:31:37 crc kubenswrapper[4890]: I1125 15:31:37.984636 4890 scope.go:117] "RemoveContainer" containerID="040c842a972c23a9692699ce5354d20d5e885f515d047301cf76e6ca7da7c2c7" Nov 25 15:31:46 crc kubenswrapper[4890]: I1125 15:31:46.173802 4890 scope.go:117] "RemoveContainer" containerID="9467aa8edc4d8bfe461dc02826750091f343018e37e7f269d305c77d21e832da" Nov 25 15:31:46 crc kubenswrapper[4890]: E1125 15:31:46.174796 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:31:54 crc kubenswrapper[4890]: I1125 15:31:54.043526 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-cwx2k"] Nov 25 15:31:54 crc kubenswrapper[4890]: I1125 15:31:54.057759 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-66a8-account-create-627gj"] Nov 25 15:31:54 crc kubenswrapper[4890]: I1125 15:31:54.070191 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-cwx2k"] Nov 25 15:31:54 crc kubenswrapper[4890]: I1125 15:31:54.081866 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-66a8-account-create-627gj"] Nov 25 15:31:54 crc kubenswrapper[4890]: I1125 15:31:54.190100 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c71a662-55d3-46c8-8e7f-f8595bf24172" path="/var/lib/kubelet/pods/0c71a662-55d3-46c8-8e7f-f8595bf24172/volumes" Nov 25 15:31:54 crc kubenswrapper[4890]: I1125 15:31:54.191008 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfecce7f-d8ed-44d9-bc97-aba2e5b19e8b" path="/var/lib/kubelet/pods/bfecce7f-d8ed-44d9-bc97-aba2e5b19e8b/volumes" Nov 25 15:31:55 crc kubenswrapper[4890]: I1125 15:31:55.058904 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-d21a-account-create-nlvf4"] Nov 25 15:31:55 crc kubenswrapper[4890]: I1125 15:31:55.070078 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-xmpnt"] Nov 25 15:31:55 crc kubenswrapper[4890]: I1125 15:31:55.085852 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-57c6-account-create-skvmk"] Nov 25 15:31:55 crc kubenswrapper[4890]: I1125 15:31:55.098577 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-d21a-account-create-nlvf4"] Nov 25 15:31:55 crc kubenswrapper[4890]: I1125 15:31:55.109236 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-xmpnt"] Nov 25 15:31:55 crc kubenswrapper[4890]: I1125 15:31:55.120273 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-n5rj9"] Nov 25 15:31:55 crc kubenswrapper[4890]: I1125 15:31:55.129691 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-57c6-account-create-skvmk"] Nov 25 15:31:55 crc kubenswrapper[4890]: I1125 15:31:55.141662 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-n5rj9"] Nov 25 15:31:56 crc kubenswrapper[4890]: I1125 15:31:56.190677 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="197b5965-ef39-4de5-9b7c-cb64c982c2a4" path="/var/lib/kubelet/pods/197b5965-ef39-4de5-9b7c-cb64c982c2a4/volumes" Nov 25 15:31:56 crc kubenswrapper[4890]: I1125 15:31:56.193194 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60cb628f-f5b8-43ef-ab79-9753d39d4c04" path="/var/lib/kubelet/pods/60cb628f-f5b8-43ef-ab79-9753d39d4c04/volumes" Nov 25 15:31:56 crc kubenswrapper[4890]: I1125 15:31:56.194243 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bf4c0cd-f20d-4fd9-8077-e733e4196b62" path="/var/lib/kubelet/pods/7bf4c0cd-f20d-4fd9-8077-e733e4196b62/volumes" Nov 25 15:31:56 crc kubenswrapper[4890]: I1125 15:31:56.195416 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fea0bc8f-0181-4f5d-8e81-47bd6d89c4fb" path="/var/lib/kubelet/pods/fea0bc8f-0181-4f5d-8e81-47bd6d89c4fb/volumes" Nov 25 15:31:59 crc kubenswrapper[4890]: I1125 15:31:59.173871 4890 scope.go:117] "RemoveContainer" containerID="9467aa8edc4d8bfe461dc02826750091f343018e37e7f269d305c77d21e832da" Nov 25 15:31:59 crc kubenswrapper[4890]: E1125 15:31:59.174918 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:32:01 crc kubenswrapper[4890]: I1125 15:32:01.055335 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-mg8z4"] Nov 25 15:32:01 crc kubenswrapper[4890]: I1125 15:32:01.066212 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-mg8z4"] Nov 25 15:32:02 crc kubenswrapper[4890]: I1125 15:32:02.188589 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13af51fd-267a-4b92-b08b-49db0d79a866" path="/var/lib/kubelet/pods/13af51fd-267a-4b92-b08b-49db0d79a866/volumes" Nov 25 15:32:12 crc kubenswrapper[4890]: I1125 15:32:12.180932 4890 scope.go:117] "RemoveContainer" containerID="9467aa8edc4d8bfe461dc02826750091f343018e37e7f269d305c77d21e832da" Nov 25 15:32:12 crc kubenswrapper[4890]: E1125 15:32:12.182423 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:32:26 crc kubenswrapper[4890]: I1125 15:32:26.173634 4890 scope.go:117] "RemoveContainer" containerID="9467aa8edc4d8bfe461dc02826750091f343018e37e7f269d305c77d21e832da" Nov 25 15:32:26 crc kubenswrapper[4890]: E1125 15:32:26.174758 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:32:38 crc kubenswrapper[4890]: I1125 15:32:38.291982 4890 scope.go:117] "RemoveContainer" containerID="57a8dfaf17074cb0e7b19059e430746740989c7d1433da6ed075a05661853b0a" Nov 25 15:32:38 crc kubenswrapper[4890]: I1125 15:32:38.359049 4890 scope.go:117] "RemoveContainer" containerID="6e7f48c14398dcfdf4962151c72cc993d0cff6a4d0863894b039abf0f5055f69" Nov 25 15:32:38 crc kubenswrapper[4890]: I1125 15:32:38.391737 4890 scope.go:117] "RemoveContainer" containerID="649bba0742590fe6749edc35bcbbb56579c0de557a05e2c389b08b8f5102f2d6" Nov 25 15:32:38 crc kubenswrapper[4890]: I1125 15:32:38.445104 4890 scope.go:117] "RemoveContainer" containerID="eb87dca0ce22763a819779b239a159b8c0b39406bd4a1afc6dbea87a86ff28c9" Nov 25 15:32:38 crc kubenswrapper[4890]: I1125 15:32:38.494983 4890 scope.go:117] "RemoveContainer" containerID="27b4254c2dd7539a7c6185ff497de399703ce0e26dee37b3760a3ea080dd76a5" Nov 25 15:32:38 crc kubenswrapper[4890]: I1125 15:32:38.540988 4890 scope.go:117] "RemoveContainer" containerID="968f71594fbf195538b3495381442d82ebb8bc7cbd156a37ede4596b9ed58f09" Nov 25 15:32:38 crc kubenswrapper[4890]: I1125 15:32:38.584382 4890 scope.go:117] "RemoveContainer" containerID="3341eafe5523ba14f01d0cde11d921109ba1d96008b5d6c483a72bdd45b3f430" Nov 25 15:32:40 crc kubenswrapper[4890]: I1125 15:32:40.172913 4890 scope.go:117] "RemoveContainer" containerID="9467aa8edc4d8bfe461dc02826750091f343018e37e7f269d305c77d21e832da" Nov 25 15:32:40 crc kubenswrapper[4890]: E1125 15:32:40.173593 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:32:52 crc kubenswrapper[4890]: I1125 15:32:52.179687 4890 scope.go:117] "RemoveContainer" containerID="9467aa8edc4d8bfe461dc02826750091f343018e37e7f269d305c77d21e832da" Nov 25 15:32:52 crc kubenswrapper[4890]: E1125 15:32:52.180490 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:33:04 crc kubenswrapper[4890]: I1125 15:33:04.049147 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-mlvqz"] Nov 25 15:33:04 crc kubenswrapper[4890]: I1125 15:33:04.059482 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-mlvqz"] Nov 25 15:33:04 crc kubenswrapper[4890]: I1125 15:33:04.187264 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0" path="/var/lib/kubelet/pods/910698d6-a2a9-4ba6-90d6-6bed4ad9ecf0/volumes" Nov 25 15:33:05 crc kubenswrapper[4890]: I1125 15:33:05.173189 4890 scope.go:117] "RemoveContainer" containerID="9467aa8edc4d8bfe461dc02826750091f343018e37e7f269d305c77d21e832da" Nov 25 15:33:05 crc kubenswrapper[4890]: I1125 15:33:05.413641 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" event={"ID":"4e4f849d-f239-4727-a73e-18327856929a","Type":"ContainerStarted","Data":"c2b9ae015432b5c3d0ae89bbacbfeadbc65550341b67aff696270c5f412db829"} Nov 25 15:33:09 crc kubenswrapper[4890]: I1125 15:33:09.452906 4890 generic.go:334] "Generic (PLEG): container finished" podID="a2e80fb3-23be-4b01-b663-554207d5d538" containerID="82d1d929bc5761143b76abb4dbac8ffe1f724b89064540f4914bf10674b0f5aa" exitCode=0 Nov 25 15:33:09 crc kubenswrapper[4890]: I1125 15:33:09.453047 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ddcgt" event={"ID":"a2e80fb3-23be-4b01-b663-554207d5d538","Type":"ContainerDied","Data":"82d1d929bc5761143b76abb4dbac8ffe1f724b89064540f4914bf10674b0f5aa"} Nov 25 15:33:10 crc kubenswrapper[4890]: I1125 15:33:10.905694 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ddcgt" Nov 25 15:33:10 crc kubenswrapper[4890]: I1125 15:33:10.992684 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6dbln\" (UniqueName: \"kubernetes.io/projected/a2e80fb3-23be-4b01-b663-554207d5d538-kube-api-access-6dbln\") pod \"a2e80fb3-23be-4b01-b663-554207d5d538\" (UID: \"a2e80fb3-23be-4b01-b663-554207d5d538\") " Nov 25 15:33:10 crc kubenswrapper[4890]: I1125 15:33:10.993119 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2e80fb3-23be-4b01-b663-554207d5d538-bootstrap-combined-ca-bundle\") pod \"a2e80fb3-23be-4b01-b663-554207d5d538\" (UID: \"a2e80fb3-23be-4b01-b663-554207d5d538\") " Nov 25 15:33:10 crc kubenswrapper[4890]: I1125 15:33:10.993193 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a2e80fb3-23be-4b01-b663-554207d5d538-ssh-key\") pod \"a2e80fb3-23be-4b01-b663-554207d5d538\" (UID: \"a2e80fb3-23be-4b01-b663-554207d5d538\") " Nov 25 15:33:10 crc kubenswrapper[4890]: I1125 15:33:10.993279 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a2e80fb3-23be-4b01-b663-554207d5d538-inventory\") pod \"a2e80fb3-23be-4b01-b663-554207d5d538\" (UID: \"a2e80fb3-23be-4b01-b663-554207d5d538\") " Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.000329 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2e80fb3-23be-4b01-b663-554207d5d538-kube-api-access-6dbln" (OuterVolumeSpecName: "kube-api-access-6dbln") pod "a2e80fb3-23be-4b01-b663-554207d5d538" (UID: "a2e80fb3-23be-4b01-b663-554207d5d538"). InnerVolumeSpecName "kube-api-access-6dbln". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.005508 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2e80fb3-23be-4b01-b663-554207d5d538-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "a2e80fb3-23be-4b01-b663-554207d5d538" (UID: "a2e80fb3-23be-4b01-b663-554207d5d538"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.026125 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2e80fb3-23be-4b01-b663-554207d5d538-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a2e80fb3-23be-4b01-b663-554207d5d538" (UID: "a2e80fb3-23be-4b01-b663-554207d5d538"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.028582 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2e80fb3-23be-4b01-b663-554207d5d538-inventory" (OuterVolumeSpecName: "inventory") pod "a2e80fb3-23be-4b01-b663-554207d5d538" (UID: "a2e80fb3-23be-4b01-b663-554207d5d538"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.095921 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6dbln\" (UniqueName: \"kubernetes.io/projected/a2e80fb3-23be-4b01-b663-554207d5d538-kube-api-access-6dbln\") on node \"crc\" DevicePath \"\"" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.095965 4890 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2e80fb3-23be-4b01-b663-554207d5d538-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.095978 4890 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a2e80fb3-23be-4b01-b663-554207d5d538-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.095993 4890 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a2e80fb3-23be-4b01-b663-554207d5d538-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.473602 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ddcgt" event={"ID":"a2e80fb3-23be-4b01-b663-554207d5d538","Type":"ContainerDied","Data":"46857d31f257187cade02b75c1cbd691288b2fb6c1644e81b34a915e28718df0"} Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.473917 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="46857d31f257187cade02b75c1cbd691288b2fb6c1644e81b34a915e28718df0" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.473827 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ddcgt" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.569555 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgklf"] Nov 25 15:33:11 crc kubenswrapper[4890]: E1125 15:33:11.570277 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c070ef9-daba-4ae9-ac05-dff449baf01a" containerName="extract-content" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.570291 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c070ef9-daba-4ae9-ac05-dff449baf01a" containerName="extract-content" Nov 25 15:33:11 crc kubenswrapper[4890]: E1125 15:33:11.570307 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9fbb88d-f5f9-41c6-9719-15be794dfb6c" containerName="extract-content" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.570313 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9fbb88d-f5f9-41c6-9719-15be794dfb6c" containerName="extract-content" Nov 25 15:33:11 crc kubenswrapper[4890]: E1125 15:33:11.570327 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08bbb8ce-17b9-4797-80a7-de3beca270f4" containerName="extract-content" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.570334 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="08bbb8ce-17b9-4797-80a7-de3beca270f4" containerName="extract-content" Nov 25 15:33:11 crc kubenswrapper[4890]: E1125 15:33:11.570345 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00771e9e-5583-42ab-b834-4643d9bbdbec" containerName="registry-server" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.570352 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="00771e9e-5583-42ab-b834-4643d9bbdbec" containerName="registry-server" Nov 25 15:33:11 crc kubenswrapper[4890]: E1125 15:33:11.570362 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0a34ff8-8257-4f49-90d2-90576cbc40b7" containerName="extract-content" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.570368 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0a34ff8-8257-4f49-90d2-90576cbc40b7" containerName="extract-content" Nov 25 15:33:11 crc kubenswrapper[4890]: E1125 15:33:11.570378 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b18eb2e9-77b4-46ba-9e32-31c1e9b73963" containerName="extract-utilities" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.570384 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="b18eb2e9-77b4-46ba-9e32-31c1e9b73963" containerName="extract-utilities" Nov 25 15:33:11 crc kubenswrapper[4890]: E1125 15:33:11.570392 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95656df6-ba37-41ef-a281-87240f41c544" containerName="collect-profiles" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.570397 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="95656df6-ba37-41ef-a281-87240f41c544" containerName="collect-profiles" Nov 25 15:33:11 crc kubenswrapper[4890]: E1125 15:33:11.570408 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fcc307a-b17f-46ab-aabe-0a0361656701" containerName="extract-utilities" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.570413 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fcc307a-b17f-46ab-aabe-0a0361656701" containerName="extract-utilities" Nov 25 15:33:11 crc kubenswrapper[4890]: E1125 15:33:11.570423 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0a34ff8-8257-4f49-90d2-90576cbc40b7" containerName="registry-server" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.570428 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0a34ff8-8257-4f49-90d2-90576cbc40b7" containerName="registry-server" Nov 25 15:33:11 crc kubenswrapper[4890]: E1125 15:33:11.570437 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7235b7e-d6fe-45a2-93fd-66b0a77b4b76" containerName="registry-server" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.570444 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7235b7e-d6fe-45a2-93fd-66b0a77b4b76" containerName="registry-server" Nov 25 15:33:11 crc kubenswrapper[4890]: E1125 15:33:11.570458 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0a34ff8-8257-4f49-90d2-90576cbc40b7" containerName="extract-utilities" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.570464 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0a34ff8-8257-4f49-90d2-90576cbc40b7" containerName="extract-utilities" Nov 25 15:33:11 crc kubenswrapper[4890]: E1125 15:33:11.570471 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c070ef9-daba-4ae9-ac05-dff449baf01a" containerName="extract-utilities" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.570477 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c070ef9-daba-4ae9-ac05-dff449baf01a" containerName="extract-utilities" Nov 25 15:33:11 crc kubenswrapper[4890]: E1125 15:33:11.570494 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00771e9e-5583-42ab-b834-4643d9bbdbec" containerName="extract-utilities" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.570501 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="00771e9e-5583-42ab-b834-4643d9bbdbec" containerName="extract-utilities" Nov 25 15:33:11 crc kubenswrapper[4890]: E1125 15:33:11.570513 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08bbb8ce-17b9-4797-80a7-de3beca270f4" containerName="extract-utilities" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.570519 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="08bbb8ce-17b9-4797-80a7-de3beca270f4" containerName="extract-utilities" Nov 25 15:33:11 crc kubenswrapper[4890]: E1125 15:33:11.570525 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c070ef9-daba-4ae9-ac05-dff449baf01a" containerName="registry-server" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.570530 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c070ef9-daba-4ae9-ac05-dff449baf01a" containerName="registry-server" Nov 25 15:33:11 crc kubenswrapper[4890]: E1125 15:33:11.570539 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb551926-2f33-4616-a423-caee14e79f74" containerName="extract-content" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.570545 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb551926-2f33-4616-a423-caee14e79f74" containerName="extract-content" Nov 25 15:33:11 crc kubenswrapper[4890]: E1125 15:33:11.570556 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fcc307a-b17f-46ab-aabe-0a0361656701" containerName="extract-content" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.570562 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fcc307a-b17f-46ab-aabe-0a0361656701" containerName="extract-content" Nov 25 15:33:11 crc kubenswrapper[4890]: E1125 15:33:11.570569 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b53e3920-fe5e-45fe-9b2d-495a5cb16b2c" containerName="registry-server" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.570574 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="b53e3920-fe5e-45fe-9b2d-495a5cb16b2c" containerName="registry-server" Nov 25 15:33:11 crc kubenswrapper[4890]: E1125 15:33:11.570590 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eacc1f05-701e-49af-b2d7-9c049f6cddb9" containerName="extract-content" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.570595 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="eacc1f05-701e-49af-b2d7-9c049f6cddb9" containerName="extract-content" Nov 25 15:33:11 crc kubenswrapper[4890]: E1125 15:33:11.570608 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f05c16a-1239-42a7-9910-ca03ef42b0f7" containerName="registry-server" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.570613 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f05c16a-1239-42a7-9910-ca03ef42b0f7" containerName="registry-server" Nov 25 15:33:11 crc kubenswrapper[4890]: E1125 15:33:11.570623 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eacc1f05-701e-49af-b2d7-9c049f6cddb9" containerName="extract-utilities" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.570628 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="eacc1f05-701e-49af-b2d7-9c049f6cddb9" containerName="extract-utilities" Nov 25 15:33:11 crc kubenswrapper[4890]: E1125 15:33:11.570640 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb551926-2f33-4616-a423-caee14e79f74" containerName="extract-utilities" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.570646 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb551926-2f33-4616-a423-caee14e79f74" containerName="extract-utilities" Nov 25 15:33:11 crc kubenswrapper[4890]: E1125 15:33:11.570653 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c9a571e-db77-4303-96cc-c111add7626d" containerName="registry-server" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.570659 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c9a571e-db77-4303-96cc-c111add7626d" containerName="registry-server" Nov 25 15:33:11 crc kubenswrapper[4890]: E1125 15:33:11.570670 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c9a571e-db77-4303-96cc-c111add7626d" containerName="extract-utilities" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.570676 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c9a571e-db77-4303-96cc-c111add7626d" containerName="extract-utilities" Nov 25 15:33:11 crc kubenswrapper[4890]: E1125 15:33:11.570688 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f05c16a-1239-42a7-9910-ca03ef42b0f7" containerName="extract-utilities" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.570694 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f05c16a-1239-42a7-9910-ca03ef42b0f7" containerName="extract-utilities" Nov 25 15:33:11 crc kubenswrapper[4890]: E1125 15:33:11.570704 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b18eb2e9-77b4-46ba-9e32-31c1e9b73963" containerName="registry-server" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.570709 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="b18eb2e9-77b4-46ba-9e32-31c1e9b73963" containerName="registry-server" Nov 25 15:33:11 crc kubenswrapper[4890]: E1125 15:33:11.570723 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9fbb88d-f5f9-41c6-9719-15be794dfb6c" containerName="registry-server" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.570728 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9fbb88d-f5f9-41c6-9719-15be794dfb6c" containerName="registry-server" Nov 25 15:33:11 crc kubenswrapper[4890]: E1125 15:33:11.570743 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6febf84d-85dd-45d8-a32e-bc978311ed98" containerName="extract-utilities" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.570749 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="6febf84d-85dd-45d8-a32e-bc978311ed98" containerName="extract-utilities" Nov 25 15:33:11 crc kubenswrapper[4890]: E1125 15:33:11.570759 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c9a571e-db77-4303-96cc-c111add7626d" containerName="extract-content" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.570765 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c9a571e-db77-4303-96cc-c111add7626d" containerName="extract-content" Nov 25 15:33:11 crc kubenswrapper[4890]: E1125 15:33:11.570776 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b53e3920-fe5e-45fe-9b2d-495a5cb16b2c" containerName="extract-utilities" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.570782 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="b53e3920-fe5e-45fe-9b2d-495a5cb16b2c" containerName="extract-utilities" Nov 25 15:33:11 crc kubenswrapper[4890]: E1125 15:33:11.570789 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7235b7e-d6fe-45a2-93fd-66b0a77b4b76" containerName="extract-content" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.570795 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7235b7e-d6fe-45a2-93fd-66b0a77b4b76" containerName="extract-content" Nov 25 15:33:11 crc kubenswrapper[4890]: E1125 15:33:11.570807 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7235b7e-d6fe-45a2-93fd-66b0a77b4b76" containerName="extract-utilities" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.570813 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7235b7e-d6fe-45a2-93fd-66b0a77b4b76" containerName="extract-utilities" Nov 25 15:33:11 crc kubenswrapper[4890]: E1125 15:33:11.570819 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b53e3920-fe5e-45fe-9b2d-495a5cb16b2c" containerName="extract-content" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.570826 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="b53e3920-fe5e-45fe-9b2d-495a5cb16b2c" containerName="extract-content" Nov 25 15:33:11 crc kubenswrapper[4890]: E1125 15:33:11.570834 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb551926-2f33-4616-a423-caee14e79f74" containerName="registry-server" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.570840 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb551926-2f33-4616-a423-caee14e79f74" containerName="registry-server" Nov 25 15:33:11 crc kubenswrapper[4890]: E1125 15:33:11.570850 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00771e9e-5583-42ab-b834-4643d9bbdbec" containerName="extract-content" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.570855 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="00771e9e-5583-42ab-b834-4643d9bbdbec" containerName="extract-content" Nov 25 15:33:11 crc kubenswrapper[4890]: E1125 15:33:11.570865 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6febf84d-85dd-45d8-a32e-bc978311ed98" containerName="extract-content" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.570871 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="6febf84d-85dd-45d8-a32e-bc978311ed98" containerName="extract-content" Nov 25 15:33:11 crc kubenswrapper[4890]: E1125 15:33:11.570883 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f05c16a-1239-42a7-9910-ca03ef42b0f7" containerName="registry-server" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.570890 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f05c16a-1239-42a7-9910-ca03ef42b0f7" containerName="registry-server" Nov 25 15:33:11 crc kubenswrapper[4890]: E1125 15:33:11.570901 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08bbb8ce-17b9-4797-80a7-de3beca270f4" containerName="registry-server" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.570908 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="08bbb8ce-17b9-4797-80a7-de3beca270f4" containerName="registry-server" Nov 25 15:33:11 crc kubenswrapper[4890]: E1125 15:33:11.570920 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6febf84d-85dd-45d8-a32e-bc978311ed98" containerName="registry-server" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.570926 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="6febf84d-85dd-45d8-a32e-bc978311ed98" containerName="registry-server" Nov 25 15:33:11 crc kubenswrapper[4890]: E1125 15:33:11.570941 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fcc307a-b17f-46ab-aabe-0a0361656701" containerName="registry-server" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.570947 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fcc307a-b17f-46ab-aabe-0a0361656701" containerName="registry-server" Nov 25 15:33:11 crc kubenswrapper[4890]: E1125 15:33:11.570958 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f05c16a-1239-42a7-9910-ca03ef42b0f7" containerName="extract-content" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.570964 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f05c16a-1239-42a7-9910-ca03ef42b0f7" containerName="extract-content" Nov 25 15:33:11 crc kubenswrapper[4890]: E1125 15:33:11.570973 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b18eb2e9-77b4-46ba-9e32-31c1e9b73963" containerName="extract-content" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.570978 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="b18eb2e9-77b4-46ba-9e32-31c1e9b73963" containerName="extract-content" Nov 25 15:33:11 crc kubenswrapper[4890]: E1125 15:33:11.570987 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eacc1f05-701e-49af-b2d7-9c049f6cddb9" containerName="registry-server" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.570992 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="eacc1f05-701e-49af-b2d7-9c049f6cddb9" containerName="registry-server" Nov 25 15:33:11 crc kubenswrapper[4890]: E1125 15:33:11.571001 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08bbb8ce-17b9-4797-80a7-de3beca270f4" containerName="registry-server" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.571007 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="08bbb8ce-17b9-4797-80a7-de3beca270f4" containerName="registry-server" Nov 25 15:33:11 crc kubenswrapper[4890]: E1125 15:33:11.571018 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2e80fb3-23be-4b01-b663-554207d5d538" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.571024 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2e80fb3-23be-4b01-b663-554207d5d538" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 25 15:33:11 crc kubenswrapper[4890]: E1125 15:33:11.571035 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9fbb88d-f5f9-41c6-9719-15be794dfb6c" containerName="extract-utilities" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.571040 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9fbb88d-f5f9-41c6-9719-15be794dfb6c" containerName="extract-utilities" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.571239 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="00771e9e-5583-42ab-b834-4643d9bbdbec" containerName="registry-server" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.571251 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="1fcc307a-b17f-46ab-aabe-0a0361656701" containerName="registry-server" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.571264 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f05c16a-1239-42a7-9910-ca03ef42b0f7" containerName="registry-server" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.571281 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb551926-2f33-4616-a423-caee14e79f74" containerName="registry-server" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.571288 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f05c16a-1239-42a7-9910-ca03ef42b0f7" containerName="registry-server" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.571299 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c070ef9-daba-4ae9-ac05-dff449baf01a" containerName="registry-server" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.571309 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="08bbb8ce-17b9-4797-80a7-de3beca270f4" containerName="registry-server" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.571318 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="08bbb8ce-17b9-4797-80a7-de3beca270f4" containerName="registry-server" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.571327 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="6febf84d-85dd-45d8-a32e-bc978311ed98" containerName="registry-server" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.571335 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="b18eb2e9-77b4-46ba-9e32-31c1e9b73963" containerName="registry-server" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.571343 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7235b7e-d6fe-45a2-93fd-66b0a77b4b76" containerName="registry-server" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.571352 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c9a571e-db77-4303-96cc-c111add7626d" containerName="registry-server" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.571361 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="95656df6-ba37-41ef-a281-87240f41c544" containerName="collect-profiles" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.571376 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9fbb88d-f5f9-41c6-9719-15be794dfb6c" containerName="registry-server" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.571388 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="eacc1f05-701e-49af-b2d7-9c049f6cddb9" containerName="registry-server" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.571399 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2e80fb3-23be-4b01-b663-554207d5d538" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.571407 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="b53e3920-fe5e-45fe-9b2d-495a5cb16b2c" containerName="registry-server" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.571417 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0a34ff8-8257-4f49-90d2-90576cbc40b7" containerName="registry-server" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.572048 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgklf" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.575254 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-22zzp" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.575263 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.578503 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.578985 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.579712 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgklf"] Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.710919 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef539e05-0227-4553-b6a8-0a70f5eadf8c-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-sgklf\" (UID: \"ef539e05-0227-4553-b6a8-0a70f5eadf8c\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgklf" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.711196 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef539e05-0227-4553-b6a8-0a70f5eadf8c-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-sgklf\" (UID: \"ef539e05-0227-4553-b6a8-0a70f5eadf8c\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgklf" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.711246 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8npm\" (UniqueName: \"kubernetes.io/projected/ef539e05-0227-4553-b6a8-0a70f5eadf8c-kube-api-access-c8npm\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-sgklf\" (UID: \"ef539e05-0227-4553-b6a8-0a70f5eadf8c\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgklf" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.813336 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef539e05-0227-4553-b6a8-0a70f5eadf8c-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-sgklf\" (UID: \"ef539e05-0227-4553-b6a8-0a70f5eadf8c\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgklf" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.813443 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef539e05-0227-4553-b6a8-0a70f5eadf8c-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-sgklf\" (UID: \"ef539e05-0227-4553-b6a8-0a70f5eadf8c\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgklf" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.813472 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8npm\" (UniqueName: \"kubernetes.io/projected/ef539e05-0227-4553-b6a8-0a70f5eadf8c-kube-api-access-c8npm\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-sgklf\" (UID: \"ef539e05-0227-4553-b6a8-0a70f5eadf8c\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgklf" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.818872 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef539e05-0227-4553-b6a8-0a70f5eadf8c-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-sgklf\" (UID: \"ef539e05-0227-4553-b6a8-0a70f5eadf8c\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgklf" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.819197 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef539e05-0227-4553-b6a8-0a70f5eadf8c-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-sgklf\" (UID: \"ef539e05-0227-4553-b6a8-0a70f5eadf8c\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgklf" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.836420 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8npm\" (UniqueName: \"kubernetes.io/projected/ef539e05-0227-4553-b6a8-0a70f5eadf8c-kube-api-access-c8npm\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-sgklf\" (UID: \"ef539e05-0227-4553-b6a8-0a70f5eadf8c\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgklf" Nov 25 15:33:11 crc kubenswrapper[4890]: I1125 15:33:11.894446 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgklf" Nov 25 15:33:12 crc kubenswrapper[4890]: I1125 15:33:12.455798 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgklf"] Nov 25 15:33:12 crc kubenswrapper[4890]: I1125 15:33:12.458864 4890 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 15:33:12 crc kubenswrapper[4890]: I1125 15:33:12.484928 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgklf" event={"ID":"ef539e05-0227-4553-b6a8-0a70f5eadf8c","Type":"ContainerStarted","Data":"be1b9b22c29bf03daa39c0ca3b0b385a444f80d029b5ab55b621958f8cfe9214"} Nov 25 15:33:13 crc kubenswrapper[4890]: I1125 15:33:13.497886 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgklf" event={"ID":"ef539e05-0227-4553-b6a8-0a70f5eadf8c","Type":"ContainerStarted","Data":"ccfe0bc53f3987150c459d172a8e2a9d7035d9db94aefa5e39c2101791addb21"} Nov 25 15:33:29 crc kubenswrapper[4890]: I1125 15:33:29.043102 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgklf" podStartSLOduration=17.597427157 podStartE2EDuration="18.043081631s" podCreationTimestamp="2025-11-25 15:33:11 +0000 UTC" firstStartedPulling="2025-11-25 15:33:12.458587496 +0000 UTC m=+1850.901050106" lastFinishedPulling="2025-11-25 15:33:12.90424197 +0000 UTC m=+1851.346704580" observedRunningTime="2025-11-25 15:33:13.52667447 +0000 UTC m=+1851.969137090" watchObservedRunningTime="2025-11-25 15:33:29.043081631 +0000 UTC m=+1867.485544241" Nov 25 15:33:29 crc kubenswrapper[4890]: I1125 15:33:29.047567 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-t4rb6"] Nov 25 15:33:29 crc kubenswrapper[4890]: I1125 15:33:29.057082 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-t4rb6"] Nov 25 15:33:30 crc kubenswrapper[4890]: I1125 15:33:30.184986 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="872e8100-002a-4ac1-b699-b35111573d7d" path="/var/lib/kubelet/pods/872e8100-002a-4ac1-b699-b35111573d7d/volumes" Nov 25 15:33:38 crc kubenswrapper[4890]: I1125 15:33:38.749219 4890 scope.go:117] "RemoveContainer" containerID="b317dbbc6ec2fb8dfd39a367f59796a433cbc2e2182fcfcad92df15433c00b42" Nov 25 15:33:38 crc kubenswrapper[4890]: I1125 15:33:38.830098 4890 scope.go:117] "RemoveContainer" containerID="a8e112324cbebe6d570d7610b4ef244ecf7bb4b8aab01e300595ae4f47c1ee92" Nov 25 15:33:38 crc kubenswrapper[4890]: I1125 15:33:38.858608 4890 scope.go:117] "RemoveContainer" containerID="f38ab6ac425c917e9cfcdf44dedf4ba6b91c5339078f0d15f992406d87b07283" Nov 25 15:33:42 crc kubenswrapper[4890]: I1125 15:33:42.025826 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-828xw"] Nov 25 15:33:42 crc kubenswrapper[4890]: I1125 15:33:42.037787 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-828xw"] Nov 25 15:33:42 crc kubenswrapper[4890]: I1125 15:33:42.183600 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70c6170b-9975-487c-ac2d-a2a309833658" path="/var/lib/kubelet/pods/70c6170b-9975-487c-ac2d-a2a309833658/volumes" Nov 25 15:34:12 crc kubenswrapper[4890]: I1125 15:34:12.043555 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-s2qx4"] Nov 25 15:34:12 crc kubenswrapper[4890]: I1125 15:34:12.055549 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-s2qx4"] Nov 25 15:34:12 crc kubenswrapper[4890]: I1125 15:34:12.187679 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="519d05c4-07a9-4749-a8ae-072cdd3fd0af" path="/var/lib/kubelet/pods/519d05c4-07a9-4749-a8ae-072cdd3fd0af/volumes" Nov 25 15:34:38 crc kubenswrapper[4890]: I1125 15:34:38.978086 4890 scope.go:117] "RemoveContainer" containerID="1557774a1642615323744cb5147878a67bf6499e0cacf33a93edbb5e592537bc" Nov 25 15:34:39 crc kubenswrapper[4890]: I1125 15:34:39.042395 4890 scope.go:117] "RemoveContainer" containerID="35098fd0841e4cedb8a48c9b6dd5cc6d32370f0a2cb7b03e08538da936e45d55" Nov 25 15:34:39 crc kubenswrapper[4890]: I1125 15:34:39.086690 4890 scope.go:117] "RemoveContainer" containerID="b76d7672c9af03f6cfa1ad825433495b765e29311771d9a0c6a518019d549333" Nov 25 15:34:49 crc kubenswrapper[4890]: I1125 15:34:49.631359 4890 generic.go:334] "Generic (PLEG): container finished" podID="ef539e05-0227-4553-b6a8-0a70f5eadf8c" containerID="ccfe0bc53f3987150c459d172a8e2a9d7035d9db94aefa5e39c2101791addb21" exitCode=0 Nov 25 15:34:49 crc kubenswrapper[4890]: I1125 15:34:49.631887 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgklf" event={"ID":"ef539e05-0227-4553-b6a8-0a70f5eadf8c","Type":"ContainerDied","Data":"ccfe0bc53f3987150c459d172a8e2a9d7035d9db94aefa5e39c2101791addb21"} Nov 25 15:34:51 crc kubenswrapper[4890]: I1125 15:34:51.070982 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgklf" Nov 25 15:34:51 crc kubenswrapper[4890]: I1125 15:34:51.193395 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef539e05-0227-4553-b6a8-0a70f5eadf8c-ssh-key\") pod \"ef539e05-0227-4553-b6a8-0a70f5eadf8c\" (UID: \"ef539e05-0227-4553-b6a8-0a70f5eadf8c\") " Nov 25 15:34:51 crc kubenswrapper[4890]: I1125 15:34:51.193616 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef539e05-0227-4553-b6a8-0a70f5eadf8c-inventory\") pod \"ef539e05-0227-4553-b6a8-0a70f5eadf8c\" (UID: \"ef539e05-0227-4553-b6a8-0a70f5eadf8c\") " Nov 25 15:34:51 crc kubenswrapper[4890]: I1125 15:34:51.193669 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8npm\" (UniqueName: \"kubernetes.io/projected/ef539e05-0227-4553-b6a8-0a70f5eadf8c-kube-api-access-c8npm\") pod \"ef539e05-0227-4553-b6a8-0a70f5eadf8c\" (UID: \"ef539e05-0227-4553-b6a8-0a70f5eadf8c\") " Nov 25 15:34:51 crc kubenswrapper[4890]: I1125 15:34:51.202978 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef539e05-0227-4553-b6a8-0a70f5eadf8c-kube-api-access-c8npm" (OuterVolumeSpecName: "kube-api-access-c8npm") pod "ef539e05-0227-4553-b6a8-0a70f5eadf8c" (UID: "ef539e05-0227-4553-b6a8-0a70f5eadf8c"). InnerVolumeSpecName "kube-api-access-c8npm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:34:51 crc kubenswrapper[4890]: I1125 15:34:51.229843 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef539e05-0227-4553-b6a8-0a70f5eadf8c-inventory" (OuterVolumeSpecName: "inventory") pod "ef539e05-0227-4553-b6a8-0a70f5eadf8c" (UID: "ef539e05-0227-4553-b6a8-0a70f5eadf8c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:34:51 crc kubenswrapper[4890]: I1125 15:34:51.233926 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef539e05-0227-4553-b6a8-0a70f5eadf8c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ef539e05-0227-4553-b6a8-0a70f5eadf8c" (UID: "ef539e05-0227-4553-b6a8-0a70f5eadf8c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:34:51 crc kubenswrapper[4890]: I1125 15:34:51.301367 4890 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef539e05-0227-4553-b6a8-0a70f5eadf8c-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 15:34:51 crc kubenswrapper[4890]: I1125 15:34:51.301623 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c8npm\" (UniqueName: \"kubernetes.io/projected/ef539e05-0227-4553-b6a8-0a70f5eadf8c-kube-api-access-c8npm\") on node \"crc\" DevicePath \"\"" Nov 25 15:34:51 crc kubenswrapper[4890]: I1125 15:34:51.301730 4890 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef539e05-0227-4553-b6a8-0a70f5eadf8c-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 15:34:51 crc kubenswrapper[4890]: I1125 15:34:51.654277 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgklf" event={"ID":"ef539e05-0227-4553-b6a8-0a70f5eadf8c","Type":"ContainerDied","Data":"be1b9b22c29bf03daa39c0ca3b0b385a444f80d029b5ab55b621958f8cfe9214"} Nov 25 15:34:51 crc kubenswrapper[4890]: I1125 15:34:51.654318 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="be1b9b22c29bf03daa39c0ca3b0b385a444f80d029b5ab55b621958f8cfe9214" Nov 25 15:34:51 crc kubenswrapper[4890]: I1125 15:34:51.654359 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgklf" Nov 25 15:34:51 crc kubenswrapper[4890]: I1125 15:34:51.747057 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9h2zh"] Nov 25 15:34:51 crc kubenswrapper[4890]: E1125 15:34:51.748153 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef539e05-0227-4553-b6a8-0a70f5eadf8c" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 25 15:34:51 crc kubenswrapper[4890]: I1125 15:34:51.748198 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef539e05-0227-4553-b6a8-0a70f5eadf8c" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 25 15:34:51 crc kubenswrapper[4890]: I1125 15:34:51.748553 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef539e05-0227-4553-b6a8-0a70f5eadf8c" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 25 15:34:51 crc kubenswrapper[4890]: I1125 15:34:51.749419 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9h2zh" Nov 25 15:34:51 crc kubenswrapper[4890]: I1125 15:34:51.752142 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 15:34:51 crc kubenswrapper[4890]: I1125 15:34:51.752927 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 15:34:51 crc kubenswrapper[4890]: I1125 15:34:51.753045 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 15:34:51 crc kubenswrapper[4890]: I1125 15:34:51.753245 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-22zzp" Nov 25 15:34:51 crc kubenswrapper[4890]: I1125 15:34:51.774006 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9h2zh"] Nov 25 15:34:51 crc kubenswrapper[4890]: I1125 15:34:51.919193 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nbnz\" (UniqueName: \"kubernetes.io/projected/35c7409b-9251-4ebb-9a04-8de78ea0fde0-kube-api-access-4nbnz\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9h2zh\" (UID: \"35c7409b-9251-4ebb-9a04-8de78ea0fde0\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9h2zh" Nov 25 15:34:51 crc kubenswrapper[4890]: I1125 15:34:51.919733 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/35c7409b-9251-4ebb-9a04-8de78ea0fde0-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9h2zh\" (UID: \"35c7409b-9251-4ebb-9a04-8de78ea0fde0\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9h2zh" Nov 25 15:34:51 crc kubenswrapper[4890]: I1125 15:34:51.919788 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/35c7409b-9251-4ebb-9a04-8de78ea0fde0-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9h2zh\" (UID: \"35c7409b-9251-4ebb-9a04-8de78ea0fde0\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9h2zh" Nov 25 15:34:52 crc kubenswrapper[4890]: I1125 15:34:52.022540 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/35c7409b-9251-4ebb-9a04-8de78ea0fde0-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9h2zh\" (UID: \"35c7409b-9251-4ebb-9a04-8de78ea0fde0\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9h2zh" Nov 25 15:34:52 crc kubenswrapper[4890]: I1125 15:34:52.023016 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/35c7409b-9251-4ebb-9a04-8de78ea0fde0-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9h2zh\" (UID: \"35c7409b-9251-4ebb-9a04-8de78ea0fde0\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9h2zh" Nov 25 15:34:52 crc kubenswrapper[4890]: I1125 15:34:52.023365 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nbnz\" (UniqueName: \"kubernetes.io/projected/35c7409b-9251-4ebb-9a04-8de78ea0fde0-kube-api-access-4nbnz\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9h2zh\" (UID: \"35c7409b-9251-4ebb-9a04-8de78ea0fde0\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9h2zh" Nov 25 15:34:52 crc kubenswrapper[4890]: I1125 15:34:52.030572 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/35c7409b-9251-4ebb-9a04-8de78ea0fde0-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9h2zh\" (UID: \"35c7409b-9251-4ebb-9a04-8de78ea0fde0\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9h2zh" Nov 25 15:34:52 crc kubenswrapper[4890]: I1125 15:34:52.047191 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/35c7409b-9251-4ebb-9a04-8de78ea0fde0-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9h2zh\" (UID: \"35c7409b-9251-4ebb-9a04-8de78ea0fde0\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9h2zh" Nov 25 15:34:52 crc kubenswrapper[4890]: I1125 15:34:52.047777 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nbnz\" (UniqueName: \"kubernetes.io/projected/35c7409b-9251-4ebb-9a04-8de78ea0fde0-kube-api-access-4nbnz\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9h2zh\" (UID: \"35c7409b-9251-4ebb-9a04-8de78ea0fde0\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9h2zh" Nov 25 15:34:52 crc kubenswrapper[4890]: I1125 15:34:52.081775 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9h2zh" Nov 25 15:34:52 crc kubenswrapper[4890]: I1125 15:34:52.607148 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9h2zh"] Nov 25 15:34:52 crc kubenswrapper[4890]: I1125 15:34:52.665714 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9h2zh" event={"ID":"35c7409b-9251-4ebb-9a04-8de78ea0fde0","Type":"ContainerStarted","Data":"90f411a0ed765eb58771e18a3411329624b68d738d4dae051d60dfb1b451998c"} Nov 25 15:34:54 crc kubenswrapper[4890]: I1125 15:34:54.683597 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9h2zh" event={"ID":"35c7409b-9251-4ebb-9a04-8de78ea0fde0","Type":"ContainerStarted","Data":"0cdc8b92cfe6d9c46ab1bf967fb70c118e83af388b12b4fc58654336a5088691"} Nov 25 15:34:54 crc kubenswrapper[4890]: I1125 15:34:54.700724 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9h2zh" podStartSLOduration=2.8865487610000002 podStartE2EDuration="3.700701341s" podCreationTimestamp="2025-11-25 15:34:51 +0000 UTC" firstStartedPulling="2025-11-25 15:34:52.613394155 +0000 UTC m=+1951.055856765" lastFinishedPulling="2025-11-25 15:34:53.427546735 +0000 UTC m=+1951.870009345" observedRunningTime="2025-11-25 15:34:54.698994948 +0000 UTC m=+1953.141457558" watchObservedRunningTime="2025-11-25 15:34:54.700701341 +0000 UTC m=+1953.143163951" Nov 25 15:35:26 crc kubenswrapper[4890]: I1125 15:35:26.447130 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 15:35:26 crc kubenswrapper[4890]: I1125 15:35:26.447758 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 15:35:56 crc kubenswrapper[4890]: I1125 15:35:56.447249 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 15:35:56 crc kubenswrapper[4890]: I1125 15:35:56.448176 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 15:36:02 crc kubenswrapper[4890]: I1125 15:36:02.423739 4890 generic.go:334] "Generic (PLEG): container finished" podID="35c7409b-9251-4ebb-9a04-8de78ea0fde0" containerID="0cdc8b92cfe6d9c46ab1bf967fb70c118e83af388b12b4fc58654336a5088691" exitCode=0 Nov 25 15:36:02 crc kubenswrapper[4890]: I1125 15:36:02.423832 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9h2zh" event={"ID":"35c7409b-9251-4ebb-9a04-8de78ea0fde0","Type":"ContainerDied","Data":"0cdc8b92cfe6d9c46ab1bf967fb70c118e83af388b12b4fc58654336a5088691"} Nov 25 15:36:03 crc kubenswrapper[4890]: I1125 15:36:03.894103 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9h2zh" Nov 25 15:36:04 crc kubenswrapper[4890]: I1125 15:36:04.003041 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4nbnz\" (UniqueName: \"kubernetes.io/projected/35c7409b-9251-4ebb-9a04-8de78ea0fde0-kube-api-access-4nbnz\") pod \"35c7409b-9251-4ebb-9a04-8de78ea0fde0\" (UID: \"35c7409b-9251-4ebb-9a04-8de78ea0fde0\") " Nov 25 15:36:04 crc kubenswrapper[4890]: I1125 15:36:04.003223 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/35c7409b-9251-4ebb-9a04-8de78ea0fde0-ssh-key\") pod \"35c7409b-9251-4ebb-9a04-8de78ea0fde0\" (UID: \"35c7409b-9251-4ebb-9a04-8de78ea0fde0\") " Nov 25 15:36:04 crc kubenswrapper[4890]: I1125 15:36:04.003257 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/35c7409b-9251-4ebb-9a04-8de78ea0fde0-inventory\") pod \"35c7409b-9251-4ebb-9a04-8de78ea0fde0\" (UID: \"35c7409b-9251-4ebb-9a04-8de78ea0fde0\") " Nov 25 15:36:04 crc kubenswrapper[4890]: I1125 15:36:04.010419 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35c7409b-9251-4ebb-9a04-8de78ea0fde0-kube-api-access-4nbnz" (OuterVolumeSpecName: "kube-api-access-4nbnz") pod "35c7409b-9251-4ebb-9a04-8de78ea0fde0" (UID: "35c7409b-9251-4ebb-9a04-8de78ea0fde0"). InnerVolumeSpecName "kube-api-access-4nbnz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:36:04 crc kubenswrapper[4890]: I1125 15:36:04.037790 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35c7409b-9251-4ebb-9a04-8de78ea0fde0-inventory" (OuterVolumeSpecName: "inventory") pod "35c7409b-9251-4ebb-9a04-8de78ea0fde0" (UID: "35c7409b-9251-4ebb-9a04-8de78ea0fde0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:36:04 crc kubenswrapper[4890]: I1125 15:36:04.038367 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35c7409b-9251-4ebb-9a04-8de78ea0fde0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "35c7409b-9251-4ebb-9a04-8de78ea0fde0" (UID: "35c7409b-9251-4ebb-9a04-8de78ea0fde0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:36:04 crc kubenswrapper[4890]: I1125 15:36:04.105505 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4nbnz\" (UniqueName: \"kubernetes.io/projected/35c7409b-9251-4ebb-9a04-8de78ea0fde0-kube-api-access-4nbnz\") on node \"crc\" DevicePath \"\"" Nov 25 15:36:04 crc kubenswrapper[4890]: I1125 15:36:04.105551 4890 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/35c7409b-9251-4ebb-9a04-8de78ea0fde0-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 15:36:04 crc kubenswrapper[4890]: I1125 15:36:04.105562 4890 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/35c7409b-9251-4ebb-9a04-8de78ea0fde0-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 15:36:04 crc kubenswrapper[4890]: I1125 15:36:04.446803 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9h2zh" event={"ID":"35c7409b-9251-4ebb-9a04-8de78ea0fde0","Type":"ContainerDied","Data":"90f411a0ed765eb58771e18a3411329624b68d738d4dae051d60dfb1b451998c"} Nov 25 15:36:04 crc kubenswrapper[4890]: I1125 15:36:04.447362 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="90f411a0ed765eb58771e18a3411329624b68d738d4dae051d60dfb1b451998c" Nov 25 15:36:04 crc kubenswrapper[4890]: I1125 15:36:04.446873 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9h2zh" Nov 25 15:36:04 crc kubenswrapper[4890]: I1125 15:36:04.556769 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6vbxl"] Nov 25 15:36:04 crc kubenswrapper[4890]: E1125 15:36:04.557244 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35c7409b-9251-4ebb-9a04-8de78ea0fde0" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 25 15:36:04 crc kubenswrapper[4890]: I1125 15:36:04.557268 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="35c7409b-9251-4ebb-9a04-8de78ea0fde0" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 25 15:36:04 crc kubenswrapper[4890]: I1125 15:36:04.557629 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="35c7409b-9251-4ebb-9a04-8de78ea0fde0" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 25 15:36:04 crc kubenswrapper[4890]: I1125 15:36:04.558587 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6vbxl" Nov 25 15:36:04 crc kubenswrapper[4890]: I1125 15:36:04.561419 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-22zzp" Nov 25 15:36:04 crc kubenswrapper[4890]: I1125 15:36:04.562176 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 15:36:04 crc kubenswrapper[4890]: I1125 15:36:04.562254 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 15:36:04 crc kubenswrapper[4890]: I1125 15:36:04.563345 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 15:36:04 crc kubenswrapper[4890]: I1125 15:36:04.579773 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6vbxl"] Nov 25 15:36:04 crc kubenswrapper[4890]: I1125 15:36:04.616921 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8394f488-cfd0-4fdc-aa22-96fe01aec73e-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6vbxl\" (UID: \"8394f488-cfd0-4fdc-aa22-96fe01aec73e\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6vbxl" Nov 25 15:36:04 crc kubenswrapper[4890]: I1125 15:36:04.617062 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8394f488-cfd0-4fdc-aa22-96fe01aec73e-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6vbxl\" (UID: \"8394f488-cfd0-4fdc-aa22-96fe01aec73e\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6vbxl" Nov 25 15:36:04 crc kubenswrapper[4890]: I1125 15:36:04.617130 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fmgc\" (UniqueName: \"kubernetes.io/projected/8394f488-cfd0-4fdc-aa22-96fe01aec73e-kube-api-access-2fmgc\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6vbxl\" (UID: \"8394f488-cfd0-4fdc-aa22-96fe01aec73e\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6vbxl" Nov 25 15:36:04 crc kubenswrapper[4890]: I1125 15:36:04.719095 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8394f488-cfd0-4fdc-aa22-96fe01aec73e-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6vbxl\" (UID: \"8394f488-cfd0-4fdc-aa22-96fe01aec73e\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6vbxl" Nov 25 15:36:04 crc kubenswrapper[4890]: I1125 15:36:04.719300 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8394f488-cfd0-4fdc-aa22-96fe01aec73e-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6vbxl\" (UID: \"8394f488-cfd0-4fdc-aa22-96fe01aec73e\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6vbxl" Nov 25 15:36:04 crc kubenswrapper[4890]: I1125 15:36:04.719390 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fmgc\" (UniqueName: \"kubernetes.io/projected/8394f488-cfd0-4fdc-aa22-96fe01aec73e-kube-api-access-2fmgc\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6vbxl\" (UID: \"8394f488-cfd0-4fdc-aa22-96fe01aec73e\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6vbxl" Nov 25 15:36:04 crc kubenswrapper[4890]: I1125 15:36:04.724451 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8394f488-cfd0-4fdc-aa22-96fe01aec73e-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6vbxl\" (UID: \"8394f488-cfd0-4fdc-aa22-96fe01aec73e\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6vbxl" Nov 25 15:36:04 crc kubenswrapper[4890]: I1125 15:36:04.724760 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8394f488-cfd0-4fdc-aa22-96fe01aec73e-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6vbxl\" (UID: \"8394f488-cfd0-4fdc-aa22-96fe01aec73e\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6vbxl" Nov 25 15:36:04 crc kubenswrapper[4890]: I1125 15:36:04.739251 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fmgc\" (UniqueName: \"kubernetes.io/projected/8394f488-cfd0-4fdc-aa22-96fe01aec73e-kube-api-access-2fmgc\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6vbxl\" (UID: \"8394f488-cfd0-4fdc-aa22-96fe01aec73e\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6vbxl" Nov 25 15:36:04 crc kubenswrapper[4890]: I1125 15:36:04.881774 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6vbxl" Nov 25 15:36:05 crc kubenswrapper[4890]: I1125 15:36:05.445659 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6vbxl"] Nov 25 15:36:06 crc kubenswrapper[4890]: I1125 15:36:06.470968 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6vbxl" event={"ID":"8394f488-cfd0-4fdc-aa22-96fe01aec73e","Type":"ContainerStarted","Data":"38f60f05116a81eb9493fe8ffa744a9cc32845fdb569f5e4169ee3561a032ce3"} Nov 25 15:36:06 crc kubenswrapper[4890]: I1125 15:36:06.473560 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6vbxl" event={"ID":"8394f488-cfd0-4fdc-aa22-96fe01aec73e","Type":"ContainerStarted","Data":"645392511cdbe24a8430233ac0fde18ab7df1a3c64023e99e87651ef8e832e53"} Nov 25 15:36:06 crc kubenswrapper[4890]: I1125 15:36:06.492214 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6vbxl" podStartSLOduration=1.880942524 podStartE2EDuration="2.492194265s" podCreationTimestamp="2025-11-25 15:36:04 +0000 UTC" firstStartedPulling="2025-11-25 15:36:05.452319916 +0000 UTC m=+2023.894782526" lastFinishedPulling="2025-11-25 15:36:06.063571657 +0000 UTC m=+2024.506034267" observedRunningTime="2025-11-25 15:36:06.48998156 +0000 UTC m=+2024.932444170" watchObservedRunningTime="2025-11-25 15:36:06.492194265 +0000 UTC m=+2024.934656875" Nov 25 15:36:11 crc kubenswrapper[4890]: I1125 15:36:11.520447 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6vbxl" event={"ID":"8394f488-cfd0-4fdc-aa22-96fe01aec73e","Type":"ContainerDied","Data":"38f60f05116a81eb9493fe8ffa744a9cc32845fdb569f5e4169ee3561a032ce3"} Nov 25 15:36:11 crc kubenswrapper[4890]: I1125 15:36:11.520484 4890 generic.go:334] "Generic (PLEG): container finished" podID="8394f488-cfd0-4fdc-aa22-96fe01aec73e" containerID="38f60f05116a81eb9493fe8ffa744a9cc32845fdb569f5e4169ee3561a032ce3" exitCode=0 Nov 25 15:36:12 crc kubenswrapper[4890]: I1125 15:36:12.954322 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6vbxl" Nov 25 15:36:13 crc kubenswrapper[4890]: I1125 15:36:13.028738 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8394f488-cfd0-4fdc-aa22-96fe01aec73e-ssh-key\") pod \"8394f488-cfd0-4fdc-aa22-96fe01aec73e\" (UID: \"8394f488-cfd0-4fdc-aa22-96fe01aec73e\") " Nov 25 15:36:13 crc kubenswrapper[4890]: I1125 15:36:13.029200 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8394f488-cfd0-4fdc-aa22-96fe01aec73e-inventory\") pod \"8394f488-cfd0-4fdc-aa22-96fe01aec73e\" (UID: \"8394f488-cfd0-4fdc-aa22-96fe01aec73e\") " Nov 25 15:36:13 crc kubenswrapper[4890]: I1125 15:36:13.029351 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2fmgc\" (UniqueName: \"kubernetes.io/projected/8394f488-cfd0-4fdc-aa22-96fe01aec73e-kube-api-access-2fmgc\") pod \"8394f488-cfd0-4fdc-aa22-96fe01aec73e\" (UID: \"8394f488-cfd0-4fdc-aa22-96fe01aec73e\") " Nov 25 15:36:13 crc kubenswrapper[4890]: I1125 15:36:13.035507 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8394f488-cfd0-4fdc-aa22-96fe01aec73e-kube-api-access-2fmgc" (OuterVolumeSpecName: "kube-api-access-2fmgc") pod "8394f488-cfd0-4fdc-aa22-96fe01aec73e" (UID: "8394f488-cfd0-4fdc-aa22-96fe01aec73e"). InnerVolumeSpecName "kube-api-access-2fmgc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:36:13 crc kubenswrapper[4890]: I1125 15:36:13.059493 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8394f488-cfd0-4fdc-aa22-96fe01aec73e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8394f488-cfd0-4fdc-aa22-96fe01aec73e" (UID: "8394f488-cfd0-4fdc-aa22-96fe01aec73e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:36:13 crc kubenswrapper[4890]: I1125 15:36:13.077134 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8394f488-cfd0-4fdc-aa22-96fe01aec73e-inventory" (OuterVolumeSpecName: "inventory") pod "8394f488-cfd0-4fdc-aa22-96fe01aec73e" (UID: "8394f488-cfd0-4fdc-aa22-96fe01aec73e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:36:13 crc kubenswrapper[4890]: I1125 15:36:13.131757 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2fmgc\" (UniqueName: \"kubernetes.io/projected/8394f488-cfd0-4fdc-aa22-96fe01aec73e-kube-api-access-2fmgc\") on node \"crc\" DevicePath \"\"" Nov 25 15:36:13 crc kubenswrapper[4890]: I1125 15:36:13.131789 4890 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8394f488-cfd0-4fdc-aa22-96fe01aec73e-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 15:36:13 crc kubenswrapper[4890]: I1125 15:36:13.131800 4890 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8394f488-cfd0-4fdc-aa22-96fe01aec73e-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 15:36:13 crc kubenswrapper[4890]: I1125 15:36:13.538122 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6vbxl" event={"ID":"8394f488-cfd0-4fdc-aa22-96fe01aec73e","Type":"ContainerDied","Data":"645392511cdbe24a8430233ac0fde18ab7df1a3c64023e99e87651ef8e832e53"} Nov 25 15:36:13 crc kubenswrapper[4890]: I1125 15:36:13.538183 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6vbxl" Nov 25 15:36:13 crc kubenswrapper[4890]: I1125 15:36:13.538188 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="645392511cdbe24a8430233ac0fde18ab7df1a3c64023e99e87651ef8e832e53" Nov 25 15:36:13 crc kubenswrapper[4890]: I1125 15:36:13.614072 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-nn65s"] Nov 25 15:36:13 crc kubenswrapper[4890]: E1125 15:36:13.614616 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8394f488-cfd0-4fdc-aa22-96fe01aec73e" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 25 15:36:13 crc kubenswrapper[4890]: I1125 15:36:13.614641 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="8394f488-cfd0-4fdc-aa22-96fe01aec73e" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 25 15:36:13 crc kubenswrapper[4890]: I1125 15:36:13.614831 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="8394f488-cfd0-4fdc-aa22-96fe01aec73e" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 25 15:36:13 crc kubenswrapper[4890]: I1125 15:36:13.615522 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nn65s" Nov 25 15:36:13 crc kubenswrapper[4890]: I1125 15:36:13.617925 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 15:36:13 crc kubenswrapper[4890]: I1125 15:36:13.619105 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-22zzp" Nov 25 15:36:13 crc kubenswrapper[4890]: I1125 15:36:13.619136 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 15:36:13 crc kubenswrapper[4890]: I1125 15:36:13.619112 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 15:36:13 crc kubenswrapper[4890]: I1125 15:36:13.625262 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-nn65s"] Nov 25 15:36:13 crc kubenswrapper[4890]: I1125 15:36:13.742222 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6679d2f8-52a9-4539-bb89-f4c9c986bc6b-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-nn65s\" (UID: \"6679d2f8-52a9-4539-bb89-f4c9c986bc6b\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nn65s" Nov 25 15:36:13 crc kubenswrapper[4890]: I1125 15:36:13.742605 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rr2qq\" (UniqueName: \"kubernetes.io/projected/6679d2f8-52a9-4539-bb89-f4c9c986bc6b-kube-api-access-rr2qq\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-nn65s\" (UID: \"6679d2f8-52a9-4539-bb89-f4c9c986bc6b\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nn65s" Nov 25 15:36:13 crc kubenswrapper[4890]: I1125 15:36:13.742822 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6679d2f8-52a9-4539-bb89-f4c9c986bc6b-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-nn65s\" (UID: \"6679d2f8-52a9-4539-bb89-f4c9c986bc6b\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nn65s" Nov 25 15:36:13 crc kubenswrapper[4890]: I1125 15:36:13.844268 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6679d2f8-52a9-4539-bb89-f4c9c986bc6b-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-nn65s\" (UID: \"6679d2f8-52a9-4539-bb89-f4c9c986bc6b\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nn65s" Nov 25 15:36:13 crc kubenswrapper[4890]: I1125 15:36:13.844402 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rr2qq\" (UniqueName: \"kubernetes.io/projected/6679d2f8-52a9-4539-bb89-f4c9c986bc6b-kube-api-access-rr2qq\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-nn65s\" (UID: \"6679d2f8-52a9-4539-bb89-f4c9c986bc6b\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nn65s" Nov 25 15:36:13 crc kubenswrapper[4890]: I1125 15:36:13.844453 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6679d2f8-52a9-4539-bb89-f4c9c986bc6b-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-nn65s\" (UID: \"6679d2f8-52a9-4539-bb89-f4c9c986bc6b\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nn65s" Nov 25 15:36:13 crc kubenswrapper[4890]: I1125 15:36:13.848684 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6679d2f8-52a9-4539-bb89-f4c9c986bc6b-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-nn65s\" (UID: \"6679d2f8-52a9-4539-bb89-f4c9c986bc6b\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nn65s" Nov 25 15:36:13 crc kubenswrapper[4890]: I1125 15:36:13.848742 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6679d2f8-52a9-4539-bb89-f4c9c986bc6b-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-nn65s\" (UID: \"6679d2f8-52a9-4539-bb89-f4c9c986bc6b\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nn65s" Nov 25 15:36:13 crc kubenswrapper[4890]: I1125 15:36:13.866019 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rr2qq\" (UniqueName: \"kubernetes.io/projected/6679d2f8-52a9-4539-bb89-f4c9c986bc6b-kube-api-access-rr2qq\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-nn65s\" (UID: \"6679d2f8-52a9-4539-bb89-f4c9c986bc6b\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nn65s" Nov 25 15:36:13 crc kubenswrapper[4890]: I1125 15:36:13.931721 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nn65s" Nov 25 15:36:14 crc kubenswrapper[4890]: I1125 15:36:14.477502 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-nn65s"] Nov 25 15:36:14 crc kubenswrapper[4890]: I1125 15:36:14.548117 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nn65s" event={"ID":"6679d2f8-52a9-4539-bb89-f4c9c986bc6b","Type":"ContainerStarted","Data":"5c069a45383ddb031befc777749069c322a5121339a62f0e6714be5b38cad5fb"} Nov 25 15:36:15 crc kubenswrapper[4890]: I1125 15:36:15.557582 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nn65s" event={"ID":"6679d2f8-52a9-4539-bb89-f4c9c986bc6b","Type":"ContainerStarted","Data":"96f08f62bb021a6b81075b815ff78d37d3c5d433a96b4b3494cadc64c7e35786"} Nov 25 15:36:15 crc kubenswrapper[4890]: I1125 15:36:15.578524 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nn65s" podStartSLOduration=2.030792403 podStartE2EDuration="2.578504666s" podCreationTimestamp="2025-11-25 15:36:13 +0000 UTC" firstStartedPulling="2025-11-25 15:36:14.467809538 +0000 UTC m=+2032.910272148" lastFinishedPulling="2025-11-25 15:36:15.015521801 +0000 UTC m=+2033.457984411" observedRunningTime="2025-11-25 15:36:15.572473466 +0000 UTC m=+2034.014936096" watchObservedRunningTime="2025-11-25 15:36:15.578504666 +0000 UTC m=+2034.020967276" Nov 25 15:36:26 crc kubenswrapper[4890]: I1125 15:36:26.447649 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 15:36:26 crc kubenswrapper[4890]: I1125 15:36:26.448405 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 15:36:26 crc kubenswrapper[4890]: I1125 15:36:26.448468 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" Nov 25 15:36:26 crc kubenswrapper[4890]: I1125 15:36:26.449372 4890 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c2b9ae015432b5c3d0ae89bbacbfeadbc65550341b67aff696270c5f412db829"} pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 15:36:26 crc kubenswrapper[4890]: I1125 15:36:26.449441 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" containerID="cri-o://c2b9ae015432b5c3d0ae89bbacbfeadbc65550341b67aff696270c5f412db829" gracePeriod=600 Nov 25 15:36:26 crc kubenswrapper[4890]: I1125 15:36:26.653207 4890 generic.go:334] "Generic (PLEG): container finished" podID="4e4f849d-f239-4727-a73e-18327856929a" containerID="c2b9ae015432b5c3d0ae89bbacbfeadbc65550341b67aff696270c5f412db829" exitCode=0 Nov 25 15:36:26 crc kubenswrapper[4890]: I1125 15:36:26.653289 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" event={"ID":"4e4f849d-f239-4727-a73e-18327856929a","Type":"ContainerDied","Data":"c2b9ae015432b5c3d0ae89bbacbfeadbc65550341b67aff696270c5f412db829"} Nov 25 15:36:26 crc kubenswrapper[4890]: I1125 15:36:26.653792 4890 scope.go:117] "RemoveContainer" containerID="9467aa8edc4d8bfe461dc02826750091f343018e37e7f269d305c77d21e832da" Nov 25 15:36:27 crc kubenswrapper[4890]: I1125 15:36:27.664928 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" event={"ID":"4e4f849d-f239-4727-a73e-18327856929a","Type":"ContainerStarted","Data":"67ca0de5cd1ef0f73ff70ee2df49e62ca714111ad20a0de1fef88f6f60fdf262"} Nov 25 15:36:39 crc kubenswrapper[4890]: I1125 15:36:39.229093 4890 scope.go:117] "RemoveContainer" containerID="7cbbc2af37308da187efc26ff75f2815686e7168fb9c0b836eee86b68b721c60" Nov 25 15:36:50 crc kubenswrapper[4890]: I1125 15:36:50.561705 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5shjx"] Nov 25 15:36:50 crc kubenswrapper[4890]: I1125 15:36:50.564316 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5shjx" Nov 25 15:36:50 crc kubenswrapper[4890]: I1125 15:36:50.568988 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a00fb2b-1fda-483a-bc58-c3560d90c49a-utilities\") pod \"certified-operators-5shjx\" (UID: \"1a00fb2b-1fda-483a-bc58-c3560d90c49a\") " pod="openshift-marketplace/certified-operators-5shjx" Nov 25 15:36:50 crc kubenswrapper[4890]: I1125 15:36:50.569120 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a00fb2b-1fda-483a-bc58-c3560d90c49a-catalog-content\") pod \"certified-operators-5shjx\" (UID: \"1a00fb2b-1fda-483a-bc58-c3560d90c49a\") " pod="openshift-marketplace/certified-operators-5shjx" Nov 25 15:36:50 crc kubenswrapper[4890]: I1125 15:36:50.569258 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrfh9\" (UniqueName: \"kubernetes.io/projected/1a00fb2b-1fda-483a-bc58-c3560d90c49a-kube-api-access-qrfh9\") pod \"certified-operators-5shjx\" (UID: \"1a00fb2b-1fda-483a-bc58-c3560d90c49a\") " pod="openshift-marketplace/certified-operators-5shjx" Nov 25 15:36:50 crc kubenswrapper[4890]: I1125 15:36:50.577500 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5shjx"] Nov 25 15:36:50 crc kubenswrapper[4890]: I1125 15:36:50.670720 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrfh9\" (UniqueName: \"kubernetes.io/projected/1a00fb2b-1fda-483a-bc58-c3560d90c49a-kube-api-access-qrfh9\") pod \"certified-operators-5shjx\" (UID: \"1a00fb2b-1fda-483a-bc58-c3560d90c49a\") " pod="openshift-marketplace/certified-operators-5shjx" Nov 25 15:36:50 crc kubenswrapper[4890]: I1125 15:36:50.670835 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a00fb2b-1fda-483a-bc58-c3560d90c49a-utilities\") pod \"certified-operators-5shjx\" (UID: \"1a00fb2b-1fda-483a-bc58-c3560d90c49a\") " pod="openshift-marketplace/certified-operators-5shjx" Nov 25 15:36:50 crc kubenswrapper[4890]: I1125 15:36:50.670959 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a00fb2b-1fda-483a-bc58-c3560d90c49a-catalog-content\") pod \"certified-operators-5shjx\" (UID: \"1a00fb2b-1fda-483a-bc58-c3560d90c49a\") " pod="openshift-marketplace/certified-operators-5shjx" Nov 25 15:36:50 crc kubenswrapper[4890]: I1125 15:36:50.671423 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a00fb2b-1fda-483a-bc58-c3560d90c49a-utilities\") pod \"certified-operators-5shjx\" (UID: \"1a00fb2b-1fda-483a-bc58-c3560d90c49a\") " pod="openshift-marketplace/certified-operators-5shjx" Nov 25 15:36:50 crc kubenswrapper[4890]: I1125 15:36:50.671455 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a00fb2b-1fda-483a-bc58-c3560d90c49a-catalog-content\") pod \"certified-operators-5shjx\" (UID: \"1a00fb2b-1fda-483a-bc58-c3560d90c49a\") " pod="openshift-marketplace/certified-operators-5shjx" Nov 25 15:36:50 crc kubenswrapper[4890]: I1125 15:36:50.692454 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrfh9\" (UniqueName: \"kubernetes.io/projected/1a00fb2b-1fda-483a-bc58-c3560d90c49a-kube-api-access-qrfh9\") pod \"certified-operators-5shjx\" (UID: \"1a00fb2b-1fda-483a-bc58-c3560d90c49a\") " pod="openshift-marketplace/certified-operators-5shjx" Nov 25 15:36:50 crc kubenswrapper[4890]: I1125 15:36:50.892881 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5shjx" Nov 25 15:36:50 crc kubenswrapper[4890]: I1125 15:36:50.921896 4890 generic.go:334] "Generic (PLEG): container finished" podID="6679d2f8-52a9-4539-bb89-f4c9c986bc6b" containerID="96f08f62bb021a6b81075b815ff78d37d3c5d433a96b4b3494cadc64c7e35786" exitCode=0 Nov 25 15:36:50 crc kubenswrapper[4890]: I1125 15:36:50.921947 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nn65s" event={"ID":"6679d2f8-52a9-4539-bb89-f4c9c986bc6b","Type":"ContainerDied","Data":"96f08f62bb021a6b81075b815ff78d37d3c5d433a96b4b3494cadc64c7e35786"} Nov 25 15:36:51 crc kubenswrapper[4890]: I1125 15:36:51.401332 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5shjx"] Nov 25 15:36:51 crc kubenswrapper[4890]: I1125 15:36:51.932027 4890 generic.go:334] "Generic (PLEG): container finished" podID="1a00fb2b-1fda-483a-bc58-c3560d90c49a" containerID="497bf48a92dc8f83c8c3adb302f4cdd0d5b2708aaad78ee84c512aed242a4f13" exitCode=0 Nov 25 15:36:51 crc kubenswrapper[4890]: I1125 15:36:51.932071 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5shjx" event={"ID":"1a00fb2b-1fda-483a-bc58-c3560d90c49a","Type":"ContainerDied","Data":"497bf48a92dc8f83c8c3adb302f4cdd0d5b2708aaad78ee84c512aed242a4f13"} Nov 25 15:36:51 crc kubenswrapper[4890]: I1125 15:36:51.932494 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5shjx" event={"ID":"1a00fb2b-1fda-483a-bc58-c3560d90c49a","Type":"ContainerStarted","Data":"138d6b40a941be13af8059a586e7c6f9f8b49cb102b50ee427c2699e0155defe"} Nov 25 15:36:52 crc kubenswrapper[4890]: I1125 15:36:52.402335 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nn65s" Nov 25 15:36:52 crc kubenswrapper[4890]: I1125 15:36:52.512761 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rr2qq\" (UniqueName: \"kubernetes.io/projected/6679d2f8-52a9-4539-bb89-f4c9c986bc6b-kube-api-access-rr2qq\") pod \"6679d2f8-52a9-4539-bb89-f4c9c986bc6b\" (UID: \"6679d2f8-52a9-4539-bb89-f4c9c986bc6b\") " Nov 25 15:36:52 crc kubenswrapper[4890]: I1125 15:36:52.512853 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6679d2f8-52a9-4539-bb89-f4c9c986bc6b-inventory\") pod \"6679d2f8-52a9-4539-bb89-f4c9c986bc6b\" (UID: \"6679d2f8-52a9-4539-bb89-f4c9c986bc6b\") " Nov 25 15:36:52 crc kubenswrapper[4890]: I1125 15:36:52.513148 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6679d2f8-52a9-4539-bb89-f4c9c986bc6b-ssh-key\") pod \"6679d2f8-52a9-4539-bb89-f4c9c986bc6b\" (UID: \"6679d2f8-52a9-4539-bb89-f4c9c986bc6b\") " Nov 25 15:36:52 crc kubenswrapper[4890]: I1125 15:36:52.519028 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6679d2f8-52a9-4539-bb89-f4c9c986bc6b-kube-api-access-rr2qq" (OuterVolumeSpecName: "kube-api-access-rr2qq") pod "6679d2f8-52a9-4539-bb89-f4c9c986bc6b" (UID: "6679d2f8-52a9-4539-bb89-f4c9c986bc6b"). InnerVolumeSpecName "kube-api-access-rr2qq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:36:52 crc kubenswrapper[4890]: I1125 15:36:52.547133 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6679d2f8-52a9-4539-bb89-f4c9c986bc6b-inventory" (OuterVolumeSpecName: "inventory") pod "6679d2f8-52a9-4539-bb89-f4c9c986bc6b" (UID: "6679d2f8-52a9-4539-bb89-f4c9c986bc6b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:36:52 crc kubenswrapper[4890]: I1125 15:36:52.549606 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6679d2f8-52a9-4539-bb89-f4c9c986bc6b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6679d2f8-52a9-4539-bb89-f4c9c986bc6b" (UID: "6679d2f8-52a9-4539-bb89-f4c9c986bc6b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:36:52 crc kubenswrapper[4890]: I1125 15:36:52.615040 4890 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6679d2f8-52a9-4539-bb89-f4c9c986bc6b-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 15:36:52 crc kubenswrapper[4890]: I1125 15:36:52.615089 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rr2qq\" (UniqueName: \"kubernetes.io/projected/6679d2f8-52a9-4539-bb89-f4c9c986bc6b-kube-api-access-rr2qq\") on node \"crc\" DevicePath \"\"" Nov 25 15:36:52 crc kubenswrapper[4890]: I1125 15:36:52.615107 4890 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6679d2f8-52a9-4539-bb89-f4c9c986bc6b-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 15:36:52 crc kubenswrapper[4890]: I1125 15:36:52.942240 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5shjx" event={"ID":"1a00fb2b-1fda-483a-bc58-c3560d90c49a","Type":"ContainerStarted","Data":"c7ae452760d571c6d4635b3db035c749a204bf5371d44ecd400744b08da6ee1f"} Nov 25 15:36:52 crc kubenswrapper[4890]: I1125 15:36:52.944348 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nn65s" event={"ID":"6679d2f8-52a9-4539-bb89-f4c9c986bc6b","Type":"ContainerDied","Data":"5c069a45383ddb031befc777749069c322a5121339a62f0e6714be5b38cad5fb"} Nov 25 15:36:52 crc kubenswrapper[4890]: I1125 15:36:52.944376 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c069a45383ddb031befc777749069c322a5121339a62f0e6714be5b38cad5fb" Nov 25 15:36:52 crc kubenswrapper[4890]: I1125 15:36:52.944423 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nn65s" Nov 25 15:36:53 crc kubenswrapper[4890]: I1125 15:36:53.040557 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wdv5b"] Nov 25 15:36:53 crc kubenswrapper[4890]: E1125 15:36:53.041020 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6679d2f8-52a9-4539-bb89-f4c9c986bc6b" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 25 15:36:53 crc kubenswrapper[4890]: I1125 15:36:53.041041 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="6679d2f8-52a9-4539-bb89-f4c9c986bc6b" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 25 15:36:53 crc kubenswrapper[4890]: I1125 15:36:53.041267 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="6679d2f8-52a9-4539-bb89-f4c9c986bc6b" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 25 15:36:53 crc kubenswrapper[4890]: I1125 15:36:53.042178 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wdv5b" Nov 25 15:36:53 crc kubenswrapper[4890]: I1125 15:36:53.044043 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 15:36:53 crc kubenswrapper[4890]: I1125 15:36:53.045495 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 15:36:53 crc kubenswrapper[4890]: I1125 15:36:53.045788 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 15:36:53 crc kubenswrapper[4890]: I1125 15:36:53.046903 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-22zzp" Nov 25 15:36:53 crc kubenswrapper[4890]: I1125 15:36:53.053653 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wdv5b"] Nov 25 15:36:53 crc kubenswrapper[4890]: I1125 15:36:53.125893 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2875b0db-cd24-42e1-91b7-3308bec9d7d1-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-wdv5b\" (UID: \"2875b0db-cd24-42e1-91b7-3308bec9d7d1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wdv5b" Nov 25 15:36:53 crc kubenswrapper[4890]: I1125 15:36:53.126142 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vh448\" (UniqueName: \"kubernetes.io/projected/2875b0db-cd24-42e1-91b7-3308bec9d7d1-kube-api-access-vh448\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-wdv5b\" (UID: \"2875b0db-cd24-42e1-91b7-3308bec9d7d1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wdv5b" Nov 25 15:36:53 crc kubenswrapper[4890]: I1125 15:36:53.126282 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2875b0db-cd24-42e1-91b7-3308bec9d7d1-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-wdv5b\" (UID: \"2875b0db-cd24-42e1-91b7-3308bec9d7d1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wdv5b" Nov 25 15:36:53 crc kubenswrapper[4890]: I1125 15:36:53.227989 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vh448\" (UniqueName: \"kubernetes.io/projected/2875b0db-cd24-42e1-91b7-3308bec9d7d1-kube-api-access-vh448\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-wdv5b\" (UID: \"2875b0db-cd24-42e1-91b7-3308bec9d7d1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wdv5b" Nov 25 15:36:53 crc kubenswrapper[4890]: I1125 15:36:53.228228 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2875b0db-cd24-42e1-91b7-3308bec9d7d1-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-wdv5b\" (UID: \"2875b0db-cd24-42e1-91b7-3308bec9d7d1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wdv5b" Nov 25 15:36:53 crc kubenswrapper[4890]: I1125 15:36:53.228442 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2875b0db-cd24-42e1-91b7-3308bec9d7d1-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-wdv5b\" (UID: \"2875b0db-cd24-42e1-91b7-3308bec9d7d1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wdv5b" Nov 25 15:36:53 crc kubenswrapper[4890]: I1125 15:36:53.238558 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2875b0db-cd24-42e1-91b7-3308bec9d7d1-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-wdv5b\" (UID: \"2875b0db-cd24-42e1-91b7-3308bec9d7d1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wdv5b" Nov 25 15:36:53 crc kubenswrapper[4890]: I1125 15:36:53.240779 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2875b0db-cd24-42e1-91b7-3308bec9d7d1-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-wdv5b\" (UID: \"2875b0db-cd24-42e1-91b7-3308bec9d7d1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wdv5b" Nov 25 15:36:53 crc kubenswrapper[4890]: I1125 15:36:53.247592 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vh448\" (UniqueName: \"kubernetes.io/projected/2875b0db-cd24-42e1-91b7-3308bec9d7d1-kube-api-access-vh448\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-wdv5b\" (UID: \"2875b0db-cd24-42e1-91b7-3308bec9d7d1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wdv5b" Nov 25 15:36:53 crc kubenswrapper[4890]: I1125 15:36:53.388557 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wdv5b" Nov 25 15:36:53 crc kubenswrapper[4890]: I1125 15:36:53.901897 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wdv5b"] Nov 25 15:36:53 crc kubenswrapper[4890]: I1125 15:36:53.956455 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wdv5b" event={"ID":"2875b0db-cd24-42e1-91b7-3308bec9d7d1","Type":"ContainerStarted","Data":"2bcf13efd08f31294e8c9fced24adac8e1b907ffb18b265e12a419c05b3073fa"} Nov 25 15:36:53 crc kubenswrapper[4890]: I1125 15:36:53.958742 4890 generic.go:334] "Generic (PLEG): container finished" podID="1a00fb2b-1fda-483a-bc58-c3560d90c49a" containerID="c7ae452760d571c6d4635b3db035c749a204bf5371d44ecd400744b08da6ee1f" exitCode=0 Nov 25 15:36:53 crc kubenswrapper[4890]: I1125 15:36:53.958813 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5shjx" event={"ID":"1a00fb2b-1fda-483a-bc58-c3560d90c49a","Type":"ContainerDied","Data":"c7ae452760d571c6d4635b3db035c749a204bf5371d44ecd400744b08da6ee1f"} Nov 25 15:36:54 crc kubenswrapper[4890]: I1125 15:36:54.972334 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wdv5b" event={"ID":"2875b0db-cd24-42e1-91b7-3308bec9d7d1","Type":"ContainerStarted","Data":"06654759287eb44061150db6c536c5d7e1dc7d2c94e239e7c8b2bceea4280278"} Nov 25 15:36:55 crc kubenswrapper[4890]: I1125 15:36:55.000586 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wdv5b" podStartSLOduration=1.339564719 podStartE2EDuration="2.000566223s" podCreationTimestamp="2025-11-25 15:36:53 +0000 UTC" firstStartedPulling="2025-11-25 15:36:53.919525175 +0000 UTC m=+2072.361987785" lastFinishedPulling="2025-11-25 15:36:54.580526679 +0000 UTC m=+2073.022989289" observedRunningTime="2025-11-25 15:36:54.991302002 +0000 UTC m=+2073.433764632" watchObservedRunningTime="2025-11-25 15:36:55.000566223 +0000 UTC m=+2073.443028823" Nov 25 15:36:55 crc kubenswrapper[4890]: I1125 15:36:55.990216 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5shjx" event={"ID":"1a00fb2b-1fda-483a-bc58-c3560d90c49a","Type":"ContainerStarted","Data":"6dbfdc39dbeb6856cfc5e7852275adc72d1ad260c1efb782fb620b7a838452d7"} Nov 25 15:36:56 crc kubenswrapper[4890]: I1125 15:36:56.022783 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5shjx" podStartSLOduration=2.601122028 podStartE2EDuration="6.02275829s" podCreationTimestamp="2025-11-25 15:36:50 +0000 UTC" firstStartedPulling="2025-11-25 15:36:51.934765261 +0000 UTC m=+2070.377227871" lastFinishedPulling="2025-11-25 15:36:55.356401523 +0000 UTC m=+2073.798864133" observedRunningTime="2025-11-25 15:36:56.012194356 +0000 UTC m=+2074.454656966" watchObservedRunningTime="2025-11-25 15:36:56.02275829 +0000 UTC m=+2074.465220900" Nov 25 15:37:00 crc kubenswrapper[4890]: I1125 15:37:00.894039 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5shjx" Nov 25 15:37:00 crc kubenswrapper[4890]: I1125 15:37:00.894703 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5shjx" Nov 25 15:37:00 crc kubenswrapper[4890]: I1125 15:37:00.967785 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5shjx" Nov 25 15:37:01 crc kubenswrapper[4890]: I1125 15:37:01.078495 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5shjx" Nov 25 15:37:01 crc kubenswrapper[4890]: I1125 15:37:01.209620 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5shjx"] Nov 25 15:37:03 crc kubenswrapper[4890]: I1125 15:37:03.049505 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5shjx" podUID="1a00fb2b-1fda-483a-bc58-c3560d90c49a" containerName="registry-server" containerID="cri-o://6dbfdc39dbeb6856cfc5e7852275adc72d1ad260c1efb782fb620b7a838452d7" gracePeriod=2 Nov 25 15:37:05 crc kubenswrapper[4890]: I1125 15:37:05.069376 4890 generic.go:334] "Generic (PLEG): container finished" podID="1a00fb2b-1fda-483a-bc58-c3560d90c49a" containerID="6dbfdc39dbeb6856cfc5e7852275adc72d1ad260c1efb782fb620b7a838452d7" exitCode=0 Nov 25 15:37:05 crc kubenswrapper[4890]: I1125 15:37:05.069424 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5shjx" event={"ID":"1a00fb2b-1fda-483a-bc58-c3560d90c49a","Type":"ContainerDied","Data":"6dbfdc39dbeb6856cfc5e7852275adc72d1ad260c1efb782fb620b7a838452d7"} Nov 25 15:37:05 crc kubenswrapper[4890]: I1125 15:37:05.752010 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5shjx" Nov 25 15:37:05 crc kubenswrapper[4890]: I1125 15:37:05.890237 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a00fb2b-1fda-483a-bc58-c3560d90c49a-utilities\") pod \"1a00fb2b-1fda-483a-bc58-c3560d90c49a\" (UID: \"1a00fb2b-1fda-483a-bc58-c3560d90c49a\") " Nov 25 15:37:05 crc kubenswrapper[4890]: I1125 15:37:05.890401 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrfh9\" (UniqueName: \"kubernetes.io/projected/1a00fb2b-1fda-483a-bc58-c3560d90c49a-kube-api-access-qrfh9\") pod \"1a00fb2b-1fda-483a-bc58-c3560d90c49a\" (UID: \"1a00fb2b-1fda-483a-bc58-c3560d90c49a\") " Nov 25 15:37:05 crc kubenswrapper[4890]: I1125 15:37:05.890458 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a00fb2b-1fda-483a-bc58-c3560d90c49a-catalog-content\") pod \"1a00fb2b-1fda-483a-bc58-c3560d90c49a\" (UID: \"1a00fb2b-1fda-483a-bc58-c3560d90c49a\") " Nov 25 15:37:05 crc kubenswrapper[4890]: I1125 15:37:05.891963 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a00fb2b-1fda-483a-bc58-c3560d90c49a-utilities" (OuterVolumeSpecName: "utilities") pod "1a00fb2b-1fda-483a-bc58-c3560d90c49a" (UID: "1a00fb2b-1fda-483a-bc58-c3560d90c49a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:37:05 crc kubenswrapper[4890]: I1125 15:37:05.905365 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a00fb2b-1fda-483a-bc58-c3560d90c49a-kube-api-access-qrfh9" (OuterVolumeSpecName: "kube-api-access-qrfh9") pod "1a00fb2b-1fda-483a-bc58-c3560d90c49a" (UID: "1a00fb2b-1fda-483a-bc58-c3560d90c49a"). InnerVolumeSpecName "kube-api-access-qrfh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:37:05 crc kubenswrapper[4890]: I1125 15:37:05.948340 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a00fb2b-1fda-483a-bc58-c3560d90c49a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1a00fb2b-1fda-483a-bc58-c3560d90c49a" (UID: "1a00fb2b-1fda-483a-bc58-c3560d90c49a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:37:05 crc kubenswrapper[4890]: I1125 15:37:05.993991 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a00fb2b-1fda-483a-bc58-c3560d90c49a-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 15:37:05 crc kubenswrapper[4890]: I1125 15:37:05.994025 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qrfh9\" (UniqueName: \"kubernetes.io/projected/1a00fb2b-1fda-483a-bc58-c3560d90c49a-kube-api-access-qrfh9\") on node \"crc\" DevicePath \"\"" Nov 25 15:37:05 crc kubenswrapper[4890]: I1125 15:37:05.994046 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a00fb2b-1fda-483a-bc58-c3560d90c49a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 15:37:06 crc kubenswrapper[4890]: I1125 15:37:06.083217 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5shjx" event={"ID":"1a00fb2b-1fda-483a-bc58-c3560d90c49a","Type":"ContainerDied","Data":"138d6b40a941be13af8059a586e7c6f9f8b49cb102b50ee427c2699e0155defe"} Nov 25 15:37:06 crc kubenswrapper[4890]: I1125 15:37:06.083283 4890 scope.go:117] "RemoveContainer" containerID="6dbfdc39dbeb6856cfc5e7852275adc72d1ad260c1efb782fb620b7a838452d7" Nov 25 15:37:06 crc kubenswrapper[4890]: I1125 15:37:06.084448 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5shjx" Nov 25 15:37:06 crc kubenswrapper[4890]: I1125 15:37:06.121463 4890 scope.go:117] "RemoveContainer" containerID="c7ae452760d571c6d4635b3db035c749a204bf5371d44ecd400744b08da6ee1f" Nov 25 15:37:06 crc kubenswrapper[4890]: I1125 15:37:06.133987 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5shjx"] Nov 25 15:37:06 crc kubenswrapper[4890]: I1125 15:37:06.148698 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5shjx"] Nov 25 15:37:06 crc kubenswrapper[4890]: I1125 15:37:06.149346 4890 scope.go:117] "RemoveContainer" containerID="497bf48a92dc8f83c8c3adb302f4cdd0d5b2708aaad78ee84c512aed242a4f13" Nov 25 15:37:06 crc kubenswrapper[4890]: I1125 15:37:06.200991 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a00fb2b-1fda-483a-bc58-c3560d90c49a" path="/var/lib/kubelet/pods/1a00fb2b-1fda-483a-bc58-c3560d90c49a/volumes" Nov 25 15:37:41 crc kubenswrapper[4890]: I1125 15:37:41.451309 4890 generic.go:334] "Generic (PLEG): container finished" podID="2875b0db-cd24-42e1-91b7-3308bec9d7d1" containerID="06654759287eb44061150db6c536c5d7e1dc7d2c94e239e7c8b2bceea4280278" exitCode=0 Nov 25 15:37:41 crc kubenswrapper[4890]: I1125 15:37:41.451407 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wdv5b" event={"ID":"2875b0db-cd24-42e1-91b7-3308bec9d7d1","Type":"ContainerDied","Data":"06654759287eb44061150db6c536c5d7e1dc7d2c94e239e7c8b2bceea4280278"} Nov 25 15:37:42 crc kubenswrapper[4890]: I1125 15:37:42.917800 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wdv5b" Nov 25 15:37:43 crc kubenswrapper[4890]: I1125 15:37:43.004836 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vh448\" (UniqueName: \"kubernetes.io/projected/2875b0db-cd24-42e1-91b7-3308bec9d7d1-kube-api-access-vh448\") pod \"2875b0db-cd24-42e1-91b7-3308bec9d7d1\" (UID: \"2875b0db-cd24-42e1-91b7-3308bec9d7d1\") " Nov 25 15:37:43 crc kubenswrapper[4890]: I1125 15:37:43.004892 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2875b0db-cd24-42e1-91b7-3308bec9d7d1-inventory\") pod \"2875b0db-cd24-42e1-91b7-3308bec9d7d1\" (UID: \"2875b0db-cd24-42e1-91b7-3308bec9d7d1\") " Nov 25 15:37:43 crc kubenswrapper[4890]: I1125 15:37:43.005032 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2875b0db-cd24-42e1-91b7-3308bec9d7d1-ssh-key\") pod \"2875b0db-cd24-42e1-91b7-3308bec9d7d1\" (UID: \"2875b0db-cd24-42e1-91b7-3308bec9d7d1\") " Nov 25 15:37:43 crc kubenswrapper[4890]: I1125 15:37:43.013393 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2875b0db-cd24-42e1-91b7-3308bec9d7d1-kube-api-access-vh448" (OuterVolumeSpecName: "kube-api-access-vh448") pod "2875b0db-cd24-42e1-91b7-3308bec9d7d1" (UID: "2875b0db-cd24-42e1-91b7-3308bec9d7d1"). InnerVolumeSpecName "kube-api-access-vh448". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:37:43 crc kubenswrapper[4890]: I1125 15:37:43.036028 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2875b0db-cd24-42e1-91b7-3308bec9d7d1-inventory" (OuterVolumeSpecName: "inventory") pod "2875b0db-cd24-42e1-91b7-3308bec9d7d1" (UID: "2875b0db-cd24-42e1-91b7-3308bec9d7d1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:37:43 crc kubenswrapper[4890]: I1125 15:37:43.040886 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2875b0db-cd24-42e1-91b7-3308bec9d7d1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2875b0db-cd24-42e1-91b7-3308bec9d7d1" (UID: "2875b0db-cd24-42e1-91b7-3308bec9d7d1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:37:43 crc kubenswrapper[4890]: I1125 15:37:43.108716 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vh448\" (UniqueName: \"kubernetes.io/projected/2875b0db-cd24-42e1-91b7-3308bec9d7d1-kube-api-access-vh448\") on node \"crc\" DevicePath \"\"" Nov 25 15:37:43 crc kubenswrapper[4890]: I1125 15:37:43.109238 4890 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2875b0db-cd24-42e1-91b7-3308bec9d7d1-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 15:37:43 crc kubenswrapper[4890]: I1125 15:37:43.109252 4890 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2875b0db-cd24-42e1-91b7-3308bec9d7d1-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 15:37:43 crc kubenswrapper[4890]: I1125 15:37:43.474506 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wdv5b" event={"ID":"2875b0db-cd24-42e1-91b7-3308bec9d7d1","Type":"ContainerDied","Data":"2bcf13efd08f31294e8c9fced24adac8e1b907ffb18b265e12a419c05b3073fa"} Nov 25 15:37:43 crc kubenswrapper[4890]: I1125 15:37:43.474545 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wdv5b" Nov 25 15:37:43 crc kubenswrapper[4890]: I1125 15:37:43.474559 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2bcf13efd08f31294e8c9fced24adac8e1b907ffb18b265e12a419c05b3073fa" Nov 25 15:37:43 crc kubenswrapper[4890]: I1125 15:37:43.606245 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-59sbs"] Nov 25 15:37:43 crc kubenswrapper[4890]: E1125 15:37:43.606763 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a00fb2b-1fda-483a-bc58-c3560d90c49a" containerName="extract-content" Nov 25 15:37:43 crc kubenswrapper[4890]: I1125 15:37:43.606785 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a00fb2b-1fda-483a-bc58-c3560d90c49a" containerName="extract-content" Nov 25 15:37:43 crc kubenswrapper[4890]: E1125 15:37:43.606812 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2875b0db-cd24-42e1-91b7-3308bec9d7d1" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 25 15:37:43 crc kubenswrapper[4890]: I1125 15:37:43.606822 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="2875b0db-cd24-42e1-91b7-3308bec9d7d1" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 25 15:37:43 crc kubenswrapper[4890]: E1125 15:37:43.606859 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a00fb2b-1fda-483a-bc58-c3560d90c49a" containerName="registry-server" Nov 25 15:37:43 crc kubenswrapper[4890]: I1125 15:37:43.606866 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a00fb2b-1fda-483a-bc58-c3560d90c49a" containerName="registry-server" Nov 25 15:37:43 crc kubenswrapper[4890]: E1125 15:37:43.606880 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a00fb2b-1fda-483a-bc58-c3560d90c49a" containerName="extract-utilities" Nov 25 15:37:43 crc kubenswrapper[4890]: I1125 15:37:43.606888 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a00fb2b-1fda-483a-bc58-c3560d90c49a" containerName="extract-utilities" Nov 25 15:37:43 crc kubenswrapper[4890]: I1125 15:37:43.607270 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="2875b0db-cd24-42e1-91b7-3308bec9d7d1" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 25 15:37:43 crc kubenswrapper[4890]: I1125 15:37:43.607297 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a00fb2b-1fda-483a-bc58-c3560d90c49a" containerName="registry-server" Nov 25 15:37:43 crc kubenswrapper[4890]: I1125 15:37:43.608221 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-59sbs" Nov 25 15:37:43 crc kubenswrapper[4890]: I1125 15:37:43.612390 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 15:37:43 crc kubenswrapper[4890]: I1125 15:37:43.612468 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 15:37:43 crc kubenswrapper[4890]: I1125 15:37:43.612468 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-22zzp" Nov 25 15:37:43 crc kubenswrapper[4890]: I1125 15:37:43.612702 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 15:37:43 crc kubenswrapper[4890]: I1125 15:37:43.630504 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtzwp\" (UniqueName: \"kubernetes.io/projected/788b5bd1-adf2-4416-a8f0-975895de101c-kube-api-access-wtzwp\") pod \"ssh-known-hosts-edpm-deployment-59sbs\" (UID: \"788b5bd1-adf2-4416-a8f0-975895de101c\") " pod="openstack/ssh-known-hosts-edpm-deployment-59sbs" Nov 25 15:37:43 crc kubenswrapper[4890]: I1125 15:37:43.630794 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/788b5bd1-adf2-4416-a8f0-975895de101c-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-59sbs\" (UID: \"788b5bd1-adf2-4416-a8f0-975895de101c\") " pod="openstack/ssh-known-hosts-edpm-deployment-59sbs" Nov 25 15:37:43 crc kubenswrapper[4890]: I1125 15:37:43.630960 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/788b5bd1-adf2-4416-a8f0-975895de101c-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-59sbs\" (UID: \"788b5bd1-adf2-4416-a8f0-975895de101c\") " pod="openstack/ssh-known-hosts-edpm-deployment-59sbs" Nov 25 15:37:43 crc kubenswrapper[4890]: I1125 15:37:43.631513 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-59sbs"] Nov 25 15:37:43 crc kubenswrapper[4890]: I1125 15:37:43.733489 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/788b5bd1-adf2-4416-a8f0-975895de101c-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-59sbs\" (UID: \"788b5bd1-adf2-4416-a8f0-975895de101c\") " pod="openstack/ssh-known-hosts-edpm-deployment-59sbs" Nov 25 15:37:43 crc kubenswrapper[4890]: I1125 15:37:43.733610 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/788b5bd1-adf2-4416-a8f0-975895de101c-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-59sbs\" (UID: \"788b5bd1-adf2-4416-a8f0-975895de101c\") " pod="openstack/ssh-known-hosts-edpm-deployment-59sbs" Nov 25 15:37:43 crc kubenswrapper[4890]: I1125 15:37:43.733702 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtzwp\" (UniqueName: \"kubernetes.io/projected/788b5bd1-adf2-4416-a8f0-975895de101c-kube-api-access-wtzwp\") pod \"ssh-known-hosts-edpm-deployment-59sbs\" (UID: \"788b5bd1-adf2-4416-a8f0-975895de101c\") " pod="openstack/ssh-known-hosts-edpm-deployment-59sbs" Nov 25 15:37:43 crc kubenswrapper[4890]: I1125 15:37:43.738018 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/788b5bd1-adf2-4416-a8f0-975895de101c-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-59sbs\" (UID: \"788b5bd1-adf2-4416-a8f0-975895de101c\") " pod="openstack/ssh-known-hosts-edpm-deployment-59sbs" Nov 25 15:37:43 crc kubenswrapper[4890]: I1125 15:37:43.748805 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/788b5bd1-adf2-4416-a8f0-975895de101c-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-59sbs\" (UID: \"788b5bd1-adf2-4416-a8f0-975895de101c\") " pod="openstack/ssh-known-hosts-edpm-deployment-59sbs" Nov 25 15:37:43 crc kubenswrapper[4890]: I1125 15:37:43.771962 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtzwp\" (UniqueName: \"kubernetes.io/projected/788b5bd1-adf2-4416-a8f0-975895de101c-kube-api-access-wtzwp\") pod \"ssh-known-hosts-edpm-deployment-59sbs\" (UID: \"788b5bd1-adf2-4416-a8f0-975895de101c\") " pod="openstack/ssh-known-hosts-edpm-deployment-59sbs" Nov 25 15:37:43 crc kubenswrapper[4890]: I1125 15:37:43.989568 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-59sbs" Nov 25 15:37:44 crc kubenswrapper[4890]: I1125 15:37:44.496228 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-59sbs"] Nov 25 15:37:45 crc kubenswrapper[4890]: I1125 15:37:45.494078 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-59sbs" event={"ID":"788b5bd1-adf2-4416-a8f0-975895de101c","Type":"ContainerStarted","Data":"4738f14404d006916d85a46a323def2300aea9e158c9a4bd2125f1efa952f395"} Nov 25 15:37:45 crc kubenswrapper[4890]: I1125 15:37:45.494552 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-59sbs" event={"ID":"788b5bd1-adf2-4416-a8f0-975895de101c","Type":"ContainerStarted","Data":"f645ed1dd317cedfe242de4a4f3eb91e45a0e5bd38c0e9571cc7c06cd4369f1a"} Nov 25 15:37:45 crc kubenswrapper[4890]: I1125 15:37:45.521896 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-59sbs" podStartSLOduration=1.916642823 podStartE2EDuration="2.521866654s" podCreationTimestamp="2025-11-25 15:37:43 +0000 UTC" firstStartedPulling="2025-11-25 15:37:44.501542233 +0000 UTC m=+2122.944004843" lastFinishedPulling="2025-11-25 15:37:45.106766064 +0000 UTC m=+2123.549228674" observedRunningTime="2025-11-25 15:37:45.517938156 +0000 UTC m=+2123.960400776" watchObservedRunningTime="2025-11-25 15:37:45.521866654 +0000 UTC m=+2123.964329284" Nov 25 15:37:52 crc kubenswrapper[4890]: I1125 15:37:52.558577 4890 generic.go:334] "Generic (PLEG): container finished" podID="788b5bd1-adf2-4416-a8f0-975895de101c" containerID="4738f14404d006916d85a46a323def2300aea9e158c9a4bd2125f1efa952f395" exitCode=0 Nov 25 15:37:52 crc kubenswrapper[4890]: I1125 15:37:52.558665 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-59sbs" event={"ID":"788b5bd1-adf2-4416-a8f0-975895de101c","Type":"ContainerDied","Data":"4738f14404d006916d85a46a323def2300aea9e158c9a4bd2125f1efa952f395"} Nov 25 15:37:54 crc kubenswrapper[4890]: I1125 15:37:54.022866 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-59sbs" Nov 25 15:37:54 crc kubenswrapper[4890]: I1125 15:37:54.049395 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/788b5bd1-adf2-4416-a8f0-975895de101c-ssh-key-openstack-edpm-ipam\") pod \"788b5bd1-adf2-4416-a8f0-975895de101c\" (UID: \"788b5bd1-adf2-4416-a8f0-975895de101c\") " Nov 25 15:37:54 crc kubenswrapper[4890]: I1125 15:37:54.084829 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/788b5bd1-adf2-4416-a8f0-975895de101c-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "788b5bd1-adf2-4416-a8f0-975895de101c" (UID: "788b5bd1-adf2-4416-a8f0-975895de101c"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:37:54 crc kubenswrapper[4890]: I1125 15:37:54.151811 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wtzwp\" (UniqueName: \"kubernetes.io/projected/788b5bd1-adf2-4416-a8f0-975895de101c-kube-api-access-wtzwp\") pod \"788b5bd1-adf2-4416-a8f0-975895de101c\" (UID: \"788b5bd1-adf2-4416-a8f0-975895de101c\") " Nov 25 15:37:54 crc kubenswrapper[4890]: I1125 15:37:54.151998 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/788b5bd1-adf2-4416-a8f0-975895de101c-inventory-0\") pod \"788b5bd1-adf2-4416-a8f0-975895de101c\" (UID: \"788b5bd1-adf2-4416-a8f0-975895de101c\") " Nov 25 15:37:54 crc kubenswrapper[4890]: I1125 15:37:54.152932 4890 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/788b5bd1-adf2-4416-a8f0-975895de101c-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 25 15:37:54 crc kubenswrapper[4890]: I1125 15:37:54.156931 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/788b5bd1-adf2-4416-a8f0-975895de101c-kube-api-access-wtzwp" (OuterVolumeSpecName: "kube-api-access-wtzwp") pod "788b5bd1-adf2-4416-a8f0-975895de101c" (UID: "788b5bd1-adf2-4416-a8f0-975895de101c"). InnerVolumeSpecName "kube-api-access-wtzwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:37:54 crc kubenswrapper[4890]: I1125 15:37:54.192826 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/788b5bd1-adf2-4416-a8f0-975895de101c-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "788b5bd1-adf2-4416-a8f0-975895de101c" (UID: "788b5bd1-adf2-4416-a8f0-975895de101c"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:37:54 crc kubenswrapper[4890]: I1125 15:37:54.253664 4890 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/788b5bd1-adf2-4416-a8f0-975895de101c-inventory-0\") on node \"crc\" DevicePath \"\"" Nov 25 15:37:54 crc kubenswrapper[4890]: I1125 15:37:54.253702 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wtzwp\" (UniqueName: \"kubernetes.io/projected/788b5bd1-adf2-4416-a8f0-975895de101c-kube-api-access-wtzwp\") on node \"crc\" DevicePath \"\"" Nov 25 15:37:54 crc kubenswrapper[4890]: I1125 15:37:54.580695 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-59sbs" event={"ID":"788b5bd1-adf2-4416-a8f0-975895de101c","Type":"ContainerDied","Data":"f645ed1dd317cedfe242de4a4f3eb91e45a0e5bd38c0e9571cc7c06cd4369f1a"} Nov 25 15:37:54 crc kubenswrapper[4890]: I1125 15:37:54.580745 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f645ed1dd317cedfe242de4a4f3eb91e45a0e5bd38c0e9571cc7c06cd4369f1a" Nov 25 15:37:54 crc kubenswrapper[4890]: I1125 15:37:54.580813 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-59sbs" Nov 25 15:37:54 crc kubenswrapper[4890]: I1125 15:37:54.682361 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-l94p8"] Nov 25 15:37:54 crc kubenswrapper[4890]: E1125 15:37:54.683575 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="788b5bd1-adf2-4416-a8f0-975895de101c" containerName="ssh-known-hosts-edpm-deployment" Nov 25 15:37:54 crc kubenswrapper[4890]: I1125 15:37:54.683602 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="788b5bd1-adf2-4416-a8f0-975895de101c" containerName="ssh-known-hosts-edpm-deployment" Nov 25 15:37:54 crc kubenswrapper[4890]: I1125 15:37:54.683942 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="788b5bd1-adf2-4416-a8f0-975895de101c" containerName="ssh-known-hosts-edpm-deployment" Nov 25 15:37:54 crc kubenswrapper[4890]: I1125 15:37:54.685089 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-l94p8" Nov 25 15:37:54 crc kubenswrapper[4890]: I1125 15:37:54.689870 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-22zzp" Nov 25 15:37:54 crc kubenswrapper[4890]: I1125 15:37:54.690129 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 15:37:54 crc kubenswrapper[4890]: I1125 15:37:54.690352 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 15:37:54 crc kubenswrapper[4890]: I1125 15:37:54.693022 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 15:37:54 crc kubenswrapper[4890]: I1125 15:37:54.695375 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-l94p8"] Nov 25 15:37:54 crc kubenswrapper[4890]: I1125 15:37:54.763337 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/90b03f38-bd28-4c62-ae05-5cc72f4e6ee7-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-l94p8\" (UID: \"90b03f38-bd28-4c62-ae05-5cc72f4e6ee7\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-l94p8" Nov 25 15:37:54 crc kubenswrapper[4890]: I1125 15:37:54.763409 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/90b03f38-bd28-4c62-ae05-5cc72f4e6ee7-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-l94p8\" (UID: \"90b03f38-bd28-4c62-ae05-5cc72f4e6ee7\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-l94p8" Nov 25 15:37:54 crc kubenswrapper[4890]: I1125 15:37:54.763526 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4ksz\" (UniqueName: \"kubernetes.io/projected/90b03f38-bd28-4c62-ae05-5cc72f4e6ee7-kube-api-access-k4ksz\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-l94p8\" (UID: \"90b03f38-bd28-4c62-ae05-5cc72f4e6ee7\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-l94p8" Nov 25 15:37:54 crc kubenswrapper[4890]: I1125 15:37:54.866435 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/90b03f38-bd28-4c62-ae05-5cc72f4e6ee7-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-l94p8\" (UID: \"90b03f38-bd28-4c62-ae05-5cc72f4e6ee7\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-l94p8" Nov 25 15:37:54 crc kubenswrapper[4890]: I1125 15:37:54.866552 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/90b03f38-bd28-4c62-ae05-5cc72f4e6ee7-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-l94p8\" (UID: \"90b03f38-bd28-4c62-ae05-5cc72f4e6ee7\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-l94p8" Nov 25 15:37:54 crc kubenswrapper[4890]: I1125 15:37:54.866632 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4ksz\" (UniqueName: \"kubernetes.io/projected/90b03f38-bd28-4c62-ae05-5cc72f4e6ee7-kube-api-access-k4ksz\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-l94p8\" (UID: \"90b03f38-bd28-4c62-ae05-5cc72f4e6ee7\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-l94p8" Nov 25 15:37:54 crc kubenswrapper[4890]: I1125 15:37:54.872458 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/90b03f38-bd28-4c62-ae05-5cc72f4e6ee7-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-l94p8\" (UID: \"90b03f38-bd28-4c62-ae05-5cc72f4e6ee7\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-l94p8" Nov 25 15:37:54 crc kubenswrapper[4890]: I1125 15:37:54.876991 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/90b03f38-bd28-4c62-ae05-5cc72f4e6ee7-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-l94p8\" (UID: \"90b03f38-bd28-4c62-ae05-5cc72f4e6ee7\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-l94p8" Nov 25 15:37:54 crc kubenswrapper[4890]: I1125 15:37:54.883481 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4ksz\" (UniqueName: \"kubernetes.io/projected/90b03f38-bd28-4c62-ae05-5cc72f4e6ee7-kube-api-access-k4ksz\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-l94p8\" (UID: \"90b03f38-bd28-4c62-ae05-5cc72f4e6ee7\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-l94p8" Nov 25 15:37:55 crc kubenswrapper[4890]: I1125 15:37:55.006235 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-l94p8" Nov 25 15:37:55 crc kubenswrapper[4890]: I1125 15:37:55.543061 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-l94p8"] Nov 25 15:37:55 crc kubenswrapper[4890]: I1125 15:37:55.593011 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-l94p8" event={"ID":"90b03f38-bd28-4c62-ae05-5cc72f4e6ee7","Type":"ContainerStarted","Data":"baa5396137e8e803340bc86db0ea548ec26d125d2c182e6149776eef5944bde8"} Nov 25 15:37:57 crc kubenswrapper[4890]: I1125 15:37:57.615505 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-l94p8" event={"ID":"90b03f38-bd28-4c62-ae05-5cc72f4e6ee7","Type":"ContainerStarted","Data":"5504c827284c6d0b3433f077e75e54157c0c31eed93dbb9efa4f773aa593ad06"} Nov 25 15:37:57 crc kubenswrapper[4890]: I1125 15:37:57.644944 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-l94p8" podStartSLOduration=2.824729701 podStartE2EDuration="3.644915111s" podCreationTimestamp="2025-11-25 15:37:54 +0000 UTC" firstStartedPulling="2025-11-25 15:37:55.547516463 +0000 UTC m=+2133.989979073" lastFinishedPulling="2025-11-25 15:37:56.367701873 +0000 UTC m=+2134.810164483" observedRunningTime="2025-11-25 15:37:57.633110866 +0000 UTC m=+2136.075573496" watchObservedRunningTime="2025-11-25 15:37:57.644915111 +0000 UTC m=+2136.087377721" Nov 25 15:38:04 crc kubenswrapper[4890]: I1125 15:38:04.687778 4890 generic.go:334] "Generic (PLEG): container finished" podID="90b03f38-bd28-4c62-ae05-5cc72f4e6ee7" containerID="5504c827284c6d0b3433f077e75e54157c0c31eed93dbb9efa4f773aa593ad06" exitCode=0 Nov 25 15:38:04 crc kubenswrapper[4890]: I1125 15:38:04.687889 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-l94p8" event={"ID":"90b03f38-bd28-4c62-ae05-5cc72f4e6ee7","Type":"ContainerDied","Data":"5504c827284c6d0b3433f077e75e54157c0c31eed93dbb9efa4f773aa593ad06"} Nov 25 15:38:06 crc kubenswrapper[4890]: I1125 15:38:06.126067 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-l94p8" Nov 25 15:38:06 crc kubenswrapper[4890]: I1125 15:38:06.300264 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/90b03f38-bd28-4c62-ae05-5cc72f4e6ee7-inventory\") pod \"90b03f38-bd28-4c62-ae05-5cc72f4e6ee7\" (UID: \"90b03f38-bd28-4c62-ae05-5cc72f4e6ee7\") " Nov 25 15:38:06 crc kubenswrapper[4890]: I1125 15:38:06.300553 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/90b03f38-bd28-4c62-ae05-5cc72f4e6ee7-ssh-key\") pod \"90b03f38-bd28-4c62-ae05-5cc72f4e6ee7\" (UID: \"90b03f38-bd28-4c62-ae05-5cc72f4e6ee7\") " Nov 25 15:38:06 crc kubenswrapper[4890]: I1125 15:38:06.300617 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k4ksz\" (UniqueName: \"kubernetes.io/projected/90b03f38-bd28-4c62-ae05-5cc72f4e6ee7-kube-api-access-k4ksz\") pod \"90b03f38-bd28-4c62-ae05-5cc72f4e6ee7\" (UID: \"90b03f38-bd28-4c62-ae05-5cc72f4e6ee7\") " Nov 25 15:38:06 crc kubenswrapper[4890]: I1125 15:38:06.306766 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90b03f38-bd28-4c62-ae05-5cc72f4e6ee7-kube-api-access-k4ksz" (OuterVolumeSpecName: "kube-api-access-k4ksz") pod "90b03f38-bd28-4c62-ae05-5cc72f4e6ee7" (UID: "90b03f38-bd28-4c62-ae05-5cc72f4e6ee7"). InnerVolumeSpecName "kube-api-access-k4ksz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:38:06 crc kubenswrapper[4890]: I1125 15:38:06.333393 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90b03f38-bd28-4c62-ae05-5cc72f4e6ee7-inventory" (OuterVolumeSpecName: "inventory") pod "90b03f38-bd28-4c62-ae05-5cc72f4e6ee7" (UID: "90b03f38-bd28-4c62-ae05-5cc72f4e6ee7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:38:06 crc kubenswrapper[4890]: I1125 15:38:06.333829 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90b03f38-bd28-4c62-ae05-5cc72f4e6ee7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "90b03f38-bd28-4c62-ae05-5cc72f4e6ee7" (UID: "90b03f38-bd28-4c62-ae05-5cc72f4e6ee7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:38:06 crc kubenswrapper[4890]: I1125 15:38:06.403462 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k4ksz\" (UniqueName: \"kubernetes.io/projected/90b03f38-bd28-4c62-ae05-5cc72f4e6ee7-kube-api-access-k4ksz\") on node \"crc\" DevicePath \"\"" Nov 25 15:38:06 crc kubenswrapper[4890]: I1125 15:38:06.403551 4890 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/90b03f38-bd28-4c62-ae05-5cc72f4e6ee7-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 15:38:06 crc kubenswrapper[4890]: I1125 15:38:06.403566 4890 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/90b03f38-bd28-4c62-ae05-5cc72f4e6ee7-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 15:38:06 crc kubenswrapper[4890]: I1125 15:38:06.712954 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-l94p8" event={"ID":"90b03f38-bd28-4c62-ae05-5cc72f4e6ee7","Type":"ContainerDied","Data":"baa5396137e8e803340bc86db0ea548ec26d125d2c182e6149776eef5944bde8"} Nov 25 15:38:06 crc kubenswrapper[4890]: I1125 15:38:06.713000 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="baa5396137e8e803340bc86db0ea548ec26d125d2c182e6149776eef5944bde8" Nov 25 15:38:06 crc kubenswrapper[4890]: I1125 15:38:06.713104 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-l94p8" Nov 25 15:38:06 crc kubenswrapper[4890]: I1125 15:38:06.788466 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7bjzl"] Nov 25 15:38:06 crc kubenswrapper[4890]: E1125 15:38:06.788919 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90b03f38-bd28-4c62-ae05-5cc72f4e6ee7" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 25 15:38:06 crc kubenswrapper[4890]: I1125 15:38:06.788932 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="90b03f38-bd28-4c62-ae05-5cc72f4e6ee7" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 25 15:38:06 crc kubenswrapper[4890]: I1125 15:38:06.789127 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="90b03f38-bd28-4c62-ae05-5cc72f4e6ee7" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 25 15:38:06 crc kubenswrapper[4890]: I1125 15:38:06.790201 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7bjzl" Nov 25 15:38:06 crc kubenswrapper[4890]: I1125 15:38:06.794002 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-22zzp" Nov 25 15:38:06 crc kubenswrapper[4890]: I1125 15:38:06.794303 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 15:38:06 crc kubenswrapper[4890]: I1125 15:38:06.794435 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 15:38:06 crc kubenswrapper[4890]: I1125 15:38:06.796429 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 15:38:06 crc kubenswrapper[4890]: I1125 15:38:06.808702 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7bjzl"] Nov 25 15:38:06 crc kubenswrapper[4890]: I1125 15:38:06.912683 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ba10718-5ae6-4670-bbc6-b02e9c5562e0-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7bjzl\" (UID: \"2ba10718-5ae6-4670-bbc6-b02e9c5562e0\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7bjzl" Nov 25 15:38:06 crc kubenswrapper[4890]: I1125 15:38:06.912789 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mqfc\" (UniqueName: \"kubernetes.io/projected/2ba10718-5ae6-4670-bbc6-b02e9c5562e0-kube-api-access-4mqfc\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7bjzl\" (UID: \"2ba10718-5ae6-4670-bbc6-b02e9c5562e0\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7bjzl" Nov 25 15:38:06 crc kubenswrapper[4890]: I1125 15:38:06.913082 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ba10718-5ae6-4670-bbc6-b02e9c5562e0-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7bjzl\" (UID: \"2ba10718-5ae6-4670-bbc6-b02e9c5562e0\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7bjzl" Nov 25 15:38:06 crc kubenswrapper[4890]: E1125 15:38:06.936433 4890 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod90b03f38_bd28_4c62_ae05_5cc72f4e6ee7.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod90b03f38_bd28_4c62_ae05_5cc72f4e6ee7.slice/crio-baa5396137e8e803340bc86db0ea548ec26d125d2c182e6149776eef5944bde8\": RecentStats: unable to find data in memory cache]" Nov 25 15:38:07 crc kubenswrapper[4890]: I1125 15:38:07.014989 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ba10718-5ae6-4670-bbc6-b02e9c5562e0-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7bjzl\" (UID: \"2ba10718-5ae6-4670-bbc6-b02e9c5562e0\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7bjzl" Nov 25 15:38:07 crc kubenswrapper[4890]: I1125 15:38:07.015382 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mqfc\" (UniqueName: \"kubernetes.io/projected/2ba10718-5ae6-4670-bbc6-b02e9c5562e0-kube-api-access-4mqfc\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7bjzl\" (UID: \"2ba10718-5ae6-4670-bbc6-b02e9c5562e0\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7bjzl" Nov 25 15:38:07 crc kubenswrapper[4890]: I1125 15:38:07.015435 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ba10718-5ae6-4670-bbc6-b02e9c5562e0-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7bjzl\" (UID: \"2ba10718-5ae6-4670-bbc6-b02e9c5562e0\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7bjzl" Nov 25 15:38:07 crc kubenswrapper[4890]: I1125 15:38:07.020962 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ba10718-5ae6-4670-bbc6-b02e9c5562e0-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7bjzl\" (UID: \"2ba10718-5ae6-4670-bbc6-b02e9c5562e0\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7bjzl" Nov 25 15:38:07 crc kubenswrapper[4890]: I1125 15:38:07.021114 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ba10718-5ae6-4670-bbc6-b02e9c5562e0-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7bjzl\" (UID: \"2ba10718-5ae6-4670-bbc6-b02e9c5562e0\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7bjzl" Nov 25 15:38:07 crc kubenswrapper[4890]: I1125 15:38:07.049962 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mqfc\" (UniqueName: \"kubernetes.io/projected/2ba10718-5ae6-4670-bbc6-b02e9c5562e0-kube-api-access-4mqfc\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7bjzl\" (UID: \"2ba10718-5ae6-4670-bbc6-b02e9c5562e0\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7bjzl" Nov 25 15:38:07 crc kubenswrapper[4890]: I1125 15:38:07.116617 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7bjzl" Nov 25 15:38:07 crc kubenswrapper[4890]: I1125 15:38:07.680422 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7bjzl"] Nov 25 15:38:07 crc kubenswrapper[4890]: I1125 15:38:07.724037 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7bjzl" event={"ID":"2ba10718-5ae6-4670-bbc6-b02e9c5562e0","Type":"ContainerStarted","Data":"98f2a04b722b9adee57405598c8a028ef1be70facede4b3eadbd6a6fadf26f7d"} Nov 25 15:38:08 crc kubenswrapper[4890]: I1125 15:38:08.735184 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7bjzl" event={"ID":"2ba10718-5ae6-4670-bbc6-b02e9c5562e0","Type":"ContainerStarted","Data":"a544aa83848b4ba0a9fbcd084f0eb67bc51aecd955eb79f5d075f4d64827fe5f"} Nov 25 15:38:08 crc kubenswrapper[4890]: I1125 15:38:08.762499 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7bjzl" podStartSLOduration=2.238204859 podStartE2EDuration="2.762418226s" podCreationTimestamp="2025-11-25 15:38:06 +0000 UTC" firstStartedPulling="2025-11-25 15:38:07.683494351 +0000 UTC m=+2146.125956961" lastFinishedPulling="2025-11-25 15:38:08.207707718 +0000 UTC m=+2146.650170328" observedRunningTime="2025-11-25 15:38:08.75178142 +0000 UTC m=+2147.194244030" watchObservedRunningTime="2025-11-25 15:38:08.762418226 +0000 UTC m=+2147.204880826" Nov 25 15:38:17 crc kubenswrapper[4890]: I1125 15:38:17.818913 4890 generic.go:334] "Generic (PLEG): container finished" podID="2ba10718-5ae6-4670-bbc6-b02e9c5562e0" containerID="a544aa83848b4ba0a9fbcd084f0eb67bc51aecd955eb79f5d075f4d64827fe5f" exitCode=0 Nov 25 15:38:17 crc kubenswrapper[4890]: I1125 15:38:17.819001 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7bjzl" event={"ID":"2ba10718-5ae6-4670-bbc6-b02e9c5562e0","Type":"ContainerDied","Data":"a544aa83848b4ba0a9fbcd084f0eb67bc51aecd955eb79f5d075f4d64827fe5f"} Nov 25 15:38:19 crc kubenswrapper[4890]: I1125 15:38:19.259726 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7bjzl" Nov 25 15:38:19 crc kubenswrapper[4890]: I1125 15:38:19.427352 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ba10718-5ae6-4670-bbc6-b02e9c5562e0-inventory\") pod \"2ba10718-5ae6-4670-bbc6-b02e9c5562e0\" (UID: \"2ba10718-5ae6-4670-bbc6-b02e9c5562e0\") " Nov 25 15:38:19 crc kubenswrapper[4890]: I1125 15:38:19.427501 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ba10718-5ae6-4670-bbc6-b02e9c5562e0-ssh-key\") pod \"2ba10718-5ae6-4670-bbc6-b02e9c5562e0\" (UID: \"2ba10718-5ae6-4670-bbc6-b02e9c5562e0\") " Nov 25 15:38:19 crc kubenswrapper[4890]: I1125 15:38:19.427686 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4mqfc\" (UniqueName: \"kubernetes.io/projected/2ba10718-5ae6-4670-bbc6-b02e9c5562e0-kube-api-access-4mqfc\") pod \"2ba10718-5ae6-4670-bbc6-b02e9c5562e0\" (UID: \"2ba10718-5ae6-4670-bbc6-b02e9c5562e0\") " Nov 25 15:38:19 crc kubenswrapper[4890]: I1125 15:38:19.435258 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ba10718-5ae6-4670-bbc6-b02e9c5562e0-kube-api-access-4mqfc" (OuterVolumeSpecName: "kube-api-access-4mqfc") pod "2ba10718-5ae6-4670-bbc6-b02e9c5562e0" (UID: "2ba10718-5ae6-4670-bbc6-b02e9c5562e0"). InnerVolumeSpecName "kube-api-access-4mqfc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:38:19 crc kubenswrapper[4890]: E1125 15:38:19.454892 4890 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2ba10718-5ae6-4670-bbc6-b02e9c5562e0-ssh-key podName:2ba10718-5ae6-4670-bbc6-b02e9c5562e0 nodeName:}" failed. No retries permitted until 2025-11-25 15:38:19.954862942 +0000 UTC m=+2158.397325552 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "ssh-key" (UniqueName: "kubernetes.io/secret/2ba10718-5ae6-4670-bbc6-b02e9c5562e0-ssh-key") pod "2ba10718-5ae6-4670-bbc6-b02e9c5562e0" (UID: "2ba10718-5ae6-4670-bbc6-b02e9c5562e0") : error deleting /var/lib/kubelet/pods/2ba10718-5ae6-4670-bbc6-b02e9c5562e0/volume-subpaths: remove /var/lib/kubelet/pods/2ba10718-5ae6-4670-bbc6-b02e9c5562e0/volume-subpaths: no such file or directory Nov 25 15:38:19 crc kubenswrapper[4890]: I1125 15:38:19.457834 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ba10718-5ae6-4670-bbc6-b02e9c5562e0-inventory" (OuterVolumeSpecName: "inventory") pod "2ba10718-5ae6-4670-bbc6-b02e9c5562e0" (UID: "2ba10718-5ae6-4670-bbc6-b02e9c5562e0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:38:19 crc kubenswrapper[4890]: I1125 15:38:19.532078 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4mqfc\" (UniqueName: \"kubernetes.io/projected/2ba10718-5ae6-4670-bbc6-b02e9c5562e0-kube-api-access-4mqfc\") on node \"crc\" DevicePath \"\"" Nov 25 15:38:19 crc kubenswrapper[4890]: I1125 15:38:19.532134 4890 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ba10718-5ae6-4670-bbc6-b02e9c5562e0-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 15:38:19 crc kubenswrapper[4890]: I1125 15:38:19.849006 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7bjzl" event={"ID":"2ba10718-5ae6-4670-bbc6-b02e9c5562e0","Type":"ContainerDied","Data":"98f2a04b722b9adee57405598c8a028ef1be70facede4b3eadbd6a6fadf26f7d"} Nov 25 15:38:19 crc kubenswrapper[4890]: I1125 15:38:19.849057 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="98f2a04b722b9adee57405598c8a028ef1be70facede4b3eadbd6a6fadf26f7d" Nov 25 15:38:19 crc kubenswrapper[4890]: I1125 15:38:19.849113 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7bjzl" Nov 25 15:38:19 crc kubenswrapper[4890]: I1125 15:38:19.951509 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg"] Nov 25 15:38:19 crc kubenswrapper[4890]: E1125 15:38:19.951935 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ba10718-5ae6-4670-bbc6-b02e9c5562e0" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 25 15:38:19 crc kubenswrapper[4890]: I1125 15:38:19.951957 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ba10718-5ae6-4670-bbc6-b02e9c5562e0" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 25 15:38:19 crc kubenswrapper[4890]: I1125 15:38:19.952372 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ba10718-5ae6-4670-bbc6-b02e9c5562e0" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 25 15:38:19 crc kubenswrapper[4890]: I1125 15:38:19.953153 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" Nov 25 15:38:19 crc kubenswrapper[4890]: I1125 15:38:19.958898 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Nov 25 15:38:19 crc kubenswrapper[4890]: I1125 15:38:19.959416 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Nov 25 15:38:19 crc kubenswrapper[4890]: I1125 15:38:19.959559 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Nov 25 15:38:19 crc kubenswrapper[4890]: I1125 15:38:19.959778 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Nov 25 15:38:19 crc kubenswrapper[4890]: I1125 15:38:19.987307 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg"] Nov 25 15:38:20 crc kubenswrapper[4890]: I1125 15:38:20.044637 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ba10718-5ae6-4670-bbc6-b02e9c5562e0-ssh-key\") pod \"2ba10718-5ae6-4670-bbc6-b02e9c5562e0\" (UID: \"2ba10718-5ae6-4670-bbc6-b02e9c5562e0\") " Nov 25 15:38:20 crc kubenswrapper[4890]: I1125 15:38:20.045130 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-j96lg\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" Nov 25 15:38:20 crc kubenswrapper[4890]: I1125 15:38:20.045196 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-j96lg\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" Nov 25 15:38:20 crc kubenswrapper[4890]: I1125 15:38:20.045227 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-j96lg\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" Nov 25 15:38:20 crc kubenswrapper[4890]: I1125 15:38:20.045370 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-j96lg\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" Nov 25 15:38:20 crc kubenswrapper[4890]: I1125 15:38:20.045447 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-j96lg\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" Nov 25 15:38:20 crc kubenswrapper[4890]: I1125 15:38:20.045600 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-j96lg\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" Nov 25 15:38:20 crc kubenswrapper[4890]: I1125 15:38:20.045831 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-j96lg\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" Nov 25 15:38:20 crc kubenswrapper[4890]: I1125 15:38:20.045900 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-j96lg\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" Nov 25 15:38:20 crc kubenswrapper[4890]: I1125 15:38:20.046038 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwg4k\" (UniqueName: \"kubernetes.io/projected/2a24b42b-12c8-43bc-9dc7-f07991224497-kube-api-access-dwg4k\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-j96lg\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" Nov 25 15:38:20 crc kubenswrapper[4890]: I1125 15:38:20.046103 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2a24b42b-12c8-43bc-9dc7-f07991224497-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-j96lg\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" Nov 25 15:38:20 crc kubenswrapper[4890]: I1125 15:38:20.046248 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-j96lg\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" Nov 25 15:38:20 crc kubenswrapper[4890]: I1125 15:38:20.046361 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2a24b42b-12c8-43bc-9dc7-f07991224497-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-j96lg\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" Nov 25 15:38:20 crc kubenswrapper[4890]: I1125 15:38:20.046455 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2a24b42b-12c8-43bc-9dc7-f07991224497-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-j96lg\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" Nov 25 15:38:20 crc kubenswrapper[4890]: I1125 15:38:20.046514 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2a24b42b-12c8-43bc-9dc7-f07991224497-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-j96lg\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" Nov 25 15:38:20 crc kubenswrapper[4890]: I1125 15:38:20.048708 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ba10718-5ae6-4670-bbc6-b02e9c5562e0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2ba10718-5ae6-4670-bbc6-b02e9c5562e0" (UID: "2ba10718-5ae6-4670-bbc6-b02e9c5562e0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:38:20 crc kubenswrapper[4890]: I1125 15:38:20.148298 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-j96lg\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" Nov 25 15:38:20 crc kubenswrapper[4890]: I1125 15:38:20.148380 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-j96lg\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" Nov 25 15:38:20 crc kubenswrapper[4890]: I1125 15:38:20.148438 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwg4k\" (UniqueName: \"kubernetes.io/projected/2a24b42b-12c8-43bc-9dc7-f07991224497-kube-api-access-dwg4k\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-j96lg\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" Nov 25 15:38:20 crc kubenswrapper[4890]: I1125 15:38:20.148473 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2a24b42b-12c8-43bc-9dc7-f07991224497-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-j96lg\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" Nov 25 15:38:20 crc kubenswrapper[4890]: I1125 15:38:20.148518 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-j96lg\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" Nov 25 15:38:20 crc kubenswrapper[4890]: I1125 15:38:20.148560 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2a24b42b-12c8-43bc-9dc7-f07991224497-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-j96lg\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" Nov 25 15:38:20 crc kubenswrapper[4890]: I1125 15:38:20.148603 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2a24b42b-12c8-43bc-9dc7-f07991224497-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-j96lg\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" Nov 25 15:38:20 crc kubenswrapper[4890]: I1125 15:38:20.148636 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2a24b42b-12c8-43bc-9dc7-f07991224497-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-j96lg\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" Nov 25 15:38:20 crc kubenswrapper[4890]: I1125 15:38:20.148686 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-j96lg\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" Nov 25 15:38:20 crc kubenswrapper[4890]: I1125 15:38:20.148715 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-j96lg\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" Nov 25 15:38:20 crc kubenswrapper[4890]: I1125 15:38:20.148742 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-j96lg\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" Nov 25 15:38:20 crc kubenswrapper[4890]: I1125 15:38:20.148773 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-j96lg\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" Nov 25 15:38:20 crc kubenswrapper[4890]: I1125 15:38:20.148812 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-j96lg\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" Nov 25 15:38:20 crc kubenswrapper[4890]: I1125 15:38:20.148839 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-j96lg\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" Nov 25 15:38:20 crc kubenswrapper[4890]: I1125 15:38:20.148911 4890 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ba10718-5ae6-4670-bbc6-b02e9c5562e0-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 15:38:20 crc kubenswrapper[4890]: I1125 15:38:20.153720 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2a24b42b-12c8-43bc-9dc7-f07991224497-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-j96lg\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" Nov 25 15:38:20 crc kubenswrapper[4890]: I1125 15:38:20.153730 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-j96lg\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" Nov 25 15:38:20 crc kubenswrapper[4890]: I1125 15:38:20.153825 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2a24b42b-12c8-43bc-9dc7-f07991224497-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-j96lg\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" Nov 25 15:38:20 crc kubenswrapper[4890]: I1125 15:38:20.154388 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-j96lg\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" Nov 25 15:38:20 crc kubenswrapper[4890]: I1125 15:38:20.154628 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-j96lg\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" Nov 25 15:38:20 crc kubenswrapper[4890]: I1125 15:38:20.154635 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-j96lg\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" Nov 25 15:38:20 crc kubenswrapper[4890]: I1125 15:38:20.155119 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2a24b42b-12c8-43bc-9dc7-f07991224497-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-j96lg\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" Nov 25 15:38:20 crc kubenswrapper[4890]: I1125 15:38:20.156650 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-j96lg\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" Nov 25 15:38:20 crc kubenswrapper[4890]: I1125 15:38:20.156740 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-j96lg\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" Nov 25 15:38:20 crc kubenswrapper[4890]: I1125 15:38:20.157083 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-j96lg\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" Nov 25 15:38:20 crc kubenswrapper[4890]: I1125 15:38:20.157806 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2a24b42b-12c8-43bc-9dc7-f07991224497-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-j96lg\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" Nov 25 15:38:20 crc kubenswrapper[4890]: I1125 15:38:20.165825 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-j96lg\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" Nov 25 15:38:20 crc kubenswrapper[4890]: I1125 15:38:20.171680 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-j96lg\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" Nov 25 15:38:20 crc kubenswrapper[4890]: I1125 15:38:20.174667 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwg4k\" (UniqueName: \"kubernetes.io/projected/2a24b42b-12c8-43bc-9dc7-f07991224497-kube-api-access-dwg4k\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-j96lg\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" Nov 25 15:38:20 crc kubenswrapper[4890]: I1125 15:38:20.278510 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" Nov 25 15:38:20 crc kubenswrapper[4890]: I1125 15:38:20.846461 4890 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 15:38:20 crc kubenswrapper[4890]: I1125 15:38:20.851449 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg"] Nov 25 15:38:20 crc kubenswrapper[4890]: I1125 15:38:20.864515 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" event={"ID":"2a24b42b-12c8-43bc-9dc7-f07991224497","Type":"ContainerStarted","Data":"1ff52b7330be1921f97fd5bd454aa1560b5167bf1e97cb4a936e2636d24aea3a"} Nov 25 15:38:22 crc kubenswrapper[4890]: I1125 15:38:22.894084 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" event={"ID":"2a24b42b-12c8-43bc-9dc7-f07991224497","Type":"ContainerStarted","Data":"324b9f6570c5c1e006280c0019f913f52f95acb86496a659e1d032146f065dd4"} Nov 25 15:38:22 crc kubenswrapper[4890]: I1125 15:38:22.918655 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" podStartSLOduration=2.482536289 podStartE2EDuration="3.918633186s" podCreationTimestamp="2025-11-25 15:38:19 +0000 UTC" firstStartedPulling="2025-11-25 15:38:20.84611993 +0000 UTC m=+2159.288582540" lastFinishedPulling="2025-11-25 15:38:22.282216827 +0000 UTC m=+2160.724679437" observedRunningTime="2025-11-25 15:38:22.913713603 +0000 UTC m=+2161.356176233" watchObservedRunningTime="2025-11-25 15:38:22.918633186 +0000 UTC m=+2161.361095796" Nov 25 15:38:26 crc kubenswrapper[4890]: I1125 15:38:26.447119 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 15:38:26 crc kubenswrapper[4890]: I1125 15:38:26.449974 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 15:38:56 crc kubenswrapper[4890]: I1125 15:38:56.447603 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 15:38:56 crc kubenswrapper[4890]: I1125 15:38:56.448619 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 15:38:58 crc kubenswrapper[4890]: I1125 15:38:58.247263 4890 generic.go:334] "Generic (PLEG): container finished" podID="2a24b42b-12c8-43bc-9dc7-f07991224497" containerID="324b9f6570c5c1e006280c0019f913f52f95acb86496a659e1d032146f065dd4" exitCode=0 Nov 25 15:38:58 crc kubenswrapper[4890]: I1125 15:38:58.247427 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" event={"ID":"2a24b42b-12c8-43bc-9dc7-f07991224497","Type":"ContainerDied","Data":"324b9f6570c5c1e006280c0019f913f52f95acb86496a659e1d032146f065dd4"} Nov 25 15:38:59 crc kubenswrapper[4890]: I1125 15:38:59.674262 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" Nov 25 15:38:59 crc kubenswrapper[4890]: I1125 15:38:59.873055 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-libvirt-combined-ca-bundle\") pod \"2a24b42b-12c8-43bc-9dc7-f07991224497\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " Nov 25 15:38:59 crc kubenswrapper[4890]: I1125 15:38:59.873139 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2a24b42b-12c8-43bc-9dc7-f07991224497-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"2a24b42b-12c8-43bc-9dc7-f07991224497\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " Nov 25 15:38:59 crc kubenswrapper[4890]: I1125 15:38:59.873206 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-ovn-combined-ca-bundle\") pod \"2a24b42b-12c8-43bc-9dc7-f07991224497\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " Nov 25 15:38:59 crc kubenswrapper[4890]: I1125 15:38:59.873498 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2a24b42b-12c8-43bc-9dc7-f07991224497-openstack-edpm-ipam-ovn-default-certs-0\") pod \"2a24b42b-12c8-43bc-9dc7-f07991224497\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " Nov 25 15:38:59 crc kubenswrapper[4890]: I1125 15:38:59.873571 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2a24b42b-12c8-43bc-9dc7-f07991224497-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"2a24b42b-12c8-43bc-9dc7-f07991224497\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " Nov 25 15:38:59 crc kubenswrapper[4890]: I1125 15:38:59.874232 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-neutron-metadata-combined-ca-bundle\") pod \"2a24b42b-12c8-43bc-9dc7-f07991224497\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " Nov 25 15:38:59 crc kubenswrapper[4890]: I1125 15:38:59.874275 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-repo-setup-combined-ca-bundle\") pod \"2a24b42b-12c8-43bc-9dc7-f07991224497\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " Nov 25 15:38:59 crc kubenswrapper[4890]: I1125 15:38:59.874308 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-telemetry-combined-ca-bundle\") pod \"2a24b42b-12c8-43bc-9dc7-f07991224497\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " Nov 25 15:38:59 crc kubenswrapper[4890]: I1125 15:38:59.874343 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwg4k\" (UniqueName: \"kubernetes.io/projected/2a24b42b-12c8-43bc-9dc7-f07991224497-kube-api-access-dwg4k\") pod \"2a24b42b-12c8-43bc-9dc7-f07991224497\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " Nov 25 15:38:59 crc kubenswrapper[4890]: I1125 15:38:59.874457 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-inventory\") pod \"2a24b42b-12c8-43bc-9dc7-f07991224497\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " Nov 25 15:38:59 crc kubenswrapper[4890]: I1125 15:38:59.874498 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-ssh-key\") pod \"2a24b42b-12c8-43bc-9dc7-f07991224497\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " Nov 25 15:38:59 crc kubenswrapper[4890]: I1125 15:38:59.874547 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-nova-combined-ca-bundle\") pod \"2a24b42b-12c8-43bc-9dc7-f07991224497\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " Nov 25 15:38:59 crc kubenswrapper[4890]: I1125 15:38:59.874619 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-bootstrap-combined-ca-bundle\") pod \"2a24b42b-12c8-43bc-9dc7-f07991224497\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " Nov 25 15:38:59 crc kubenswrapper[4890]: I1125 15:38:59.874675 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2a24b42b-12c8-43bc-9dc7-f07991224497-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"2a24b42b-12c8-43bc-9dc7-f07991224497\" (UID: \"2a24b42b-12c8-43bc-9dc7-f07991224497\") " Nov 25 15:38:59 crc kubenswrapper[4890]: I1125 15:38:59.881197 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "2a24b42b-12c8-43bc-9dc7-f07991224497" (UID: "2a24b42b-12c8-43bc-9dc7-f07991224497"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:38:59 crc kubenswrapper[4890]: I1125 15:38:59.881520 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "2a24b42b-12c8-43bc-9dc7-f07991224497" (UID: "2a24b42b-12c8-43bc-9dc7-f07991224497"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:38:59 crc kubenswrapper[4890]: I1125 15:38:59.881611 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a24b42b-12c8-43bc-9dc7-f07991224497-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "2a24b42b-12c8-43bc-9dc7-f07991224497" (UID: "2a24b42b-12c8-43bc-9dc7-f07991224497"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:38:59 crc kubenswrapper[4890]: I1125 15:38:59.883632 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "2a24b42b-12c8-43bc-9dc7-f07991224497" (UID: "2a24b42b-12c8-43bc-9dc7-f07991224497"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:38:59 crc kubenswrapper[4890]: I1125 15:38:59.883717 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "2a24b42b-12c8-43bc-9dc7-f07991224497" (UID: "2a24b42b-12c8-43bc-9dc7-f07991224497"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:38:59 crc kubenswrapper[4890]: I1125 15:38:59.883804 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a24b42b-12c8-43bc-9dc7-f07991224497-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "2a24b42b-12c8-43bc-9dc7-f07991224497" (UID: "2a24b42b-12c8-43bc-9dc7-f07991224497"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:38:59 crc kubenswrapper[4890]: I1125 15:38:59.883813 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "2a24b42b-12c8-43bc-9dc7-f07991224497" (UID: "2a24b42b-12c8-43bc-9dc7-f07991224497"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:38:59 crc kubenswrapper[4890]: I1125 15:38:59.883838 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a24b42b-12c8-43bc-9dc7-f07991224497-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "2a24b42b-12c8-43bc-9dc7-f07991224497" (UID: "2a24b42b-12c8-43bc-9dc7-f07991224497"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:38:59 crc kubenswrapper[4890]: I1125 15:38:59.883975 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a24b42b-12c8-43bc-9dc7-f07991224497-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "2a24b42b-12c8-43bc-9dc7-f07991224497" (UID: "2a24b42b-12c8-43bc-9dc7-f07991224497"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:38:59 crc kubenswrapper[4890]: I1125 15:38:59.884405 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a24b42b-12c8-43bc-9dc7-f07991224497-kube-api-access-dwg4k" (OuterVolumeSpecName: "kube-api-access-dwg4k") pod "2a24b42b-12c8-43bc-9dc7-f07991224497" (UID: "2a24b42b-12c8-43bc-9dc7-f07991224497"). InnerVolumeSpecName "kube-api-access-dwg4k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:38:59 crc kubenswrapper[4890]: I1125 15:38:59.885635 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "2a24b42b-12c8-43bc-9dc7-f07991224497" (UID: "2a24b42b-12c8-43bc-9dc7-f07991224497"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:38:59 crc kubenswrapper[4890]: I1125 15:38:59.895446 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "2a24b42b-12c8-43bc-9dc7-f07991224497" (UID: "2a24b42b-12c8-43bc-9dc7-f07991224497"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:38:59 crc kubenswrapper[4890]: I1125 15:38:59.913793 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2a24b42b-12c8-43bc-9dc7-f07991224497" (UID: "2a24b42b-12c8-43bc-9dc7-f07991224497"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:38:59 crc kubenswrapper[4890]: I1125 15:38:59.917703 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-inventory" (OuterVolumeSpecName: "inventory") pod "2a24b42b-12c8-43bc-9dc7-f07991224497" (UID: "2a24b42b-12c8-43bc-9dc7-f07991224497"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:38:59 crc kubenswrapper[4890]: I1125 15:38:59.977999 4890 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 15:38:59 crc kubenswrapper[4890]: I1125 15:38:59.978042 4890 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 15:38:59 crc kubenswrapper[4890]: I1125 15:38:59.978052 4890 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:38:59 crc kubenswrapper[4890]: I1125 15:38:59.978065 4890 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:38:59 crc kubenswrapper[4890]: I1125 15:38:59.978076 4890 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2a24b42b-12c8-43bc-9dc7-f07991224497-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 25 15:38:59 crc kubenswrapper[4890]: I1125 15:38:59.978085 4890 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:38:59 crc kubenswrapper[4890]: I1125 15:38:59.978096 4890 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2a24b42b-12c8-43bc-9dc7-f07991224497-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 25 15:38:59 crc kubenswrapper[4890]: I1125 15:38:59.978108 4890 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:38:59 crc kubenswrapper[4890]: I1125 15:38:59.978118 4890 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2a24b42b-12c8-43bc-9dc7-f07991224497-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 25 15:38:59 crc kubenswrapper[4890]: I1125 15:38:59.978137 4890 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2a24b42b-12c8-43bc-9dc7-f07991224497-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 25 15:38:59 crc kubenswrapper[4890]: I1125 15:38:59.978177 4890 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:38:59 crc kubenswrapper[4890]: I1125 15:38:59.978191 4890 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:38:59 crc kubenswrapper[4890]: I1125 15:38:59.978200 4890 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a24b42b-12c8-43bc-9dc7-f07991224497-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:38:59 crc kubenswrapper[4890]: I1125 15:38:59.978211 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwg4k\" (UniqueName: \"kubernetes.io/projected/2a24b42b-12c8-43bc-9dc7-f07991224497-kube-api-access-dwg4k\") on node \"crc\" DevicePath \"\"" Nov 25 15:39:00 crc kubenswrapper[4890]: I1125 15:39:00.267211 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" event={"ID":"2a24b42b-12c8-43bc-9dc7-f07991224497","Type":"ContainerDied","Data":"1ff52b7330be1921f97fd5bd454aa1560b5167bf1e97cb4a936e2636d24aea3a"} Nov 25 15:39:00 crc kubenswrapper[4890]: I1125 15:39:00.267269 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ff52b7330be1921f97fd5bd454aa1560b5167bf1e97cb4a936e2636d24aea3a" Nov 25 15:39:00 crc kubenswrapper[4890]: I1125 15:39:00.267243 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-j96lg" Nov 25 15:39:00 crc kubenswrapper[4890]: I1125 15:39:00.378996 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-llz86"] Nov 25 15:39:00 crc kubenswrapper[4890]: E1125 15:39:00.379575 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a24b42b-12c8-43bc-9dc7-f07991224497" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 25 15:39:00 crc kubenswrapper[4890]: I1125 15:39:00.379601 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a24b42b-12c8-43bc-9dc7-f07991224497" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 25 15:39:00 crc kubenswrapper[4890]: I1125 15:39:00.379810 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a24b42b-12c8-43bc-9dc7-f07991224497" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 25 15:39:00 crc kubenswrapper[4890]: I1125 15:39:00.380630 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-llz86" Nov 25 15:39:00 crc kubenswrapper[4890]: I1125 15:39:00.383846 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-22zzp" Nov 25 15:39:00 crc kubenswrapper[4890]: I1125 15:39:00.383872 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Nov 25 15:39:00 crc kubenswrapper[4890]: I1125 15:39:00.383976 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 15:39:00 crc kubenswrapper[4890]: I1125 15:39:00.383851 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 15:39:00 crc kubenswrapper[4890]: I1125 15:39:00.387007 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 15:39:00 crc kubenswrapper[4890]: I1125 15:39:00.389592 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-llz86"] Nov 25 15:39:00 crc kubenswrapper[4890]: I1125 15:39:00.486863 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/995132c2-f199-4d15-b3ef-3c71fed454f5-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-llz86\" (UID: \"995132c2-f199-4d15-b3ef-3c71fed454f5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-llz86" Nov 25 15:39:00 crc kubenswrapper[4890]: I1125 15:39:00.486936 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t44ql\" (UniqueName: \"kubernetes.io/projected/995132c2-f199-4d15-b3ef-3c71fed454f5-kube-api-access-t44ql\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-llz86\" (UID: \"995132c2-f199-4d15-b3ef-3c71fed454f5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-llz86" Nov 25 15:39:00 crc kubenswrapper[4890]: I1125 15:39:00.486980 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/995132c2-f199-4d15-b3ef-3c71fed454f5-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-llz86\" (UID: \"995132c2-f199-4d15-b3ef-3c71fed454f5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-llz86" Nov 25 15:39:00 crc kubenswrapper[4890]: I1125 15:39:00.487132 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/995132c2-f199-4d15-b3ef-3c71fed454f5-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-llz86\" (UID: \"995132c2-f199-4d15-b3ef-3c71fed454f5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-llz86" Nov 25 15:39:00 crc kubenswrapper[4890]: I1125 15:39:00.487284 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/995132c2-f199-4d15-b3ef-3c71fed454f5-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-llz86\" (UID: \"995132c2-f199-4d15-b3ef-3c71fed454f5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-llz86" Nov 25 15:39:00 crc kubenswrapper[4890]: I1125 15:39:00.589476 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/995132c2-f199-4d15-b3ef-3c71fed454f5-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-llz86\" (UID: \"995132c2-f199-4d15-b3ef-3c71fed454f5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-llz86" Nov 25 15:39:00 crc kubenswrapper[4890]: I1125 15:39:00.589913 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/995132c2-f199-4d15-b3ef-3c71fed454f5-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-llz86\" (UID: \"995132c2-f199-4d15-b3ef-3c71fed454f5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-llz86" Nov 25 15:39:00 crc kubenswrapper[4890]: I1125 15:39:00.589983 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/995132c2-f199-4d15-b3ef-3c71fed454f5-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-llz86\" (UID: \"995132c2-f199-4d15-b3ef-3c71fed454f5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-llz86" Nov 25 15:39:00 crc kubenswrapper[4890]: I1125 15:39:00.590073 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/995132c2-f199-4d15-b3ef-3c71fed454f5-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-llz86\" (UID: \"995132c2-f199-4d15-b3ef-3c71fed454f5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-llz86" Nov 25 15:39:00 crc kubenswrapper[4890]: I1125 15:39:00.590105 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t44ql\" (UniqueName: \"kubernetes.io/projected/995132c2-f199-4d15-b3ef-3c71fed454f5-kube-api-access-t44ql\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-llz86\" (UID: \"995132c2-f199-4d15-b3ef-3c71fed454f5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-llz86" Nov 25 15:39:00 crc kubenswrapper[4890]: I1125 15:39:00.593126 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/995132c2-f199-4d15-b3ef-3c71fed454f5-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-llz86\" (UID: \"995132c2-f199-4d15-b3ef-3c71fed454f5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-llz86" Nov 25 15:39:00 crc kubenswrapper[4890]: I1125 15:39:00.597151 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/995132c2-f199-4d15-b3ef-3c71fed454f5-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-llz86\" (UID: \"995132c2-f199-4d15-b3ef-3c71fed454f5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-llz86" Nov 25 15:39:00 crc kubenswrapper[4890]: I1125 15:39:00.597810 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/995132c2-f199-4d15-b3ef-3c71fed454f5-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-llz86\" (UID: \"995132c2-f199-4d15-b3ef-3c71fed454f5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-llz86" Nov 25 15:39:00 crc kubenswrapper[4890]: I1125 15:39:00.599056 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/995132c2-f199-4d15-b3ef-3c71fed454f5-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-llz86\" (UID: \"995132c2-f199-4d15-b3ef-3c71fed454f5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-llz86" Nov 25 15:39:00 crc kubenswrapper[4890]: I1125 15:39:00.608966 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t44ql\" (UniqueName: \"kubernetes.io/projected/995132c2-f199-4d15-b3ef-3c71fed454f5-kube-api-access-t44ql\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-llz86\" (UID: \"995132c2-f199-4d15-b3ef-3c71fed454f5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-llz86" Nov 25 15:39:00 crc kubenswrapper[4890]: I1125 15:39:00.701037 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-llz86" Nov 25 15:39:01 crc kubenswrapper[4890]: I1125 15:39:01.391969 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-llz86"] Nov 25 15:39:02 crc kubenswrapper[4890]: I1125 15:39:02.286822 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-llz86" event={"ID":"995132c2-f199-4d15-b3ef-3c71fed454f5","Type":"ContainerStarted","Data":"bba15ff5f74e18952e2cae0e9d8c72f4c71326f8c3c783422f1b6716845e7a6d"} Nov 25 15:39:02 crc kubenswrapper[4890]: I1125 15:39:02.287476 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-llz86" event={"ID":"995132c2-f199-4d15-b3ef-3c71fed454f5","Type":"ContainerStarted","Data":"3ef2357de5ab5029106cba56ef25c2288bd83ebc7e652cbb6c5b5e77347f4429"} Nov 25 15:39:02 crc kubenswrapper[4890]: I1125 15:39:02.319141 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-llz86" podStartSLOduration=1.7701410800000001 podStartE2EDuration="2.319108615s" podCreationTimestamp="2025-11-25 15:39:00 +0000 UTC" firstStartedPulling="2025-11-25 15:39:01.401570372 +0000 UTC m=+2199.844032992" lastFinishedPulling="2025-11-25 15:39:01.950537917 +0000 UTC m=+2200.393000527" observedRunningTime="2025-11-25 15:39:02.306380237 +0000 UTC m=+2200.748842837" watchObservedRunningTime="2025-11-25 15:39:02.319108615 +0000 UTC m=+2200.761571225" Nov 25 15:39:26 crc kubenswrapper[4890]: I1125 15:39:26.447701 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 15:39:26 crc kubenswrapper[4890]: I1125 15:39:26.448441 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 15:39:26 crc kubenswrapper[4890]: I1125 15:39:26.448501 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" Nov 25 15:39:26 crc kubenswrapper[4890]: I1125 15:39:26.449646 4890 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"67ca0de5cd1ef0f73ff70ee2df49e62ca714111ad20a0de1fef88f6f60fdf262"} pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 15:39:26 crc kubenswrapper[4890]: I1125 15:39:26.449726 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" containerID="cri-o://67ca0de5cd1ef0f73ff70ee2df49e62ca714111ad20a0de1fef88f6f60fdf262" gracePeriod=600 Nov 25 15:39:26 crc kubenswrapper[4890]: E1125 15:39:26.579411 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:39:27 crc kubenswrapper[4890]: I1125 15:39:27.555803 4890 generic.go:334] "Generic (PLEG): container finished" podID="4e4f849d-f239-4727-a73e-18327856929a" containerID="67ca0de5cd1ef0f73ff70ee2df49e62ca714111ad20a0de1fef88f6f60fdf262" exitCode=0 Nov 25 15:39:27 crc kubenswrapper[4890]: I1125 15:39:27.555858 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" event={"ID":"4e4f849d-f239-4727-a73e-18327856929a","Type":"ContainerDied","Data":"67ca0de5cd1ef0f73ff70ee2df49e62ca714111ad20a0de1fef88f6f60fdf262"} Nov 25 15:39:27 crc kubenswrapper[4890]: I1125 15:39:27.555897 4890 scope.go:117] "RemoveContainer" containerID="c2b9ae015432b5c3d0ae89bbacbfeadbc65550341b67aff696270c5f412db829" Nov 25 15:39:27 crc kubenswrapper[4890]: I1125 15:39:27.556683 4890 scope.go:117] "RemoveContainer" containerID="67ca0de5cd1ef0f73ff70ee2df49e62ca714111ad20a0de1fef88f6f60fdf262" Nov 25 15:39:27 crc kubenswrapper[4890]: E1125 15:39:27.556951 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:39:38 crc kubenswrapper[4890]: I1125 15:39:38.347609 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dbb29"] Nov 25 15:39:38 crc kubenswrapper[4890]: I1125 15:39:38.350690 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dbb29" Nov 25 15:39:38 crc kubenswrapper[4890]: I1125 15:39:38.362716 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dbb29"] Nov 25 15:39:38 crc kubenswrapper[4890]: I1125 15:39:38.472806 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c839380c-eefd-4354-81b5-03b3bc50d606-catalog-content\") pod \"redhat-marketplace-dbb29\" (UID: \"c839380c-eefd-4354-81b5-03b3bc50d606\") " pod="openshift-marketplace/redhat-marketplace-dbb29" Nov 25 15:39:38 crc kubenswrapper[4890]: I1125 15:39:38.472926 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5hwl\" (UniqueName: \"kubernetes.io/projected/c839380c-eefd-4354-81b5-03b3bc50d606-kube-api-access-s5hwl\") pod \"redhat-marketplace-dbb29\" (UID: \"c839380c-eefd-4354-81b5-03b3bc50d606\") " pod="openshift-marketplace/redhat-marketplace-dbb29" Nov 25 15:39:38 crc kubenswrapper[4890]: I1125 15:39:38.472975 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c839380c-eefd-4354-81b5-03b3bc50d606-utilities\") pod \"redhat-marketplace-dbb29\" (UID: \"c839380c-eefd-4354-81b5-03b3bc50d606\") " pod="openshift-marketplace/redhat-marketplace-dbb29" Nov 25 15:39:38 crc kubenswrapper[4890]: I1125 15:39:38.575273 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c839380c-eefd-4354-81b5-03b3bc50d606-catalog-content\") pod \"redhat-marketplace-dbb29\" (UID: \"c839380c-eefd-4354-81b5-03b3bc50d606\") " pod="openshift-marketplace/redhat-marketplace-dbb29" Nov 25 15:39:38 crc kubenswrapper[4890]: I1125 15:39:38.575401 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5hwl\" (UniqueName: \"kubernetes.io/projected/c839380c-eefd-4354-81b5-03b3bc50d606-kube-api-access-s5hwl\") pod \"redhat-marketplace-dbb29\" (UID: \"c839380c-eefd-4354-81b5-03b3bc50d606\") " pod="openshift-marketplace/redhat-marketplace-dbb29" Nov 25 15:39:38 crc kubenswrapper[4890]: I1125 15:39:38.575457 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c839380c-eefd-4354-81b5-03b3bc50d606-utilities\") pod \"redhat-marketplace-dbb29\" (UID: \"c839380c-eefd-4354-81b5-03b3bc50d606\") " pod="openshift-marketplace/redhat-marketplace-dbb29" Nov 25 15:39:38 crc kubenswrapper[4890]: I1125 15:39:38.576007 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c839380c-eefd-4354-81b5-03b3bc50d606-utilities\") pod \"redhat-marketplace-dbb29\" (UID: \"c839380c-eefd-4354-81b5-03b3bc50d606\") " pod="openshift-marketplace/redhat-marketplace-dbb29" Nov 25 15:39:38 crc kubenswrapper[4890]: I1125 15:39:38.576101 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c839380c-eefd-4354-81b5-03b3bc50d606-catalog-content\") pod \"redhat-marketplace-dbb29\" (UID: \"c839380c-eefd-4354-81b5-03b3bc50d606\") " pod="openshift-marketplace/redhat-marketplace-dbb29" Nov 25 15:39:38 crc kubenswrapper[4890]: I1125 15:39:38.606241 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5hwl\" (UniqueName: \"kubernetes.io/projected/c839380c-eefd-4354-81b5-03b3bc50d606-kube-api-access-s5hwl\") pod \"redhat-marketplace-dbb29\" (UID: \"c839380c-eefd-4354-81b5-03b3bc50d606\") " pod="openshift-marketplace/redhat-marketplace-dbb29" Nov 25 15:39:38 crc kubenswrapper[4890]: I1125 15:39:38.678246 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dbb29" Nov 25 15:39:39 crc kubenswrapper[4890]: I1125 15:39:39.231846 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dbb29"] Nov 25 15:39:39 crc kubenswrapper[4890]: I1125 15:39:39.671834 4890 generic.go:334] "Generic (PLEG): container finished" podID="c839380c-eefd-4354-81b5-03b3bc50d606" containerID="fcfd6f7a3355b44a6fdc42b9e5bcaf08781d7df47582b00a767f8a6023a3735c" exitCode=0 Nov 25 15:39:39 crc kubenswrapper[4890]: I1125 15:39:39.671881 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dbb29" event={"ID":"c839380c-eefd-4354-81b5-03b3bc50d606","Type":"ContainerDied","Data":"fcfd6f7a3355b44a6fdc42b9e5bcaf08781d7df47582b00a767f8a6023a3735c"} Nov 25 15:39:39 crc kubenswrapper[4890]: I1125 15:39:39.672202 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dbb29" event={"ID":"c839380c-eefd-4354-81b5-03b3bc50d606","Type":"ContainerStarted","Data":"2687d761d3e8715f016a6345d53396c2b566adb5376b65a52e0d54d5b4e11522"} Nov 25 15:39:41 crc kubenswrapper[4890]: I1125 15:39:41.693572 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dbb29" event={"ID":"c839380c-eefd-4354-81b5-03b3bc50d606","Type":"ContainerStarted","Data":"306590db133433f60ea9fbcfb2196899438c26f323bf28153257e6bc29b93bdb"} Nov 25 15:39:42 crc kubenswrapper[4890]: I1125 15:39:42.181354 4890 scope.go:117] "RemoveContainer" containerID="67ca0de5cd1ef0f73ff70ee2df49e62ca714111ad20a0de1fef88f6f60fdf262" Nov 25 15:39:42 crc kubenswrapper[4890]: E1125 15:39:42.181772 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:39:42 crc kubenswrapper[4890]: I1125 15:39:42.705303 4890 generic.go:334] "Generic (PLEG): container finished" podID="c839380c-eefd-4354-81b5-03b3bc50d606" containerID="306590db133433f60ea9fbcfb2196899438c26f323bf28153257e6bc29b93bdb" exitCode=0 Nov 25 15:39:42 crc kubenswrapper[4890]: I1125 15:39:42.705356 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dbb29" event={"ID":"c839380c-eefd-4354-81b5-03b3bc50d606","Type":"ContainerDied","Data":"306590db133433f60ea9fbcfb2196899438c26f323bf28153257e6bc29b93bdb"} Nov 25 15:39:43 crc kubenswrapper[4890]: I1125 15:39:43.522479 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-52ndf"] Nov 25 15:39:43 crc kubenswrapper[4890]: I1125 15:39:43.525905 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-52ndf" Nov 25 15:39:43 crc kubenswrapper[4890]: I1125 15:39:43.541082 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-52ndf"] Nov 25 15:39:43 crc kubenswrapper[4890]: I1125 15:39:43.579410 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f4fe22b-545b-4fac-a084-5842ef91a897-catalog-content\") pod \"community-operators-52ndf\" (UID: \"8f4fe22b-545b-4fac-a084-5842ef91a897\") " pod="openshift-marketplace/community-operators-52ndf" Nov 25 15:39:43 crc kubenswrapper[4890]: I1125 15:39:43.579642 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htgdv\" (UniqueName: \"kubernetes.io/projected/8f4fe22b-545b-4fac-a084-5842ef91a897-kube-api-access-htgdv\") pod \"community-operators-52ndf\" (UID: \"8f4fe22b-545b-4fac-a084-5842ef91a897\") " pod="openshift-marketplace/community-operators-52ndf" Nov 25 15:39:43 crc kubenswrapper[4890]: I1125 15:39:43.579712 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f4fe22b-545b-4fac-a084-5842ef91a897-utilities\") pod \"community-operators-52ndf\" (UID: \"8f4fe22b-545b-4fac-a084-5842ef91a897\") " pod="openshift-marketplace/community-operators-52ndf" Nov 25 15:39:43 crc kubenswrapper[4890]: I1125 15:39:43.681501 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htgdv\" (UniqueName: \"kubernetes.io/projected/8f4fe22b-545b-4fac-a084-5842ef91a897-kube-api-access-htgdv\") pod \"community-operators-52ndf\" (UID: \"8f4fe22b-545b-4fac-a084-5842ef91a897\") " pod="openshift-marketplace/community-operators-52ndf" Nov 25 15:39:43 crc kubenswrapper[4890]: I1125 15:39:43.681848 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f4fe22b-545b-4fac-a084-5842ef91a897-utilities\") pod \"community-operators-52ndf\" (UID: \"8f4fe22b-545b-4fac-a084-5842ef91a897\") " pod="openshift-marketplace/community-operators-52ndf" Nov 25 15:39:43 crc kubenswrapper[4890]: I1125 15:39:43.682061 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f4fe22b-545b-4fac-a084-5842ef91a897-catalog-content\") pod \"community-operators-52ndf\" (UID: \"8f4fe22b-545b-4fac-a084-5842ef91a897\") " pod="openshift-marketplace/community-operators-52ndf" Nov 25 15:39:43 crc kubenswrapper[4890]: I1125 15:39:43.682358 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f4fe22b-545b-4fac-a084-5842ef91a897-utilities\") pod \"community-operators-52ndf\" (UID: \"8f4fe22b-545b-4fac-a084-5842ef91a897\") " pod="openshift-marketplace/community-operators-52ndf" Nov 25 15:39:43 crc kubenswrapper[4890]: I1125 15:39:43.682424 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f4fe22b-545b-4fac-a084-5842ef91a897-catalog-content\") pod \"community-operators-52ndf\" (UID: \"8f4fe22b-545b-4fac-a084-5842ef91a897\") " pod="openshift-marketplace/community-operators-52ndf" Nov 25 15:39:43 crc kubenswrapper[4890]: I1125 15:39:43.703343 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htgdv\" (UniqueName: \"kubernetes.io/projected/8f4fe22b-545b-4fac-a084-5842ef91a897-kube-api-access-htgdv\") pod \"community-operators-52ndf\" (UID: \"8f4fe22b-545b-4fac-a084-5842ef91a897\") " pod="openshift-marketplace/community-operators-52ndf" Nov 25 15:39:43 crc kubenswrapper[4890]: I1125 15:39:43.716715 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dbb29" event={"ID":"c839380c-eefd-4354-81b5-03b3bc50d606","Type":"ContainerStarted","Data":"74cb15200bef7f1f456164fbc3b67a6deb6b23caca431382e58109eec5d3e322"} Nov 25 15:39:43 crc kubenswrapper[4890]: I1125 15:39:43.847676 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-52ndf" Nov 25 15:39:44 crc kubenswrapper[4890]: I1125 15:39:44.424117 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dbb29" podStartSLOduration=2.887594601 podStartE2EDuration="6.424097043s" podCreationTimestamp="2025-11-25 15:39:38 +0000 UTC" firstStartedPulling="2025-11-25 15:39:39.674424751 +0000 UTC m=+2238.116887361" lastFinishedPulling="2025-11-25 15:39:43.210927193 +0000 UTC m=+2241.653389803" observedRunningTime="2025-11-25 15:39:43.742521388 +0000 UTC m=+2242.184984018" watchObservedRunningTime="2025-11-25 15:39:44.424097043 +0000 UTC m=+2242.866559653" Nov 25 15:39:44 crc kubenswrapper[4890]: I1125 15:39:44.435058 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-52ndf"] Nov 25 15:39:44 crc kubenswrapper[4890]: W1125 15:39:44.437577 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f4fe22b_545b_4fac_a084_5842ef91a897.slice/crio-8ac93240db8260a825f77fd07ebace4ccf24f857ed01a4dd1399107434f586eb WatchSource:0}: Error finding container 8ac93240db8260a825f77fd07ebace4ccf24f857ed01a4dd1399107434f586eb: Status 404 returned error can't find the container with id 8ac93240db8260a825f77fd07ebace4ccf24f857ed01a4dd1399107434f586eb Nov 25 15:39:44 crc kubenswrapper[4890]: I1125 15:39:44.741113 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-52ndf" event={"ID":"8f4fe22b-545b-4fac-a084-5842ef91a897","Type":"ContainerStarted","Data":"e938ecb06d4f1c6cc712d9c0f21c0c0a6d13b89984f2132a8f89ed379435e11d"} Nov 25 15:39:44 crc kubenswrapper[4890]: I1125 15:39:44.741461 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-52ndf" event={"ID":"8f4fe22b-545b-4fac-a084-5842ef91a897","Type":"ContainerStarted","Data":"8ac93240db8260a825f77fd07ebace4ccf24f857ed01a4dd1399107434f586eb"} Nov 25 15:39:45 crc kubenswrapper[4890]: I1125 15:39:45.751896 4890 generic.go:334] "Generic (PLEG): container finished" podID="8f4fe22b-545b-4fac-a084-5842ef91a897" containerID="e938ecb06d4f1c6cc712d9c0f21c0c0a6d13b89984f2132a8f89ed379435e11d" exitCode=0 Nov 25 15:39:45 crc kubenswrapper[4890]: I1125 15:39:45.752328 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-52ndf" event={"ID":"8f4fe22b-545b-4fac-a084-5842ef91a897","Type":"ContainerDied","Data":"e938ecb06d4f1c6cc712d9c0f21c0c0a6d13b89984f2132a8f89ed379435e11d"} Nov 25 15:39:46 crc kubenswrapper[4890]: I1125 15:39:46.764406 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-52ndf" event={"ID":"8f4fe22b-545b-4fac-a084-5842ef91a897","Type":"ContainerStarted","Data":"ca207b50410223feb991ca8688fca1d2afe9faacb0035d2163224dfe1c3e262a"} Nov 25 15:39:47 crc kubenswrapper[4890]: I1125 15:39:47.774946 4890 generic.go:334] "Generic (PLEG): container finished" podID="8f4fe22b-545b-4fac-a084-5842ef91a897" containerID="ca207b50410223feb991ca8688fca1d2afe9faacb0035d2163224dfe1c3e262a" exitCode=0 Nov 25 15:39:47 crc kubenswrapper[4890]: I1125 15:39:47.775004 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-52ndf" event={"ID":"8f4fe22b-545b-4fac-a084-5842ef91a897","Type":"ContainerDied","Data":"ca207b50410223feb991ca8688fca1d2afe9faacb0035d2163224dfe1c3e262a"} Nov 25 15:39:48 crc kubenswrapper[4890]: I1125 15:39:48.679303 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dbb29" Nov 25 15:39:48 crc kubenswrapper[4890]: I1125 15:39:48.680048 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dbb29" Nov 25 15:39:48 crc kubenswrapper[4890]: I1125 15:39:48.728475 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dbb29" Nov 25 15:39:48 crc kubenswrapper[4890]: I1125 15:39:48.791974 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-52ndf" event={"ID":"8f4fe22b-545b-4fac-a084-5842ef91a897","Type":"ContainerStarted","Data":"d4118e0727a084e5b427ddac3acb71ccb134221ae530337bb2542c5a5270fc36"} Nov 25 15:39:48 crc kubenswrapper[4890]: I1125 15:39:48.815244 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-52ndf" podStartSLOduration=3.363981555 podStartE2EDuration="5.81521773s" podCreationTimestamp="2025-11-25 15:39:43 +0000 UTC" firstStartedPulling="2025-11-25 15:39:45.755868383 +0000 UTC m=+2244.198330993" lastFinishedPulling="2025-11-25 15:39:48.207104548 +0000 UTC m=+2246.649567168" observedRunningTime="2025-11-25 15:39:48.813862757 +0000 UTC m=+2247.256325377" watchObservedRunningTime="2025-11-25 15:39:48.81521773 +0000 UTC m=+2247.257680340" Nov 25 15:39:48 crc kubenswrapper[4890]: I1125 15:39:48.843311 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dbb29" Nov 25 15:39:51 crc kubenswrapper[4890]: I1125 15:39:51.110260 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dbb29"] Nov 25 15:39:51 crc kubenswrapper[4890]: I1125 15:39:51.841322 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dbb29" podUID="c839380c-eefd-4354-81b5-03b3bc50d606" containerName="registry-server" containerID="cri-o://74cb15200bef7f1f456164fbc3b67a6deb6b23caca431382e58109eec5d3e322" gracePeriod=2 Nov 25 15:39:52 crc kubenswrapper[4890]: I1125 15:39:52.328498 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dbb29" Nov 25 15:39:52 crc kubenswrapper[4890]: I1125 15:39:52.379514 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c839380c-eefd-4354-81b5-03b3bc50d606-utilities\") pod \"c839380c-eefd-4354-81b5-03b3bc50d606\" (UID: \"c839380c-eefd-4354-81b5-03b3bc50d606\") " Nov 25 15:39:52 crc kubenswrapper[4890]: I1125 15:39:52.379667 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s5hwl\" (UniqueName: \"kubernetes.io/projected/c839380c-eefd-4354-81b5-03b3bc50d606-kube-api-access-s5hwl\") pod \"c839380c-eefd-4354-81b5-03b3bc50d606\" (UID: \"c839380c-eefd-4354-81b5-03b3bc50d606\") " Nov 25 15:39:52 crc kubenswrapper[4890]: I1125 15:39:52.379725 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c839380c-eefd-4354-81b5-03b3bc50d606-catalog-content\") pod \"c839380c-eefd-4354-81b5-03b3bc50d606\" (UID: \"c839380c-eefd-4354-81b5-03b3bc50d606\") " Nov 25 15:39:52 crc kubenswrapper[4890]: I1125 15:39:52.380600 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c839380c-eefd-4354-81b5-03b3bc50d606-utilities" (OuterVolumeSpecName: "utilities") pod "c839380c-eefd-4354-81b5-03b3bc50d606" (UID: "c839380c-eefd-4354-81b5-03b3bc50d606"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:39:52 crc kubenswrapper[4890]: I1125 15:39:52.387635 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c839380c-eefd-4354-81b5-03b3bc50d606-kube-api-access-s5hwl" (OuterVolumeSpecName: "kube-api-access-s5hwl") pod "c839380c-eefd-4354-81b5-03b3bc50d606" (UID: "c839380c-eefd-4354-81b5-03b3bc50d606"). InnerVolumeSpecName "kube-api-access-s5hwl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:39:52 crc kubenswrapper[4890]: I1125 15:39:52.401291 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c839380c-eefd-4354-81b5-03b3bc50d606-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c839380c-eefd-4354-81b5-03b3bc50d606" (UID: "c839380c-eefd-4354-81b5-03b3bc50d606"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:39:52 crc kubenswrapper[4890]: I1125 15:39:52.482514 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s5hwl\" (UniqueName: \"kubernetes.io/projected/c839380c-eefd-4354-81b5-03b3bc50d606-kube-api-access-s5hwl\") on node \"crc\" DevicePath \"\"" Nov 25 15:39:52 crc kubenswrapper[4890]: I1125 15:39:52.482555 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c839380c-eefd-4354-81b5-03b3bc50d606-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 15:39:52 crc kubenswrapper[4890]: I1125 15:39:52.482565 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c839380c-eefd-4354-81b5-03b3bc50d606-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 15:39:52 crc kubenswrapper[4890]: I1125 15:39:52.853109 4890 generic.go:334] "Generic (PLEG): container finished" podID="c839380c-eefd-4354-81b5-03b3bc50d606" containerID="74cb15200bef7f1f456164fbc3b67a6deb6b23caca431382e58109eec5d3e322" exitCode=0 Nov 25 15:39:52 crc kubenswrapper[4890]: I1125 15:39:52.853178 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dbb29" event={"ID":"c839380c-eefd-4354-81b5-03b3bc50d606","Type":"ContainerDied","Data":"74cb15200bef7f1f456164fbc3b67a6deb6b23caca431382e58109eec5d3e322"} Nov 25 15:39:52 crc kubenswrapper[4890]: I1125 15:39:52.853207 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dbb29" event={"ID":"c839380c-eefd-4354-81b5-03b3bc50d606","Type":"ContainerDied","Data":"2687d761d3e8715f016a6345d53396c2b566adb5376b65a52e0d54d5b4e11522"} Nov 25 15:39:52 crc kubenswrapper[4890]: I1125 15:39:52.853227 4890 scope.go:117] "RemoveContainer" containerID="74cb15200bef7f1f456164fbc3b67a6deb6b23caca431382e58109eec5d3e322" Nov 25 15:39:52 crc kubenswrapper[4890]: I1125 15:39:52.853653 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dbb29" Nov 25 15:39:52 crc kubenswrapper[4890]: I1125 15:39:52.887362 4890 scope.go:117] "RemoveContainer" containerID="306590db133433f60ea9fbcfb2196899438c26f323bf28153257e6bc29b93bdb" Nov 25 15:39:52 crc kubenswrapper[4890]: I1125 15:39:52.897378 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dbb29"] Nov 25 15:39:52 crc kubenswrapper[4890]: I1125 15:39:52.908573 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dbb29"] Nov 25 15:39:52 crc kubenswrapper[4890]: I1125 15:39:52.913962 4890 scope.go:117] "RemoveContainer" containerID="fcfd6f7a3355b44a6fdc42b9e5bcaf08781d7df47582b00a767f8a6023a3735c" Nov 25 15:39:52 crc kubenswrapper[4890]: I1125 15:39:52.966469 4890 scope.go:117] "RemoveContainer" containerID="74cb15200bef7f1f456164fbc3b67a6deb6b23caca431382e58109eec5d3e322" Nov 25 15:39:52 crc kubenswrapper[4890]: E1125 15:39:52.967213 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74cb15200bef7f1f456164fbc3b67a6deb6b23caca431382e58109eec5d3e322\": container with ID starting with 74cb15200bef7f1f456164fbc3b67a6deb6b23caca431382e58109eec5d3e322 not found: ID does not exist" containerID="74cb15200bef7f1f456164fbc3b67a6deb6b23caca431382e58109eec5d3e322" Nov 25 15:39:52 crc kubenswrapper[4890]: I1125 15:39:52.967255 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74cb15200bef7f1f456164fbc3b67a6deb6b23caca431382e58109eec5d3e322"} err="failed to get container status \"74cb15200bef7f1f456164fbc3b67a6deb6b23caca431382e58109eec5d3e322\": rpc error: code = NotFound desc = could not find container \"74cb15200bef7f1f456164fbc3b67a6deb6b23caca431382e58109eec5d3e322\": container with ID starting with 74cb15200bef7f1f456164fbc3b67a6deb6b23caca431382e58109eec5d3e322 not found: ID does not exist" Nov 25 15:39:52 crc kubenswrapper[4890]: I1125 15:39:52.967283 4890 scope.go:117] "RemoveContainer" containerID="306590db133433f60ea9fbcfb2196899438c26f323bf28153257e6bc29b93bdb" Nov 25 15:39:52 crc kubenswrapper[4890]: E1125 15:39:52.967808 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"306590db133433f60ea9fbcfb2196899438c26f323bf28153257e6bc29b93bdb\": container with ID starting with 306590db133433f60ea9fbcfb2196899438c26f323bf28153257e6bc29b93bdb not found: ID does not exist" containerID="306590db133433f60ea9fbcfb2196899438c26f323bf28153257e6bc29b93bdb" Nov 25 15:39:52 crc kubenswrapper[4890]: I1125 15:39:52.967867 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"306590db133433f60ea9fbcfb2196899438c26f323bf28153257e6bc29b93bdb"} err="failed to get container status \"306590db133433f60ea9fbcfb2196899438c26f323bf28153257e6bc29b93bdb\": rpc error: code = NotFound desc = could not find container \"306590db133433f60ea9fbcfb2196899438c26f323bf28153257e6bc29b93bdb\": container with ID starting with 306590db133433f60ea9fbcfb2196899438c26f323bf28153257e6bc29b93bdb not found: ID does not exist" Nov 25 15:39:52 crc kubenswrapper[4890]: I1125 15:39:52.967886 4890 scope.go:117] "RemoveContainer" containerID="fcfd6f7a3355b44a6fdc42b9e5bcaf08781d7df47582b00a767f8a6023a3735c" Nov 25 15:39:52 crc kubenswrapper[4890]: E1125 15:39:52.968290 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fcfd6f7a3355b44a6fdc42b9e5bcaf08781d7df47582b00a767f8a6023a3735c\": container with ID starting with fcfd6f7a3355b44a6fdc42b9e5bcaf08781d7df47582b00a767f8a6023a3735c not found: ID does not exist" containerID="fcfd6f7a3355b44a6fdc42b9e5bcaf08781d7df47582b00a767f8a6023a3735c" Nov 25 15:39:52 crc kubenswrapper[4890]: I1125 15:39:52.968316 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fcfd6f7a3355b44a6fdc42b9e5bcaf08781d7df47582b00a767f8a6023a3735c"} err="failed to get container status \"fcfd6f7a3355b44a6fdc42b9e5bcaf08781d7df47582b00a767f8a6023a3735c\": rpc error: code = NotFound desc = could not find container \"fcfd6f7a3355b44a6fdc42b9e5bcaf08781d7df47582b00a767f8a6023a3735c\": container with ID starting with fcfd6f7a3355b44a6fdc42b9e5bcaf08781d7df47582b00a767f8a6023a3735c not found: ID does not exist" Nov 25 15:39:53 crc kubenswrapper[4890]: I1125 15:39:53.848239 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-52ndf" Nov 25 15:39:53 crc kubenswrapper[4890]: I1125 15:39:53.848973 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-52ndf" Nov 25 15:39:53 crc kubenswrapper[4890]: I1125 15:39:53.908585 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-52ndf" Nov 25 15:39:54 crc kubenswrapper[4890]: I1125 15:39:54.173780 4890 scope.go:117] "RemoveContainer" containerID="67ca0de5cd1ef0f73ff70ee2df49e62ca714111ad20a0de1fef88f6f60fdf262" Nov 25 15:39:54 crc kubenswrapper[4890]: E1125 15:39:54.174219 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:39:54 crc kubenswrapper[4890]: I1125 15:39:54.184236 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c839380c-eefd-4354-81b5-03b3bc50d606" path="/var/lib/kubelet/pods/c839380c-eefd-4354-81b5-03b3bc50d606/volumes" Nov 25 15:39:54 crc kubenswrapper[4890]: I1125 15:39:54.961689 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-52ndf" Nov 25 15:39:56 crc kubenswrapper[4890]: I1125 15:39:56.114327 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-52ndf"] Nov 25 15:39:56 crc kubenswrapper[4890]: I1125 15:39:56.921439 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-52ndf" podUID="8f4fe22b-545b-4fac-a084-5842ef91a897" containerName="registry-server" containerID="cri-o://d4118e0727a084e5b427ddac3acb71ccb134221ae530337bb2542c5a5270fc36" gracePeriod=2 Nov 25 15:39:57 crc kubenswrapper[4890]: I1125 15:39:57.399932 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-52ndf" Nov 25 15:39:57 crc kubenswrapper[4890]: I1125 15:39:57.483560 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f4fe22b-545b-4fac-a084-5842ef91a897-catalog-content\") pod \"8f4fe22b-545b-4fac-a084-5842ef91a897\" (UID: \"8f4fe22b-545b-4fac-a084-5842ef91a897\") " Nov 25 15:39:57 crc kubenswrapper[4890]: I1125 15:39:57.483632 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htgdv\" (UniqueName: \"kubernetes.io/projected/8f4fe22b-545b-4fac-a084-5842ef91a897-kube-api-access-htgdv\") pod \"8f4fe22b-545b-4fac-a084-5842ef91a897\" (UID: \"8f4fe22b-545b-4fac-a084-5842ef91a897\") " Nov 25 15:39:57 crc kubenswrapper[4890]: I1125 15:39:57.483834 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f4fe22b-545b-4fac-a084-5842ef91a897-utilities\") pod \"8f4fe22b-545b-4fac-a084-5842ef91a897\" (UID: \"8f4fe22b-545b-4fac-a084-5842ef91a897\") " Nov 25 15:39:57 crc kubenswrapper[4890]: I1125 15:39:57.485010 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f4fe22b-545b-4fac-a084-5842ef91a897-utilities" (OuterVolumeSpecName: "utilities") pod "8f4fe22b-545b-4fac-a084-5842ef91a897" (UID: "8f4fe22b-545b-4fac-a084-5842ef91a897"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:39:57 crc kubenswrapper[4890]: I1125 15:39:57.490982 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f4fe22b-545b-4fac-a084-5842ef91a897-kube-api-access-htgdv" (OuterVolumeSpecName: "kube-api-access-htgdv") pod "8f4fe22b-545b-4fac-a084-5842ef91a897" (UID: "8f4fe22b-545b-4fac-a084-5842ef91a897"). InnerVolumeSpecName "kube-api-access-htgdv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:39:57 crc kubenswrapper[4890]: I1125 15:39:57.556479 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f4fe22b-545b-4fac-a084-5842ef91a897-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8f4fe22b-545b-4fac-a084-5842ef91a897" (UID: "8f4fe22b-545b-4fac-a084-5842ef91a897"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:39:57 crc kubenswrapper[4890]: I1125 15:39:57.586946 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f4fe22b-545b-4fac-a084-5842ef91a897-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 15:39:57 crc kubenswrapper[4890]: I1125 15:39:57.586994 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htgdv\" (UniqueName: \"kubernetes.io/projected/8f4fe22b-545b-4fac-a084-5842ef91a897-kube-api-access-htgdv\") on node \"crc\" DevicePath \"\"" Nov 25 15:39:57 crc kubenswrapper[4890]: I1125 15:39:57.587009 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f4fe22b-545b-4fac-a084-5842ef91a897-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 15:39:57 crc kubenswrapper[4890]: I1125 15:39:57.936427 4890 generic.go:334] "Generic (PLEG): container finished" podID="8f4fe22b-545b-4fac-a084-5842ef91a897" containerID="d4118e0727a084e5b427ddac3acb71ccb134221ae530337bb2542c5a5270fc36" exitCode=0 Nov 25 15:39:57 crc kubenswrapper[4890]: I1125 15:39:57.936491 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-52ndf" event={"ID":"8f4fe22b-545b-4fac-a084-5842ef91a897","Type":"ContainerDied","Data":"d4118e0727a084e5b427ddac3acb71ccb134221ae530337bb2542c5a5270fc36"} Nov 25 15:39:57 crc kubenswrapper[4890]: I1125 15:39:57.936523 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-52ndf" event={"ID":"8f4fe22b-545b-4fac-a084-5842ef91a897","Type":"ContainerDied","Data":"8ac93240db8260a825f77fd07ebace4ccf24f857ed01a4dd1399107434f586eb"} Nov 25 15:39:57 crc kubenswrapper[4890]: I1125 15:39:57.936541 4890 scope.go:117] "RemoveContainer" containerID="d4118e0727a084e5b427ddac3acb71ccb134221ae530337bb2542c5a5270fc36" Nov 25 15:39:57 crc kubenswrapper[4890]: I1125 15:39:57.936656 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-52ndf" Nov 25 15:39:57 crc kubenswrapper[4890]: I1125 15:39:57.972272 4890 scope.go:117] "RemoveContainer" containerID="ca207b50410223feb991ca8688fca1d2afe9faacb0035d2163224dfe1c3e262a" Nov 25 15:39:57 crc kubenswrapper[4890]: I1125 15:39:57.983089 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-52ndf"] Nov 25 15:39:57 crc kubenswrapper[4890]: I1125 15:39:57.992845 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-52ndf"] Nov 25 15:39:58 crc kubenswrapper[4890]: I1125 15:39:58.004660 4890 scope.go:117] "RemoveContainer" containerID="e938ecb06d4f1c6cc712d9c0f21c0c0a6d13b89984f2132a8f89ed379435e11d" Nov 25 15:39:58 crc kubenswrapper[4890]: I1125 15:39:58.069129 4890 scope.go:117] "RemoveContainer" containerID="d4118e0727a084e5b427ddac3acb71ccb134221ae530337bb2542c5a5270fc36" Nov 25 15:39:58 crc kubenswrapper[4890]: E1125 15:39:58.070589 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4118e0727a084e5b427ddac3acb71ccb134221ae530337bb2542c5a5270fc36\": container with ID starting with d4118e0727a084e5b427ddac3acb71ccb134221ae530337bb2542c5a5270fc36 not found: ID does not exist" containerID="d4118e0727a084e5b427ddac3acb71ccb134221ae530337bb2542c5a5270fc36" Nov 25 15:39:58 crc kubenswrapper[4890]: I1125 15:39:58.070669 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4118e0727a084e5b427ddac3acb71ccb134221ae530337bb2542c5a5270fc36"} err="failed to get container status \"d4118e0727a084e5b427ddac3acb71ccb134221ae530337bb2542c5a5270fc36\": rpc error: code = NotFound desc = could not find container \"d4118e0727a084e5b427ddac3acb71ccb134221ae530337bb2542c5a5270fc36\": container with ID starting with d4118e0727a084e5b427ddac3acb71ccb134221ae530337bb2542c5a5270fc36 not found: ID does not exist" Nov 25 15:39:58 crc kubenswrapper[4890]: I1125 15:39:58.070746 4890 scope.go:117] "RemoveContainer" containerID="ca207b50410223feb991ca8688fca1d2afe9faacb0035d2163224dfe1c3e262a" Nov 25 15:39:58 crc kubenswrapper[4890]: E1125 15:39:58.071744 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca207b50410223feb991ca8688fca1d2afe9faacb0035d2163224dfe1c3e262a\": container with ID starting with ca207b50410223feb991ca8688fca1d2afe9faacb0035d2163224dfe1c3e262a not found: ID does not exist" containerID="ca207b50410223feb991ca8688fca1d2afe9faacb0035d2163224dfe1c3e262a" Nov 25 15:39:58 crc kubenswrapper[4890]: I1125 15:39:58.071798 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca207b50410223feb991ca8688fca1d2afe9faacb0035d2163224dfe1c3e262a"} err="failed to get container status \"ca207b50410223feb991ca8688fca1d2afe9faacb0035d2163224dfe1c3e262a\": rpc error: code = NotFound desc = could not find container \"ca207b50410223feb991ca8688fca1d2afe9faacb0035d2163224dfe1c3e262a\": container with ID starting with ca207b50410223feb991ca8688fca1d2afe9faacb0035d2163224dfe1c3e262a not found: ID does not exist" Nov 25 15:39:58 crc kubenswrapper[4890]: I1125 15:39:58.071889 4890 scope.go:117] "RemoveContainer" containerID="e938ecb06d4f1c6cc712d9c0f21c0c0a6d13b89984f2132a8f89ed379435e11d" Nov 25 15:39:58 crc kubenswrapper[4890]: E1125 15:39:58.072611 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e938ecb06d4f1c6cc712d9c0f21c0c0a6d13b89984f2132a8f89ed379435e11d\": container with ID starting with e938ecb06d4f1c6cc712d9c0f21c0c0a6d13b89984f2132a8f89ed379435e11d not found: ID does not exist" containerID="e938ecb06d4f1c6cc712d9c0f21c0c0a6d13b89984f2132a8f89ed379435e11d" Nov 25 15:39:58 crc kubenswrapper[4890]: I1125 15:39:58.072691 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e938ecb06d4f1c6cc712d9c0f21c0c0a6d13b89984f2132a8f89ed379435e11d"} err="failed to get container status \"e938ecb06d4f1c6cc712d9c0f21c0c0a6d13b89984f2132a8f89ed379435e11d\": rpc error: code = NotFound desc = could not find container \"e938ecb06d4f1c6cc712d9c0f21c0c0a6d13b89984f2132a8f89ed379435e11d\": container with ID starting with e938ecb06d4f1c6cc712d9c0f21c0c0a6d13b89984f2132a8f89ed379435e11d not found: ID does not exist" Nov 25 15:39:58 crc kubenswrapper[4890]: I1125 15:39:58.186223 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f4fe22b-545b-4fac-a084-5842ef91a897" path="/var/lib/kubelet/pods/8f4fe22b-545b-4fac-a084-5842ef91a897/volumes" Nov 25 15:40:04 crc kubenswrapper[4890]: I1125 15:40:04.032926 4890 generic.go:334] "Generic (PLEG): container finished" podID="995132c2-f199-4d15-b3ef-3c71fed454f5" containerID="bba15ff5f74e18952e2cae0e9d8c72f4c71326f8c3c783422f1b6716845e7a6d" exitCode=0 Nov 25 15:40:04 crc kubenswrapper[4890]: I1125 15:40:04.033025 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-llz86" event={"ID":"995132c2-f199-4d15-b3ef-3c71fed454f5","Type":"ContainerDied","Data":"bba15ff5f74e18952e2cae0e9d8c72f4c71326f8c3c783422f1b6716845e7a6d"} Nov 25 15:40:05 crc kubenswrapper[4890]: I1125 15:40:05.433799 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-llz86" Nov 25 15:40:05 crc kubenswrapper[4890]: I1125 15:40:05.562886 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/995132c2-f199-4d15-b3ef-3c71fed454f5-ovn-combined-ca-bundle\") pod \"995132c2-f199-4d15-b3ef-3c71fed454f5\" (UID: \"995132c2-f199-4d15-b3ef-3c71fed454f5\") " Nov 25 15:40:05 crc kubenswrapper[4890]: I1125 15:40:05.563344 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/995132c2-f199-4d15-b3ef-3c71fed454f5-ssh-key\") pod \"995132c2-f199-4d15-b3ef-3c71fed454f5\" (UID: \"995132c2-f199-4d15-b3ef-3c71fed454f5\") " Nov 25 15:40:05 crc kubenswrapper[4890]: I1125 15:40:05.563498 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/995132c2-f199-4d15-b3ef-3c71fed454f5-inventory\") pod \"995132c2-f199-4d15-b3ef-3c71fed454f5\" (UID: \"995132c2-f199-4d15-b3ef-3c71fed454f5\") " Nov 25 15:40:05 crc kubenswrapper[4890]: I1125 15:40:05.563701 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/995132c2-f199-4d15-b3ef-3c71fed454f5-ovncontroller-config-0\") pod \"995132c2-f199-4d15-b3ef-3c71fed454f5\" (UID: \"995132c2-f199-4d15-b3ef-3c71fed454f5\") " Nov 25 15:40:05 crc kubenswrapper[4890]: I1125 15:40:05.563743 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t44ql\" (UniqueName: \"kubernetes.io/projected/995132c2-f199-4d15-b3ef-3c71fed454f5-kube-api-access-t44ql\") pod \"995132c2-f199-4d15-b3ef-3c71fed454f5\" (UID: \"995132c2-f199-4d15-b3ef-3c71fed454f5\") " Nov 25 15:40:05 crc kubenswrapper[4890]: I1125 15:40:05.569902 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/995132c2-f199-4d15-b3ef-3c71fed454f5-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "995132c2-f199-4d15-b3ef-3c71fed454f5" (UID: "995132c2-f199-4d15-b3ef-3c71fed454f5"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:40:05 crc kubenswrapper[4890]: I1125 15:40:05.574063 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/995132c2-f199-4d15-b3ef-3c71fed454f5-kube-api-access-t44ql" (OuterVolumeSpecName: "kube-api-access-t44ql") pod "995132c2-f199-4d15-b3ef-3c71fed454f5" (UID: "995132c2-f199-4d15-b3ef-3c71fed454f5"). InnerVolumeSpecName "kube-api-access-t44ql". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:40:05 crc kubenswrapper[4890]: I1125 15:40:05.594406 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/995132c2-f199-4d15-b3ef-3c71fed454f5-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "995132c2-f199-4d15-b3ef-3c71fed454f5" (UID: "995132c2-f199-4d15-b3ef-3c71fed454f5"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:40:05 crc kubenswrapper[4890]: I1125 15:40:05.594616 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/995132c2-f199-4d15-b3ef-3c71fed454f5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "995132c2-f199-4d15-b3ef-3c71fed454f5" (UID: "995132c2-f199-4d15-b3ef-3c71fed454f5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:40:05 crc kubenswrapper[4890]: I1125 15:40:05.594888 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/995132c2-f199-4d15-b3ef-3c71fed454f5-inventory" (OuterVolumeSpecName: "inventory") pod "995132c2-f199-4d15-b3ef-3c71fed454f5" (UID: "995132c2-f199-4d15-b3ef-3c71fed454f5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:40:05 crc kubenswrapper[4890]: I1125 15:40:05.665581 4890 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/995132c2-f199-4d15-b3ef-3c71fed454f5-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Nov 25 15:40:05 crc kubenswrapper[4890]: I1125 15:40:05.665639 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t44ql\" (UniqueName: \"kubernetes.io/projected/995132c2-f199-4d15-b3ef-3c71fed454f5-kube-api-access-t44ql\") on node \"crc\" DevicePath \"\"" Nov 25 15:40:05 crc kubenswrapper[4890]: I1125 15:40:05.665655 4890 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/995132c2-f199-4d15-b3ef-3c71fed454f5-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:40:05 crc kubenswrapper[4890]: I1125 15:40:05.665668 4890 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/995132c2-f199-4d15-b3ef-3c71fed454f5-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 15:40:05 crc kubenswrapper[4890]: I1125 15:40:05.665680 4890 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/995132c2-f199-4d15-b3ef-3c71fed454f5-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 15:40:06 crc kubenswrapper[4890]: I1125 15:40:06.057986 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-llz86" event={"ID":"995132c2-f199-4d15-b3ef-3c71fed454f5","Type":"ContainerDied","Data":"3ef2357de5ab5029106cba56ef25c2288bd83ebc7e652cbb6c5b5e77347f4429"} Nov 25 15:40:06 crc kubenswrapper[4890]: I1125 15:40:06.058042 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3ef2357de5ab5029106cba56ef25c2288bd83ebc7e652cbb6c5b5e77347f4429" Nov 25 15:40:06 crc kubenswrapper[4890]: I1125 15:40:06.058114 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-llz86" Nov 25 15:40:06 crc kubenswrapper[4890]: I1125 15:40:06.169420 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dvxc6"] Nov 25 15:40:06 crc kubenswrapper[4890]: E1125 15:40:06.170054 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c839380c-eefd-4354-81b5-03b3bc50d606" containerName="extract-utilities" Nov 25 15:40:06 crc kubenswrapper[4890]: I1125 15:40:06.170087 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="c839380c-eefd-4354-81b5-03b3bc50d606" containerName="extract-utilities" Nov 25 15:40:06 crc kubenswrapper[4890]: E1125 15:40:06.170133 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c839380c-eefd-4354-81b5-03b3bc50d606" containerName="registry-server" Nov 25 15:40:06 crc kubenswrapper[4890]: I1125 15:40:06.170142 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="c839380c-eefd-4354-81b5-03b3bc50d606" containerName="registry-server" Nov 25 15:40:06 crc kubenswrapper[4890]: E1125 15:40:06.170183 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="995132c2-f199-4d15-b3ef-3c71fed454f5" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 25 15:40:06 crc kubenswrapper[4890]: I1125 15:40:06.170190 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="995132c2-f199-4d15-b3ef-3c71fed454f5" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 25 15:40:06 crc kubenswrapper[4890]: E1125 15:40:06.170214 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f4fe22b-545b-4fac-a084-5842ef91a897" containerName="registry-server" Nov 25 15:40:06 crc kubenswrapper[4890]: I1125 15:40:06.170220 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f4fe22b-545b-4fac-a084-5842ef91a897" containerName="registry-server" Nov 25 15:40:06 crc kubenswrapper[4890]: E1125 15:40:06.170238 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f4fe22b-545b-4fac-a084-5842ef91a897" containerName="extract-utilities" Nov 25 15:40:06 crc kubenswrapper[4890]: I1125 15:40:06.170252 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f4fe22b-545b-4fac-a084-5842ef91a897" containerName="extract-utilities" Nov 25 15:40:06 crc kubenswrapper[4890]: E1125 15:40:06.170270 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f4fe22b-545b-4fac-a084-5842ef91a897" containerName="extract-content" Nov 25 15:40:06 crc kubenswrapper[4890]: I1125 15:40:06.170278 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f4fe22b-545b-4fac-a084-5842ef91a897" containerName="extract-content" Nov 25 15:40:06 crc kubenswrapper[4890]: E1125 15:40:06.170293 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c839380c-eefd-4354-81b5-03b3bc50d606" containerName="extract-content" Nov 25 15:40:06 crc kubenswrapper[4890]: I1125 15:40:06.170301 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="c839380c-eefd-4354-81b5-03b3bc50d606" containerName="extract-content" Nov 25 15:40:06 crc kubenswrapper[4890]: I1125 15:40:06.170577 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f4fe22b-545b-4fac-a084-5842ef91a897" containerName="registry-server" Nov 25 15:40:06 crc kubenswrapper[4890]: I1125 15:40:06.170603 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="c839380c-eefd-4354-81b5-03b3bc50d606" containerName="registry-server" Nov 25 15:40:06 crc kubenswrapper[4890]: I1125 15:40:06.170631 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="995132c2-f199-4d15-b3ef-3c71fed454f5" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 25 15:40:06 crc kubenswrapper[4890]: I1125 15:40:06.171696 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dvxc6" Nov 25 15:40:06 crc kubenswrapper[4890]: I1125 15:40:06.176768 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Nov 25 15:40:06 crc kubenswrapper[4890]: I1125 15:40:06.176939 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 15:40:06 crc kubenswrapper[4890]: I1125 15:40:06.177092 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-22zzp" Nov 25 15:40:06 crc kubenswrapper[4890]: I1125 15:40:06.176772 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 15:40:06 crc kubenswrapper[4890]: I1125 15:40:06.176837 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Nov 25 15:40:06 crc kubenswrapper[4890]: I1125 15:40:06.181509 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 15:40:06 crc kubenswrapper[4890]: I1125 15:40:06.196804 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dvxc6"] Nov 25 15:40:06 crc kubenswrapper[4890]: I1125 15:40:06.280456 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cb4c455d-9d31-430d-b225-9bdb0b9690e0-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-dvxc6\" (UID: \"cb4c455d-9d31-430d-b225-9bdb0b9690e0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dvxc6" Nov 25 15:40:06 crc kubenswrapper[4890]: I1125 15:40:06.280573 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb4c455d-9d31-430d-b225-9bdb0b9690e0-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-dvxc6\" (UID: \"cb4c455d-9d31-430d-b225-9bdb0b9690e0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dvxc6" Nov 25 15:40:06 crc kubenswrapper[4890]: I1125 15:40:06.280620 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cb4c455d-9d31-430d-b225-9bdb0b9690e0-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-dvxc6\" (UID: \"cb4c455d-9d31-430d-b225-9bdb0b9690e0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dvxc6" Nov 25 15:40:06 crc kubenswrapper[4890]: I1125 15:40:06.280655 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8px4q\" (UniqueName: \"kubernetes.io/projected/cb4c455d-9d31-430d-b225-9bdb0b9690e0-kube-api-access-8px4q\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-dvxc6\" (UID: \"cb4c455d-9d31-430d-b225-9bdb0b9690e0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dvxc6" Nov 25 15:40:06 crc kubenswrapper[4890]: I1125 15:40:06.280685 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cb4c455d-9d31-430d-b225-9bdb0b9690e0-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-dvxc6\" (UID: \"cb4c455d-9d31-430d-b225-9bdb0b9690e0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dvxc6" Nov 25 15:40:06 crc kubenswrapper[4890]: I1125 15:40:06.280734 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cb4c455d-9d31-430d-b225-9bdb0b9690e0-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-dvxc6\" (UID: \"cb4c455d-9d31-430d-b225-9bdb0b9690e0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dvxc6" Nov 25 15:40:06 crc kubenswrapper[4890]: I1125 15:40:06.382204 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb4c455d-9d31-430d-b225-9bdb0b9690e0-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-dvxc6\" (UID: \"cb4c455d-9d31-430d-b225-9bdb0b9690e0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dvxc6" Nov 25 15:40:06 crc kubenswrapper[4890]: I1125 15:40:06.382294 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cb4c455d-9d31-430d-b225-9bdb0b9690e0-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-dvxc6\" (UID: \"cb4c455d-9d31-430d-b225-9bdb0b9690e0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dvxc6" Nov 25 15:40:06 crc kubenswrapper[4890]: I1125 15:40:06.382318 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8px4q\" (UniqueName: \"kubernetes.io/projected/cb4c455d-9d31-430d-b225-9bdb0b9690e0-kube-api-access-8px4q\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-dvxc6\" (UID: \"cb4c455d-9d31-430d-b225-9bdb0b9690e0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dvxc6" Nov 25 15:40:06 crc kubenswrapper[4890]: I1125 15:40:06.382368 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cb4c455d-9d31-430d-b225-9bdb0b9690e0-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-dvxc6\" (UID: \"cb4c455d-9d31-430d-b225-9bdb0b9690e0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dvxc6" Nov 25 15:40:06 crc kubenswrapper[4890]: I1125 15:40:06.382436 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cb4c455d-9d31-430d-b225-9bdb0b9690e0-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-dvxc6\" (UID: \"cb4c455d-9d31-430d-b225-9bdb0b9690e0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dvxc6" Nov 25 15:40:06 crc kubenswrapper[4890]: I1125 15:40:06.382503 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cb4c455d-9d31-430d-b225-9bdb0b9690e0-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-dvxc6\" (UID: \"cb4c455d-9d31-430d-b225-9bdb0b9690e0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dvxc6" Nov 25 15:40:06 crc kubenswrapper[4890]: I1125 15:40:06.387070 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cb4c455d-9d31-430d-b225-9bdb0b9690e0-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-dvxc6\" (UID: \"cb4c455d-9d31-430d-b225-9bdb0b9690e0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dvxc6" Nov 25 15:40:06 crc kubenswrapper[4890]: I1125 15:40:06.387224 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cb4c455d-9d31-430d-b225-9bdb0b9690e0-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-dvxc6\" (UID: \"cb4c455d-9d31-430d-b225-9bdb0b9690e0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dvxc6" Nov 25 15:40:06 crc kubenswrapper[4890]: I1125 15:40:06.387293 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb4c455d-9d31-430d-b225-9bdb0b9690e0-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-dvxc6\" (UID: \"cb4c455d-9d31-430d-b225-9bdb0b9690e0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dvxc6" Nov 25 15:40:06 crc kubenswrapper[4890]: I1125 15:40:06.387716 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cb4c455d-9d31-430d-b225-9bdb0b9690e0-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-dvxc6\" (UID: \"cb4c455d-9d31-430d-b225-9bdb0b9690e0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dvxc6" Nov 25 15:40:06 crc kubenswrapper[4890]: I1125 15:40:06.387946 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cb4c455d-9d31-430d-b225-9bdb0b9690e0-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-dvxc6\" (UID: \"cb4c455d-9d31-430d-b225-9bdb0b9690e0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dvxc6" Nov 25 15:40:06 crc kubenswrapper[4890]: I1125 15:40:06.406578 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8px4q\" (UniqueName: \"kubernetes.io/projected/cb4c455d-9d31-430d-b225-9bdb0b9690e0-kube-api-access-8px4q\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-dvxc6\" (UID: \"cb4c455d-9d31-430d-b225-9bdb0b9690e0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dvxc6" Nov 25 15:40:06 crc kubenswrapper[4890]: I1125 15:40:06.507361 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dvxc6" Nov 25 15:40:07 crc kubenswrapper[4890]: I1125 15:40:07.031666 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dvxc6"] Nov 25 15:40:07 crc kubenswrapper[4890]: I1125 15:40:07.067754 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dvxc6" event={"ID":"cb4c455d-9d31-430d-b225-9bdb0b9690e0","Type":"ContainerStarted","Data":"f2075edab5ef5cfb1d38e9bf6ab42928b3ffed289e36631f2e927ab5126f0833"} Nov 25 15:40:08 crc kubenswrapper[4890]: I1125 15:40:08.077518 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dvxc6" event={"ID":"cb4c455d-9d31-430d-b225-9bdb0b9690e0","Type":"ContainerStarted","Data":"6ecb52d4ccee55537ab49aa9ef94fabda51c27749ade90f91b83602cac666aaa"} Nov 25 15:40:08 crc kubenswrapper[4890]: I1125 15:40:08.098691 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dvxc6" podStartSLOduration=1.644180974 podStartE2EDuration="2.098671257s" podCreationTimestamp="2025-11-25 15:40:06 +0000 UTC" firstStartedPulling="2025-11-25 15:40:07.03519134 +0000 UTC m=+2265.477653940" lastFinishedPulling="2025-11-25 15:40:07.489681613 +0000 UTC m=+2265.932144223" observedRunningTime="2025-11-25 15:40:08.094418354 +0000 UTC m=+2266.536880964" watchObservedRunningTime="2025-11-25 15:40:08.098671257 +0000 UTC m=+2266.541133867" Nov 25 15:40:08 crc kubenswrapper[4890]: I1125 15:40:08.173662 4890 scope.go:117] "RemoveContainer" containerID="67ca0de5cd1ef0f73ff70ee2df49e62ca714111ad20a0de1fef88f6f60fdf262" Nov 25 15:40:08 crc kubenswrapper[4890]: E1125 15:40:08.173996 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:40:21 crc kubenswrapper[4890]: I1125 15:40:21.173577 4890 scope.go:117] "RemoveContainer" containerID="67ca0de5cd1ef0f73ff70ee2df49e62ca714111ad20a0de1fef88f6f60fdf262" Nov 25 15:40:21 crc kubenswrapper[4890]: E1125 15:40:21.174881 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:40:35 crc kubenswrapper[4890]: I1125 15:40:35.173971 4890 scope.go:117] "RemoveContainer" containerID="67ca0de5cd1ef0f73ff70ee2df49e62ca714111ad20a0de1fef88f6f60fdf262" Nov 25 15:40:35 crc kubenswrapper[4890]: E1125 15:40:35.175070 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:40:47 crc kubenswrapper[4890]: I1125 15:40:47.173878 4890 scope.go:117] "RemoveContainer" containerID="67ca0de5cd1ef0f73ff70ee2df49e62ca714111ad20a0de1fef88f6f60fdf262" Nov 25 15:40:47 crc kubenswrapper[4890]: E1125 15:40:47.175483 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:40:52 crc kubenswrapper[4890]: I1125 15:40:52.499050 4890 generic.go:334] "Generic (PLEG): container finished" podID="cb4c455d-9d31-430d-b225-9bdb0b9690e0" containerID="6ecb52d4ccee55537ab49aa9ef94fabda51c27749ade90f91b83602cac666aaa" exitCode=0 Nov 25 15:40:52 crc kubenswrapper[4890]: I1125 15:40:52.500143 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dvxc6" event={"ID":"cb4c455d-9d31-430d-b225-9bdb0b9690e0","Type":"ContainerDied","Data":"6ecb52d4ccee55537ab49aa9ef94fabda51c27749ade90f91b83602cac666aaa"} Nov 25 15:40:53 crc kubenswrapper[4890]: I1125 15:40:53.961770 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dvxc6" Nov 25 15:40:54 crc kubenswrapper[4890]: I1125 15:40:54.028353 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb4c455d-9d31-430d-b225-9bdb0b9690e0-neutron-metadata-combined-ca-bundle\") pod \"cb4c455d-9d31-430d-b225-9bdb0b9690e0\" (UID: \"cb4c455d-9d31-430d-b225-9bdb0b9690e0\") " Nov 25 15:40:54 crc kubenswrapper[4890]: I1125 15:40:54.028514 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cb4c455d-9d31-430d-b225-9bdb0b9690e0-neutron-ovn-metadata-agent-neutron-config-0\") pod \"cb4c455d-9d31-430d-b225-9bdb0b9690e0\" (UID: \"cb4c455d-9d31-430d-b225-9bdb0b9690e0\") " Nov 25 15:40:54 crc kubenswrapper[4890]: I1125 15:40:54.028609 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cb4c455d-9d31-430d-b225-9bdb0b9690e0-nova-metadata-neutron-config-0\") pod \"cb4c455d-9d31-430d-b225-9bdb0b9690e0\" (UID: \"cb4c455d-9d31-430d-b225-9bdb0b9690e0\") " Nov 25 15:40:54 crc kubenswrapper[4890]: I1125 15:40:54.028692 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8px4q\" (UniqueName: \"kubernetes.io/projected/cb4c455d-9d31-430d-b225-9bdb0b9690e0-kube-api-access-8px4q\") pod \"cb4c455d-9d31-430d-b225-9bdb0b9690e0\" (UID: \"cb4c455d-9d31-430d-b225-9bdb0b9690e0\") " Nov 25 15:40:54 crc kubenswrapper[4890]: I1125 15:40:54.028808 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cb4c455d-9d31-430d-b225-9bdb0b9690e0-ssh-key\") pod \"cb4c455d-9d31-430d-b225-9bdb0b9690e0\" (UID: \"cb4c455d-9d31-430d-b225-9bdb0b9690e0\") " Nov 25 15:40:54 crc kubenswrapper[4890]: I1125 15:40:54.029657 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cb4c455d-9d31-430d-b225-9bdb0b9690e0-inventory\") pod \"cb4c455d-9d31-430d-b225-9bdb0b9690e0\" (UID: \"cb4c455d-9d31-430d-b225-9bdb0b9690e0\") " Nov 25 15:40:54 crc kubenswrapper[4890]: I1125 15:40:54.035958 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb4c455d-9d31-430d-b225-9bdb0b9690e0-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "cb4c455d-9d31-430d-b225-9bdb0b9690e0" (UID: "cb4c455d-9d31-430d-b225-9bdb0b9690e0"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:40:54 crc kubenswrapper[4890]: I1125 15:40:54.036880 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb4c455d-9d31-430d-b225-9bdb0b9690e0-kube-api-access-8px4q" (OuterVolumeSpecName: "kube-api-access-8px4q") pod "cb4c455d-9d31-430d-b225-9bdb0b9690e0" (UID: "cb4c455d-9d31-430d-b225-9bdb0b9690e0"). InnerVolumeSpecName "kube-api-access-8px4q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:40:54 crc kubenswrapper[4890]: I1125 15:40:54.061586 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb4c455d-9d31-430d-b225-9bdb0b9690e0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "cb4c455d-9d31-430d-b225-9bdb0b9690e0" (UID: "cb4c455d-9d31-430d-b225-9bdb0b9690e0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:40:54 crc kubenswrapper[4890]: I1125 15:40:54.064093 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb4c455d-9d31-430d-b225-9bdb0b9690e0-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "cb4c455d-9d31-430d-b225-9bdb0b9690e0" (UID: "cb4c455d-9d31-430d-b225-9bdb0b9690e0"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:40:54 crc kubenswrapper[4890]: I1125 15:40:54.066129 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb4c455d-9d31-430d-b225-9bdb0b9690e0-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "cb4c455d-9d31-430d-b225-9bdb0b9690e0" (UID: "cb4c455d-9d31-430d-b225-9bdb0b9690e0"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:40:54 crc kubenswrapper[4890]: I1125 15:40:54.074063 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb4c455d-9d31-430d-b225-9bdb0b9690e0-inventory" (OuterVolumeSpecName: "inventory") pod "cb4c455d-9d31-430d-b225-9bdb0b9690e0" (UID: "cb4c455d-9d31-430d-b225-9bdb0b9690e0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:40:54 crc kubenswrapper[4890]: I1125 15:40:54.132140 4890 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cb4c455d-9d31-430d-b225-9bdb0b9690e0-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 25 15:40:54 crc kubenswrapper[4890]: I1125 15:40:54.132188 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8px4q\" (UniqueName: \"kubernetes.io/projected/cb4c455d-9d31-430d-b225-9bdb0b9690e0-kube-api-access-8px4q\") on node \"crc\" DevicePath \"\"" Nov 25 15:40:54 crc kubenswrapper[4890]: I1125 15:40:54.132199 4890 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cb4c455d-9d31-430d-b225-9bdb0b9690e0-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 15:40:54 crc kubenswrapper[4890]: I1125 15:40:54.132208 4890 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cb4c455d-9d31-430d-b225-9bdb0b9690e0-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 15:40:54 crc kubenswrapper[4890]: I1125 15:40:54.132218 4890 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb4c455d-9d31-430d-b225-9bdb0b9690e0-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:40:54 crc kubenswrapper[4890]: I1125 15:40:54.132229 4890 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cb4c455d-9d31-430d-b225-9bdb0b9690e0-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 25 15:40:54 crc kubenswrapper[4890]: I1125 15:40:54.525331 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dvxc6" event={"ID":"cb4c455d-9d31-430d-b225-9bdb0b9690e0","Type":"ContainerDied","Data":"f2075edab5ef5cfb1d38e9bf6ab42928b3ffed289e36631f2e927ab5126f0833"} Nov 25 15:40:54 crc kubenswrapper[4890]: I1125 15:40:54.525848 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f2075edab5ef5cfb1d38e9bf6ab42928b3ffed289e36631f2e927ab5126f0833" Nov 25 15:40:54 crc kubenswrapper[4890]: I1125 15:40:54.525416 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dvxc6" Nov 25 15:40:54 crc kubenswrapper[4890]: I1125 15:40:54.644256 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4fkzh"] Nov 25 15:40:54 crc kubenswrapper[4890]: E1125 15:40:54.665867 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb4c455d-9d31-430d-b225-9bdb0b9690e0" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 25 15:40:54 crc kubenswrapper[4890]: I1125 15:40:54.665970 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb4c455d-9d31-430d-b225-9bdb0b9690e0" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 25 15:40:54 crc kubenswrapper[4890]: I1125 15:40:54.666717 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb4c455d-9d31-430d-b225-9bdb0b9690e0" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 25 15:40:54 crc kubenswrapper[4890]: I1125 15:40:54.668263 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4fkzh" Nov 25 15:40:54 crc kubenswrapper[4890]: I1125 15:40:54.672651 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 15:40:54 crc kubenswrapper[4890]: I1125 15:40:54.674822 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Nov 25 15:40:54 crc kubenswrapper[4890]: I1125 15:40:54.675137 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-22zzp" Nov 25 15:40:54 crc kubenswrapper[4890]: I1125 15:40:54.676653 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 15:40:54 crc kubenswrapper[4890]: I1125 15:40:54.677125 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 15:40:54 crc kubenswrapper[4890]: I1125 15:40:54.701898 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4fkzh"] Nov 25 15:40:54 crc kubenswrapper[4890]: I1125 15:40:54.744810 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpvqs\" (UniqueName: \"kubernetes.io/projected/ca9e616d-0ce2-4784-945c-9e99ba4f437b-kube-api-access-cpvqs\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4fkzh\" (UID: \"ca9e616d-0ce2-4784-945c-9e99ba4f437b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4fkzh" Nov 25 15:40:54 crc kubenswrapper[4890]: I1125 15:40:54.744908 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ca9e616d-0ce2-4784-945c-9e99ba4f437b-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4fkzh\" (UID: \"ca9e616d-0ce2-4784-945c-9e99ba4f437b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4fkzh" Nov 25 15:40:54 crc kubenswrapper[4890]: I1125 15:40:54.745026 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/ca9e616d-0ce2-4784-945c-9e99ba4f437b-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4fkzh\" (UID: \"ca9e616d-0ce2-4784-945c-9e99ba4f437b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4fkzh" Nov 25 15:40:54 crc kubenswrapper[4890]: I1125 15:40:54.745060 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca9e616d-0ce2-4784-945c-9e99ba4f437b-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4fkzh\" (UID: \"ca9e616d-0ce2-4784-945c-9e99ba4f437b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4fkzh" Nov 25 15:40:54 crc kubenswrapper[4890]: I1125 15:40:54.745375 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ca9e616d-0ce2-4784-945c-9e99ba4f437b-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4fkzh\" (UID: \"ca9e616d-0ce2-4784-945c-9e99ba4f437b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4fkzh" Nov 25 15:40:54 crc kubenswrapper[4890]: I1125 15:40:54.847859 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ca9e616d-0ce2-4784-945c-9e99ba4f437b-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4fkzh\" (UID: \"ca9e616d-0ce2-4784-945c-9e99ba4f437b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4fkzh" Nov 25 15:40:54 crc kubenswrapper[4890]: I1125 15:40:54.847996 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/ca9e616d-0ce2-4784-945c-9e99ba4f437b-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4fkzh\" (UID: \"ca9e616d-0ce2-4784-945c-9e99ba4f437b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4fkzh" Nov 25 15:40:54 crc kubenswrapper[4890]: I1125 15:40:54.848033 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca9e616d-0ce2-4784-945c-9e99ba4f437b-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4fkzh\" (UID: \"ca9e616d-0ce2-4784-945c-9e99ba4f437b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4fkzh" Nov 25 15:40:54 crc kubenswrapper[4890]: I1125 15:40:54.848111 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ca9e616d-0ce2-4784-945c-9e99ba4f437b-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4fkzh\" (UID: \"ca9e616d-0ce2-4784-945c-9e99ba4f437b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4fkzh" Nov 25 15:40:54 crc kubenswrapper[4890]: I1125 15:40:54.848149 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpvqs\" (UniqueName: \"kubernetes.io/projected/ca9e616d-0ce2-4784-945c-9e99ba4f437b-kube-api-access-cpvqs\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4fkzh\" (UID: \"ca9e616d-0ce2-4784-945c-9e99ba4f437b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4fkzh" Nov 25 15:40:54 crc kubenswrapper[4890]: I1125 15:40:54.853939 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ca9e616d-0ce2-4784-945c-9e99ba4f437b-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4fkzh\" (UID: \"ca9e616d-0ce2-4784-945c-9e99ba4f437b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4fkzh" Nov 25 15:40:54 crc kubenswrapper[4890]: I1125 15:40:54.853957 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/ca9e616d-0ce2-4784-945c-9e99ba4f437b-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4fkzh\" (UID: \"ca9e616d-0ce2-4784-945c-9e99ba4f437b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4fkzh" Nov 25 15:40:54 crc kubenswrapper[4890]: I1125 15:40:54.854796 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca9e616d-0ce2-4784-945c-9e99ba4f437b-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4fkzh\" (UID: \"ca9e616d-0ce2-4784-945c-9e99ba4f437b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4fkzh" Nov 25 15:40:54 crc kubenswrapper[4890]: I1125 15:40:54.854897 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ca9e616d-0ce2-4784-945c-9e99ba4f437b-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4fkzh\" (UID: \"ca9e616d-0ce2-4784-945c-9e99ba4f437b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4fkzh" Nov 25 15:40:54 crc kubenswrapper[4890]: I1125 15:40:54.869139 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpvqs\" (UniqueName: \"kubernetes.io/projected/ca9e616d-0ce2-4784-945c-9e99ba4f437b-kube-api-access-cpvqs\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4fkzh\" (UID: \"ca9e616d-0ce2-4784-945c-9e99ba4f437b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4fkzh" Nov 25 15:40:55 crc kubenswrapper[4890]: I1125 15:40:55.005969 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4fkzh" Nov 25 15:40:55 crc kubenswrapper[4890]: I1125 15:40:55.410022 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4fkzh"] Nov 25 15:40:55 crc kubenswrapper[4890]: I1125 15:40:55.539228 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4fkzh" event={"ID":"ca9e616d-0ce2-4784-945c-9e99ba4f437b","Type":"ContainerStarted","Data":"5154655cd8559d5057573c28ce6b02e89c2a80f8580101a266396f165a9a83cf"} Nov 25 15:40:56 crc kubenswrapper[4890]: I1125 15:40:56.552307 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4fkzh" event={"ID":"ca9e616d-0ce2-4784-945c-9e99ba4f437b","Type":"ContainerStarted","Data":"9c0c6b10820010684ae74d46d8b4b424e45b34e3094ec3d1cafdd1634893ec12"} Nov 25 15:40:56 crc kubenswrapper[4890]: I1125 15:40:56.574609 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4fkzh" podStartSLOduration=2.140734458 podStartE2EDuration="2.57458566s" podCreationTimestamp="2025-11-25 15:40:54 +0000 UTC" firstStartedPulling="2025-11-25 15:40:55.412979721 +0000 UTC m=+2313.855442321" lastFinishedPulling="2025-11-25 15:40:55.846830913 +0000 UTC m=+2314.289293523" observedRunningTime="2025-11-25 15:40:56.571603368 +0000 UTC m=+2315.014065978" watchObservedRunningTime="2025-11-25 15:40:56.57458566 +0000 UTC m=+2315.017048270" Nov 25 15:40:59 crc kubenswrapper[4890]: I1125 15:40:59.007518 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-q8bnq"] Nov 25 15:40:59 crc kubenswrapper[4890]: I1125 15:40:59.010212 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q8bnq" Nov 25 15:40:59 crc kubenswrapper[4890]: I1125 15:40:59.025418 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q8bnq"] Nov 25 15:40:59 crc kubenswrapper[4890]: I1125 15:40:59.150478 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70a067f2-ef2e-4abc-ae6c-6c8d6eb63254-catalog-content\") pod \"redhat-operators-q8bnq\" (UID: \"70a067f2-ef2e-4abc-ae6c-6c8d6eb63254\") " pod="openshift-marketplace/redhat-operators-q8bnq" Nov 25 15:40:59 crc kubenswrapper[4890]: I1125 15:40:59.150624 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70a067f2-ef2e-4abc-ae6c-6c8d6eb63254-utilities\") pod \"redhat-operators-q8bnq\" (UID: \"70a067f2-ef2e-4abc-ae6c-6c8d6eb63254\") " pod="openshift-marketplace/redhat-operators-q8bnq" Nov 25 15:40:59 crc kubenswrapper[4890]: I1125 15:40:59.150651 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmzwm\" (UniqueName: \"kubernetes.io/projected/70a067f2-ef2e-4abc-ae6c-6c8d6eb63254-kube-api-access-fmzwm\") pod \"redhat-operators-q8bnq\" (UID: \"70a067f2-ef2e-4abc-ae6c-6c8d6eb63254\") " pod="openshift-marketplace/redhat-operators-q8bnq" Nov 25 15:40:59 crc kubenswrapper[4890]: I1125 15:40:59.252593 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70a067f2-ef2e-4abc-ae6c-6c8d6eb63254-catalog-content\") pod \"redhat-operators-q8bnq\" (UID: \"70a067f2-ef2e-4abc-ae6c-6c8d6eb63254\") " pod="openshift-marketplace/redhat-operators-q8bnq" Nov 25 15:40:59 crc kubenswrapper[4890]: I1125 15:40:59.252742 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70a067f2-ef2e-4abc-ae6c-6c8d6eb63254-utilities\") pod \"redhat-operators-q8bnq\" (UID: \"70a067f2-ef2e-4abc-ae6c-6c8d6eb63254\") " pod="openshift-marketplace/redhat-operators-q8bnq" Nov 25 15:40:59 crc kubenswrapper[4890]: I1125 15:40:59.252770 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmzwm\" (UniqueName: \"kubernetes.io/projected/70a067f2-ef2e-4abc-ae6c-6c8d6eb63254-kube-api-access-fmzwm\") pod \"redhat-operators-q8bnq\" (UID: \"70a067f2-ef2e-4abc-ae6c-6c8d6eb63254\") " pod="openshift-marketplace/redhat-operators-q8bnq" Nov 25 15:40:59 crc kubenswrapper[4890]: I1125 15:40:59.266401 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70a067f2-ef2e-4abc-ae6c-6c8d6eb63254-catalog-content\") pod \"redhat-operators-q8bnq\" (UID: \"70a067f2-ef2e-4abc-ae6c-6c8d6eb63254\") " pod="openshift-marketplace/redhat-operators-q8bnq" Nov 25 15:40:59 crc kubenswrapper[4890]: I1125 15:40:59.266615 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70a067f2-ef2e-4abc-ae6c-6c8d6eb63254-utilities\") pod \"redhat-operators-q8bnq\" (UID: \"70a067f2-ef2e-4abc-ae6c-6c8d6eb63254\") " pod="openshift-marketplace/redhat-operators-q8bnq" Nov 25 15:40:59 crc kubenswrapper[4890]: I1125 15:40:59.303657 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmzwm\" (UniqueName: \"kubernetes.io/projected/70a067f2-ef2e-4abc-ae6c-6c8d6eb63254-kube-api-access-fmzwm\") pod \"redhat-operators-q8bnq\" (UID: \"70a067f2-ef2e-4abc-ae6c-6c8d6eb63254\") " pod="openshift-marketplace/redhat-operators-q8bnq" Nov 25 15:40:59 crc kubenswrapper[4890]: I1125 15:40:59.333119 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q8bnq" Nov 25 15:40:59 crc kubenswrapper[4890]: I1125 15:40:59.871105 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q8bnq"] Nov 25 15:40:59 crc kubenswrapper[4890]: W1125 15:40:59.875603 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod70a067f2_ef2e_4abc_ae6c_6c8d6eb63254.slice/crio-72eafc40af2a2d88d6149d8e1dd80167e84141eedbbadf94870e04450b3e6ffa WatchSource:0}: Error finding container 72eafc40af2a2d88d6149d8e1dd80167e84141eedbbadf94870e04450b3e6ffa: Status 404 returned error can't find the container with id 72eafc40af2a2d88d6149d8e1dd80167e84141eedbbadf94870e04450b3e6ffa Nov 25 15:41:00 crc kubenswrapper[4890]: I1125 15:41:00.596944 4890 generic.go:334] "Generic (PLEG): container finished" podID="70a067f2-ef2e-4abc-ae6c-6c8d6eb63254" containerID="dd04894537926fa44ad23401ad4f0fef21b5f8d95287f1445d3f49eb09b9efbc" exitCode=0 Nov 25 15:41:00 crc kubenswrapper[4890]: I1125 15:41:00.597016 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q8bnq" event={"ID":"70a067f2-ef2e-4abc-ae6c-6c8d6eb63254","Type":"ContainerDied","Data":"dd04894537926fa44ad23401ad4f0fef21b5f8d95287f1445d3f49eb09b9efbc"} Nov 25 15:41:00 crc kubenswrapper[4890]: I1125 15:41:00.597353 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q8bnq" event={"ID":"70a067f2-ef2e-4abc-ae6c-6c8d6eb63254","Type":"ContainerStarted","Data":"72eafc40af2a2d88d6149d8e1dd80167e84141eedbbadf94870e04450b3e6ffa"} Nov 25 15:41:01 crc kubenswrapper[4890]: I1125 15:41:01.172858 4890 scope.go:117] "RemoveContainer" containerID="67ca0de5cd1ef0f73ff70ee2df49e62ca714111ad20a0de1fef88f6f60fdf262" Nov 25 15:41:01 crc kubenswrapper[4890]: E1125 15:41:01.173574 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:41:02 crc kubenswrapper[4890]: I1125 15:41:02.617994 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q8bnq" event={"ID":"70a067f2-ef2e-4abc-ae6c-6c8d6eb63254","Type":"ContainerStarted","Data":"12290ce58469d8c5921fe232d0b010238c73857ab11642116444cad874db42d6"} Nov 25 15:41:05 crc kubenswrapper[4890]: I1125 15:41:05.645879 4890 generic.go:334] "Generic (PLEG): container finished" podID="70a067f2-ef2e-4abc-ae6c-6c8d6eb63254" containerID="12290ce58469d8c5921fe232d0b010238c73857ab11642116444cad874db42d6" exitCode=0 Nov 25 15:41:05 crc kubenswrapper[4890]: I1125 15:41:05.646095 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q8bnq" event={"ID":"70a067f2-ef2e-4abc-ae6c-6c8d6eb63254","Type":"ContainerDied","Data":"12290ce58469d8c5921fe232d0b010238c73857ab11642116444cad874db42d6"} Nov 25 15:41:06 crc kubenswrapper[4890]: I1125 15:41:06.657026 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q8bnq" event={"ID":"70a067f2-ef2e-4abc-ae6c-6c8d6eb63254","Type":"ContainerStarted","Data":"55c4b87d3ed4834cbf8e8003d1c8da7fc834cb540e2dfd2b47f187a79c6109b2"} Nov 25 15:41:07 crc kubenswrapper[4890]: I1125 15:41:07.688648 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-q8bnq" podStartSLOduration=4.168923618 podStartE2EDuration="9.688631661s" podCreationTimestamp="2025-11-25 15:40:58 +0000 UTC" firstStartedPulling="2025-11-25 15:41:00.598577326 +0000 UTC m=+2319.041039936" lastFinishedPulling="2025-11-25 15:41:06.118285369 +0000 UTC m=+2324.560747979" observedRunningTime="2025-11-25 15:41:07.686024368 +0000 UTC m=+2326.128486978" watchObservedRunningTime="2025-11-25 15:41:07.688631661 +0000 UTC m=+2326.131094271" Nov 25 15:41:09 crc kubenswrapper[4890]: I1125 15:41:09.335363 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-q8bnq" Nov 25 15:41:09 crc kubenswrapper[4890]: I1125 15:41:09.335813 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-q8bnq" Nov 25 15:41:10 crc kubenswrapper[4890]: I1125 15:41:10.382033 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-q8bnq" podUID="70a067f2-ef2e-4abc-ae6c-6c8d6eb63254" containerName="registry-server" probeResult="failure" output=< Nov 25 15:41:10 crc kubenswrapper[4890]: timeout: failed to connect service ":50051" within 1s Nov 25 15:41:10 crc kubenswrapper[4890]: > Nov 25 15:41:15 crc kubenswrapper[4890]: I1125 15:41:15.173171 4890 scope.go:117] "RemoveContainer" containerID="67ca0de5cd1ef0f73ff70ee2df49e62ca714111ad20a0de1fef88f6f60fdf262" Nov 25 15:41:15 crc kubenswrapper[4890]: E1125 15:41:15.173726 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:41:19 crc kubenswrapper[4890]: I1125 15:41:19.387367 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-q8bnq" Nov 25 15:41:19 crc kubenswrapper[4890]: I1125 15:41:19.442298 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-q8bnq" Nov 25 15:41:19 crc kubenswrapper[4890]: I1125 15:41:19.633773 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q8bnq"] Nov 25 15:41:20 crc kubenswrapper[4890]: I1125 15:41:20.795025 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-q8bnq" podUID="70a067f2-ef2e-4abc-ae6c-6c8d6eb63254" containerName="registry-server" containerID="cri-o://55c4b87d3ed4834cbf8e8003d1c8da7fc834cb540e2dfd2b47f187a79c6109b2" gracePeriod=2 Nov 25 15:41:21 crc kubenswrapper[4890]: I1125 15:41:21.280703 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q8bnq" Nov 25 15:41:21 crc kubenswrapper[4890]: I1125 15:41:21.368877 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fmzwm\" (UniqueName: \"kubernetes.io/projected/70a067f2-ef2e-4abc-ae6c-6c8d6eb63254-kube-api-access-fmzwm\") pod \"70a067f2-ef2e-4abc-ae6c-6c8d6eb63254\" (UID: \"70a067f2-ef2e-4abc-ae6c-6c8d6eb63254\") " Nov 25 15:41:21 crc kubenswrapper[4890]: I1125 15:41:21.368955 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70a067f2-ef2e-4abc-ae6c-6c8d6eb63254-utilities\") pod \"70a067f2-ef2e-4abc-ae6c-6c8d6eb63254\" (UID: \"70a067f2-ef2e-4abc-ae6c-6c8d6eb63254\") " Nov 25 15:41:21 crc kubenswrapper[4890]: I1125 15:41:21.369100 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70a067f2-ef2e-4abc-ae6c-6c8d6eb63254-catalog-content\") pod \"70a067f2-ef2e-4abc-ae6c-6c8d6eb63254\" (UID: \"70a067f2-ef2e-4abc-ae6c-6c8d6eb63254\") " Nov 25 15:41:21 crc kubenswrapper[4890]: I1125 15:41:21.370065 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70a067f2-ef2e-4abc-ae6c-6c8d6eb63254-utilities" (OuterVolumeSpecName: "utilities") pod "70a067f2-ef2e-4abc-ae6c-6c8d6eb63254" (UID: "70a067f2-ef2e-4abc-ae6c-6c8d6eb63254"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:41:21 crc kubenswrapper[4890]: I1125 15:41:21.376749 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70a067f2-ef2e-4abc-ae6c-6c8d6eb63254-kube-api-access-fmzwm" (OuterVolumeSpecName: "kube-api-access-fmzwm") pod "70a067f2-ef2e-4abc-ae6c-6c8d6eb63254" (UID: "70a067f2-ef2e-4abc-ae6c-6c8d6eb63254"). InnerVolumeSpecName "kube-api-access-fmzwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:41:21 crc kubenswrapper[4890]: I1125 15:41:21.464327 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70a067f2-ef2e-4abc-ae6c-6c8d6eb63254-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "70a067f2-ef2e-4abc-ae6c-6c8d6eb63254" (UID: "70a067f2-ef2e-4abc-ae6c-6c8d6eb63254"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:41:21 crc kubenswrapper[4890]: I1125 15:41:21.473764 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fmzwm\" (UniqueName: \"kubernetes.io/projected/70a067f2-ef2e-4abc-ae6c-6c8d6eb63254-kube-api-access-fmzwm\") on node \"crc\" DevicePath \"\"" Nov 25 15:41:21 crc kubenswrapper[4890]: I1125 15:41:21.473815 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70a067f2-ef2e-4abc-ae6c-6c8d6eb63254-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 15:41:21 crc kubenswrapper[4890]: I1125 15:41:21.473828 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70a067f2-ef2e-4abc-ae6c-6c8d6eb63254-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 15:41:21 crc kubenswrapper[4890]: I1125 15:41:21.805709 4890 generic.go:334] "Generic (PLEG): container finished" podID="70a067f2-ef2e-4abc-ae6c-6c8d6eb63254" containerID="55c4b87d3ed4834cbf8e8003d1c8da7fc834cb540e2dfd2b47f187a79c6109b2" exitCode=0 Nov 25 15:41:21 crc kubenswrapper[4890]: I1125 15:41:21.805790 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q8bnq" event={"ID":"70a067f2-ef2e-4abc-ae6c-6c8d6eb63254","Type":"ContainerDied","Data":"55c4b87d3ed4834cbf8e8003d1c8da7fc834cb540e2dfd2b47f187a79c6109b2"} Nov 25 15:41:21 crc kubenswrapper[4890]: I1125 15:41:21.805796 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q8bnq" Nov 25 15:41:21 crc kubenswrapper[4890]: I1125 15:41:21.805872 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q8bnq" event={"ID":"70a067f2-ef2e-4abc-ae6c-6c8d6eb63254","Type":"ContainerDied","Data":"72eafc40af2a2d88d6149d8e1dd80167e84141eedbbadf94870e04450b3e6ffa"} Nov 25 15:41:21 crc kubenswrapper[4890]: I1125 15:41:21.805904 4890 scope.go:117] "RemoveContainer" containerID="55c4b87d3ed4834cbf8e8003d1c8da7fc834cb540e2dfd2b47f187a79c6109b2" Nov 25 15:41:21 crc kubenswrapper[4890]: I1125 15:41:21.845500 4890 scope.go:117] "RemoveContainer" containerID="12290ce58469d8c5921fe232d0b010238c73857ab11642116444cad874db42d6" Nov 25 15:41:21 crc kubenswrapper[4890]: I1125 15:41:21.851202 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q8bnq"] Nov 25 15:41:21 crc kubenswrapper[4890]: I1125 15:41:21.865196 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-q8bnq"] Nov 25 15:41:21 crc kubenswrapper[4890]: I1125 15:41:21.879371 4890 scope.go:117] "RemoveContainer" containerID="dd04894537926fa44ad23401ad4f0fef21b5f8d95287f1445d3f49eb09b9efbc" Nov 25 15:41:21 crc kubenswrapper[4890]: I1125 15:41:21.923663 4890 scope.go:117] "RemoveContainer" containerID="55c4b87d3ed4834cbf8e8003d1c8da7fc834cb540e2dfd2b47f187a79c6109b2" Nov 25 15:41:21 crc kubenswrapper[4890]: E1125 15:41:21.924290 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55c4b87d3ed4834cbf8e8003d1c8da7fc834cb540e2dfd2b47f187a79c6109b2\": container with ID starting with 55c4b87d3ed4834cbf8e8003d1c8da7fc834cb540e2dfd2b47f187a79c6109b2 not found: ID does not exist" containerID="55c4b87d3ed4834cbf8e8003d1c8da7fc834cb540e2dfd2b47f187a79c6109b2" Nov 25 15:41:21 crc kubenswrapper[4890]: I1125 15:41:21.924347 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55c4b87d3ed4834cbf8e8003d1c8da7fc834cb540e2dfd2b47f187a79c6109b2"} err="failed to get container status \"55c4b87d3ed4834cbf8e8003d1c8da7fc834cb540e2dfd2b47f187a79c6109b2\": rpc error: code = NotFound desc = could not find container \"55c4b87d3ed4834cbf8e8003d1c8da7fc834cb540e2dfd2b47f187a79c6109b2\": container with ID starting with 55c4b87d3ed4834cbf8e8003d1c8da7fc834cb540e2dfd2b47f187a79c6109b2 not found: ID does not exist" Nov 25 15:41:21 crc kubenswrapper[4890]: I1125 15:41:21.924387 4890 scope.go:117] "RemoveContainer" containerID="12290ce58469d8c5921fe232d0b010238c73857ab11642116444cad874db42d6" Nov 25 15:41:21 crc kubenswrapper[4890]: E1125 15:41:21.925302 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12290ce58469d8c5921fe232d0b010238c73857ab11642116444cad874db42d6\": container with ID starting with 12290ce58469d8c5921fe232d0b010238c73857ab11642116444cad874db42d6 not found: ID does not exist" containerID="12290ce58469d8c5921fe232d0b010238c73857ab11642116444cad874db42d6" Nov 25 15:41:21 crc kubenswrapper[4890]: I1125 15:41:21.925342 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12290ce58469d8c5921fe232d0b010238c73857ab11642116444cad874db42d6"} err="failed to get container status \"12290ce58469d8c5921fe232d0b010238c73857ab11642116444cad874db42d6\": rpc error: code = NotFound desc = could not find container \"12290ce58469d8c5921fe232d0b010238c73857ab11642116444cad874db42d6\": container with ID starting with 12290ce58469d8c5921fe232d0b010238c73857ab11642116444cad874db42d6 not found: ID does not exist" Nov 25 15:41:21 crc kubenswrapper[4890]: I1125 15:41:21.925363 4890 scope.go:117] "RemoveContainer" containerID="dd04894537926fa44ad23401ad4f0fef21b5f8d95287f1445d3f49eb09b9efbc" Nov 25 15:41:21 crc kubenswrapper[4890]: E1125 15:41:21.925873 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd04894537926fa44ad23401ad4f0fef21b5f8d95287f1445d3f49eb09b9efbc\": container with ID starting with dd04894537926fa44ad23401ad4f0fef21b5f8d95287f1445d3f49eb09b9efbc not found: ID does not exist" containerID="dd04894537926fa44ad23401ad4f0fef21b5f8d95287f1445d3f49eb09b9efbc" Nov 25 15:41:21 crc kubenswrapper[4890]: I1125 15:41:21.925900 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd04894537926fa44ad23401ad4f0fef21b5f8d95287f1445d3f49eb09b9efbc"} err="failed to get container status \"dd04894537926fa44ad23401ad4f0fef21b5f8d95287f1445d3f49eb09b9efbc\": rpc error: code = NotFound desc = could not find container \"dd04894537926fa44ad23401ad4f0fef21b5f8d95287f1445d3f49eb09b9efbc\": container with ID starting with dd04894537926fa44ad23401ad4f0fef21b5f8d95287f1445d3f49eb09b9efbc not found: ID does not exist" Nov 25 15:41:22 crc kubenswrapper[4890]: I1125 15:41:22.188694 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70a067f2-ef2e-4abc-ae6c-6c8d6eb63254" path="/var/lib/kubelet/pods/70a067f2-ef2e-4abc-ae6c-6c8d6eb63254/volumes" Nov 25 15:41:27 crc kubenswrapper[4890]: I1125 15:41:27.173071 4890 scope.go:117] "RemoveContainer" containerID="67ca0de5cd1ef0f73ff70ee2df49e62ca714111ad20a0de1fef88f6f60fdf262" Nov 25 15:41:27 crc kubenswrapper[4890]: E1125 15:41:27.174098 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:41:38 crc kubenswrapper[4890]: I1125 15:41:38.174520 4890 scope.go:117] "RemoveContainer" containerID="67ca0de5cd1ef0f73ff70ee2df49e62ca714111ad20a0de1fef88f6f60fdf262" Nov 25 15:41:38 crc kubenswrapper[4890]: E1125 15:41:38.175910 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:41:53 crc kubenswrapper[4890]: I1125 15:41:53.174541 4890 scope.go:117] "RemoveContainer" containerID="67ca0de5cd1ef0f73ff70ee2df49e62ca714111ad20a0de1fef88f6f60fdf262" Nov 25 15:41:53 crc kubenswrapper[4890]: E1125 15:41:53.175836 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:42:04 crc kubenswrapper[4890]: I1125 15:42:04.173100 4890 scope.go:117] "RemoveContainer" containerID="67ca0de5cd1ef0f73ff70ee2df49e62ca714111ad20a0de1fef88f6f60fdf262" Nov 25 15:42:04 crc kubenswrapper[4890]: E1125 15:42:04.174143 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:42:19 crc kubenswrapper[4890]: I1125 15:42:19.173458 4890 scope.go:117] "RemoveContainer" containerID="67ca0de5cd1ef0f73ff70ee2df49e62ca714111ad20a0de1fef88f6f60fdf262" Nov 25 15:42:19 crc kubenswrapper[4890]: E1125 15:42:19.174639 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:42:32 crc kubenswrapper[4890]: I1125 15:42:32.181197 4890 scope.go:117] "RemoveContainer" containerID="67ca0de5cd1ef0f73ff70ee2df49e62ca714111ad20a0de1fef88f6f60fdf262" Nov 25 15:42:32 crc kubenswrapper[4890]: E1125 15:42:32.182044 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:42:45 crc kubenswrapper[4890]: I1125 15:42:45.173963 4890 scope.go:117] "RemoveContainer" containerID="67ca0de5cd1ef0f73ff70ee2df49e62ca714111ad20a0de1fef88f6f60fdf262" Nov 25 15:42:45 crc kubenswrapper[4890]: E1125 15:42:45.174833 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:42:59 crc kubenswrapper[4890]: I1125 15:42:59.172617 4890 scope.go:117] "RemoveContainer" containerID="67ca0de5cd1ef0f73ff70ee2df49e62ca714111ad20a0de1fef88f6f60fdf262" Nov 25 15:42:59 crc kubenswrapper[4890]: E1125 15:42:59.173575 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:43:10 crc kubenswrapper[4890]: I1125 15:43:10.173929 4890 scope.go:117] "RemoveContainer" containerID="67ca0de5cd1ef0f73ff70ee2df49e62ca714111ad20a0de1fef88f6f60fdf262" Nov 25 15:43:10 crc kubenswrapper[4890]: E1125 15:43:10.174918 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:43:24 crc kubenswrapper[4890]: I1125 15:43:24.173380 4890 scope.go:117] "RemoveContainer" containerID="67ca0de5cd1ef0f73ff70ee2df49e62ca714111ad20a0de1fef88f6f60fdf262" Nov 25 15:43:24 crc kubenswrapper[4890]: E1125 15:43:24.174587 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:43:36 crc kubenswrapper[4890]: I1125 15:43:36.174210 4890 scope.go:117] "RemoveContainer" containerID="67ca0de5cd1ef0f73ff70ee2df49e62ca714111ad20a0de1fef88f6f60fdf262" Nov 25 15:43:36 crc kubenswrapper[4890]: E1125 15:43:36.175403 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:43:50 crc kubenswrapper[4890]: I1125 15:43:50.173056 4890 scope.go:117] "RemoveContainer" containerID="67ca0de5cd1ef0f73ff70ee2df49e62ca714111ad20a0de1fef88f6f60fdf262" Nov 25 15:43:50 crc kubenswrapper[4890]: E1125 15:43:50.174002 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:44:01 crc kubenswrapper[4890]: I1125 15:44:01.173858 4890 scope.go:117] "RemoveContainer" containerID="67ca0de5cd1ef0f73ff70ee2df49e62ca714111ad20a0de1fef88f6f60fdf262" Nov 25 15:44:01 crc kubenswrapper[4890]: E1125 15:44:01.174755 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:44:15 crc kubenswrapper[4890]: I1125 15:44:15.173576 4890 scope.go:117] "RemoveContainer" containerID="67ca0de5cd1ef0f73ff70ee2df49e62ca714111ad20a0de1fef88f6f60fdf262" Nov 25 15:44:15 crc kubenswrapper[4890]: E1125 15:44:15.174509 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:44:30 crc kubenswrapper[4890]: I1125 15:44:30.173425 4890 scope.go:117] "RemoveContainer" containerID="67ca0de5cd1ef0f73ff70ee2df49e62ca714111ad20a0de1fef88f6f60fdf262" Nov 25 15:44:30 crc kubenswrapper[4890]: I1125 15:44:30.688732 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" event={"ID":"4e4f849d-f239-4727-a73e-18327856929a","Type":"ContainerStarted","Data":"3add3ce883ec44a5429bf3e4879e91159a451c26e6f8d63ebfd5527dc360e6b8"} Nov 25 15:44:52 crc kubenswrapper[4890]: I1125 15:44:52.914789 4890 generic.go:334] "Generic (PLEG): container finished" podID="ca9e616d-0ce2-4784-945c-9e99ba4f437b" containerID="9c0c6b10820010684ae74d46d8b4b424e45b34e3094ec3d1cafdd1634893ec12" exitCode=0 Nov 25 15:44:52 crc kubenswrapper[4890]: I1125 15:44:52.914884 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4fkzh" event={"ID":"ca9e616d-0ce2-4784-945c-9e99ba4f437b","Type":"ContainerDied","Data":"9c0c6b10820010684ae74d46d8b4b424e45b34e3094ec3d1cafdd1634893ec12"} Nov 25 15:44:54 crc kubenswrapper[4890]: I1125 15:44:54.370285 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4fkzh" Nov 25 15:44:54 crc kubenswrapper[4890]: I1125 15:44:54.544306 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cpvqs\" (UniqueName: \"kubernetes.io/projected/ca9e616d-0ce2-4784-945c-9e99ba4f437b-kube-api-access-cpvqs\") pod \"ca9e616d-0ce2-4784-945c-9e99ba4f437b\" (UID: \"ca9e616d-0ce2-4784-945c-9e99ba4f437b\") " Nov 25 15:44:54 crc kubenswrapper[4890]: I1125 15:44:54.544722 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ca9e616d-0ce2-4784-945c-9e99ba4f437b-ssh-key\") pod \"ca9e616d-0ce2-4784-945c-9e99ba4f437b\" (UID: \"ca9e616d-0ce2-4784-945c-9e99ba4f437b\") " Nov 25 15:44:54 crc kubenswrapper[4890]: I1125 15:44:54.544879 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/ca9e616d-0ce2-4784-945c-9e99ba4f437b-libvirt-secret-0\") pod \"ca9e616d-0ce2-4784-945c-9e99ba4f437b\" (UID: \"ca9e616d-0ce2-4784-945c-9e99ba4f437b\") " Nov 25 15:44:54 crc kubenswrapper[4890]: I1125 15:44:54.545001 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ca9e616d-0ce2-4784-945c-9e99ba4f437b-inventory\") pod \"ca9e616d-0ce2-4784-945c-9e99ba4f437b\" (UID: \"ca9e616d-0ce2-4784-945c-9e99ba4f437b\") " Nov 25 15:44:54 crc kubenswrapper[4890]: I1125 15:44:54.545764 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca9e616d-0ce2-4784-945c-9e99ba4f437b-libvirt-combined-ca-bundle\") pod \"ca9e616d-0ce2-4784-945c-9e99ba4f437b\" (UID: \"ca9e616d-0ce2-4784-945c-9e99ba4f437b\") " Nov 25 15:44:54 crc kubenswrapper[4890]: I1125 15:44:54.553359 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca9e616d-0ce2-4784-945c-9e99ba4f437b-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "ca9e616d-0ce2-4784-945c-9e99ba4f437b" (UID: "ca9e616d-0ce2-4784-945c-9e99ba4f437b"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:44:54 crc kubenswrapper[4890]: I1125 15:44:54.553406 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca9e616d-0ce2-4784-945c-9e99ba4f437b-kube-api-access-cpvqs" (OuterVolumeSpecName: "kube-api-access-cpvqs") pod "ca9e616d-0ce2-4784-945c-9e99ba4f437b" (UID: "ca9e616d-0ce2-4784-945c-9e99ba4f437b"). InnerVolumeSpecName "kube-api-access-cpvqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:44:54 crc kubenswrapper[4890]: I1125 15:44:54.578540 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca9e616d-0ce2-4784-945c-9e99ba4f437b-inventory" (OuterVolumeSpecName: "inventory") pod "ca9e616d-0ce2-4784-945c-9e99ba4f437b" (UID: "ca9e616d-0ce2-4784-945c-9e99ba4f437b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:44:54 crc kubenswrapper[4890]: I1125 15:44:54.583443 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca9e616d-0ce2-4784-945c-9e99ba4f437b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ca9e616d-0ce2-4784-945c-9e99ba4f437b" (UID: "ca9e616d-0ce2-4784-945c-9e99ba4f437b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:44:54 crc kubenswrapper[4890]: I1125 15:44:54.587555 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca9e616d-0ce2-4784-945c-9e99ba4f437b-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "ca9e616d-0ce2-4784-945c-9e99ba4f437b" (UID: "ca9e616d-0ce2-4784-945c-9e99ba4f437b"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:44:54 crc kubenswrapper[4890]: I1125 15:44:54.649322 4890 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ca9e616d-0ce2-4784-945c-9e99ba4f437b-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 15:44:54 crc kubenswrapper[4890]: I1125 15:44:54.649366 4890 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca9e616d-0ce2-4784-945c-9e99ba4f437b-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:44:54 crc kubenswrapper[4890]: I1125 15:44:54.649380 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cpvqs\" (UniqueName: \"kubernetes.io/projected/ca9e616d-0ce2-4784-945c-9e99ba4f437b-kube-api-access-cpvqs\") on node \"crc\" DevicePath \"\"" Nov 25 15:44:54 crc kubenswrapper[4890]: I1125 15:44:54.649394 4890 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ca9e616d-0ce2-4784-945c-9e99ba4f437b-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 15:44:54 crc kubenswrapper[4890]: I1125 15:44:54.649403 4890 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/ca9e616d-0ce2-4784-945c-9e99ba4f437b-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Nov 25 15:44:54 crc kubenswrapper[4890]: I1125 15:44:54.935154 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4fkzh" event={"ID":"ca9e616d-0ce2-4784-945c-9e99ba4f437b","Type":"ContainerDied","Data":"5154655cd8559d5057573c28ce6b02e89c2a80f8580101a266396f165a9a83cf"} Nov 25 15:44:54 crc kubenswrapper[4890]: I1125 15:44:54.935214 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4fkzh" Nov 25 15:44:54 crc kubenswrapper[4890]: I1125 15:44:54.935218 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5154655cd8559d5057573c28ce6b02e89c2a80f8580101a266396f165a9a83cf" Nov 25 15:44:55 crc kubenswrapper[4890]: I1125 15:44:55.028561 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-49nlj"] Nov 25 15:44:55 crc kubenswrapper[4890]: E1125 15:44:55.031229 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70a067f2-ef2e-4abc-ae6c-6c8d6eb63254" containerName="extract-utilities" Nov 25 15:44:55 crc kubenswrapper[4890]: I1125 15:44:55.031262 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="70a067f2-ef2e-4abc-ae6c-6c8d6eb63254" containerName="extract-utilities" Nov 25 15:44:55 crc kubenswrapper[4890]: E1125 15:44:55.031346 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70a067f2-ef2e-4abc-ae6c-6c8d6eb63254" containerName="registry-server" Nov 25 15:44:55 crc kubenswrapper[4890]: I1125 15:44:55.031362 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="70a067f2-ef2e-4abc-ae6c-6c8d6eb63254" containerName="registry-server" Nov 25 15:44:55 crc kubenswrapper[4890]: E1125 15:44:55.031710 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca9e616d-0ce2-4784-945c-9e99ba4f437b" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 25 15:44:55 crc kubenswrapper[4890]: I1125 15:44:55.031733 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca9e616d-0ce2-4784-945c-9e99ba4f437b" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 25 15:44:55 crc kubenswrapper[4890]: E1125 15:44:55.031767 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70a067f2-ef2e-4abc-ae6c-6c8d6eb63254" containerName="extract-content" Nov 25 15:44:55 crc kubenswrapper[4890]: I1125 15:44:55.031774 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="70a067f2-ef2e-4abc-ae6c-6c8d6eb63254" containerName="extract-content" Nov 25 15:44:55 crc kubenswrapper[4890]: I1125 15:44:55.031986 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="70a067f2-ef2e-4abc-ae6c-6c8d6eb63254" containerName="registry-server" Nov 25 15:44:55 crc kubenswrapper[4890]: I1125 15:44:55.032007 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca9e616d-0ce2-4784-945c-9e99ba4f437b" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 25 15:44:55 crc kubenswrapper[4890]: I1125 15:44:55.032831 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-49nlj" Nov 25 15:44:55 crc kubenswrapper[4890]: I1125 15:44:55.035553 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Nov 25 15:44:55 crc kubenswrapper[4890]: I1125 15:44:55.035656 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 15:44:55 crc kubenswrapper[4890]: I1125 15:44:55.035688 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Nov 25 15:44:55 crc kubenswrapper[4890]: I1125 15:44:55.035770 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 15:44:55 crc kubenswrapper[4890]: I1125 15:44:55.035789 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-22zzp" Nov 25 15:44:55 crc kubenswrapper[4890]: I1125 15:44:55.035977 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Nov 25 15:44:55 crc kubenswrapper[4890]: I1125 15:44:55.037374 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 15:44:55 crc kubenswrapper[4890]: I1125 15:44:55.044003 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-49nlj"] Nov 25 15:44:55 crc kubenswrapper[4890]: I1125 15:44:55.162787 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-49nlj\" (UID: \"e890bd11-32e9-4d73-9dd0-ae5d69a1fe58\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-49nlj" Nov 25 15:44:55 crc kubenswrapper[4890]: I1125 15:44:55.162853 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-49nlj\" (UID: \"e890bd11-32e9-4d73-9dd0-ae5d69a1fe58\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-49nlj" Nov 25 15:44:55 crc kubenswrapper[4890]: I1125 15:44:55.163099 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-49nlj\" (UID: \"e890bd11-32e9-4d73-9dd0-ae5d69a1fe58\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-49nlj" Nov 25 15:44:55 crc kubenswrapper[4890]: I1125 15:44:55.163228 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-49nlj\" (UID: \"e890bd11-32e9-4d73-9dd0-ae5d69a1fe58\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-49nlj" Nov 25 15:44:55 crc kubenswrapper[4890]: I1125 15:44:55.163345 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpnl2\" (UniqueName: \"kubernetes.io/projected/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-kube-api-access-qpnl2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-49nlj\" (UID: \"e890bd11-32e9-4d73-9dd0-ae5d69a1fe58\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-49nlj" Nov 25 15:44:55 crc kubenswrapper[4890]: I1125 15:44:55.163468 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-49nlj\" (UID: \"e890bd11-32e9-4d73-9dd0-ae5d69a1fe58\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-49nlj" Nov 25 15:44:55 crc kubenswrapper[4890]: I1125 15:44:55.163627 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-49nlj\" (UID: \"e890bd11-32e9-4d73-9dd0-ae5d69a1fe58\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-49nlj" Nov 25 15:44:55 crc kubenswrapper[4890]: I1125 15:44:55.163672 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-49nlj\" (UID: \"e890bd11-32e9-4d73-9dd0-ae5d69a1fe58\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-49nlj" Nov 25 15:44:55 crc kubenswrapper[4890]: I1125 15:44:55.163744 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-49nlj\" (UID: \"e890bd11-32e9-4d73-9dd0-ae5d69a1fe58\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-49nlj" Nov 25 15:44:55 crc kubenswrapper[4890]: I1125 15:44:55.266060 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-49nlj\" (UID: \"e890bd11-32e9-4d73-9dd0-ae5d69a1fe58\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-49nlj" Nov 25 15:44:55 crc kubenswrapper[4890]: I1125 15:44:55.266637 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-49nlj\" (UID: \"e890bd11-32e9-4d73-9dd0-ae5d69a1fe58\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-49nlj" Nov 25 15:44:55 crc kubenswrapper[4890]: I1125 15:44:55.266703 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-49nlj\" (UID: \"e890bd11-32e9-4d73-9dd0-ae5d69a1fe58\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-49nlj" Nov 25 15:44:55 crc kubenswrapper[4890]: I1125 15:44:55.266753 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-49nlj\" (UID: \"e890bd11-32e9-4d73-9dd0-ae5d69a1fe58\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-49nlj" Nov 25 15:44:55 crc kubenswrapper[4890]: I1125 15:44:55.266799 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpnl2\" (UniqueName: \"kubernetes.io/projected/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-kube-api-access-qpnl2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-49nlj\" (UID: \"e890bd11-32e9-4d73-9dd0-ae5d69a1fe58\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-49nlj" Nov 25 15:44:55 crc kubenswrapper[4890]: I1125 15:44:55.266853 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-49nlj\" (UID: \"e890bd11-32e9-4d73-9dd0-ae5d69a1fe58\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-49nlj" Nov 25 15:44:55 crc kubenswrapper[4890]: I1125 15:44:55.266919 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-49nlj\" (UID: \"e890bd11-32e9-4d73-9dd0-ae5d69a1fe58\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-49nlj" Nov 25 15:44:55 crc kubenswrapper[4890]: I1125 15:44:55.266949 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-49nlj\" (UID: \"e890bd11-32e9-4d73-9dd0-ae5d69a1fe58\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-49nlj" Nov 25 15:44:55 crc kubenswrapper[4890]: I1125 15:44:55.266978 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-49nlj\" (UID: \"e890bd11-32e9-4d73-9dd0-ae5d69a1fe58\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-49nlj" Nov 25 15:44:55 crc kubenswrapper[4890]: I1125 15:44:55.267639 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-49nlj\" (UID: \"e890bd11-32e9-4d73-9dd0-ae5d69a1fe58\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-49nlj" Nov 25 15:44:55 crc kubenswrapper[4890]: I1125 15:44:55.273019 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-49nlj\" (UID: \"e890bd11-32e9-4d73-9dd0-ae5d69a1fe58\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-49nlj" Nov 25 15:44:55 crc kubenswrapper[4890]: I1125 15:44:55.273510 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-49nlj\" (UID: \"e890bd11-32e9-4d73-9dd0-ae5d69a1fe58\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-49nlj" Nov 25 15:44:55 crc kubenswrapper[4890]: I1125 15:44:55.273765 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-49nlj\" (UID: \"e890bd11-32e9-4d73-9dd0-ae5d69a1fe58\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-49nlj" Nov 25 15:44:55 crc kubenswrapper[4890]: I1125 15:44:55.273887 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-49nlj\" (UID: \"e890bd11-32e9-4d73-9dd0-ae5d69a1fe58\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-49nlj" Nov 25 15:44:55 crc kubenswrapper[4890]: I1125 15:44:55.274236 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-49nlj\" (UID: \"e890bd11-32e9-4d73-9dd0-ae5d69a1fe58\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-49nlj" Nov 25 15:44:55 crc kubenswrapper[4890]: I1125 15:44:55.275002 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-49nlj\" (UID: \"e890bd11-32e9-4d73-9dd0-ae5d69a1fe58\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-49nlj" Nov 25 15:44:55 crc kubenswrapper[4890]: I1125 15:44:55.275474 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-49nlj\" (UID: \"e890bd11-32e9-4d73-9dd0-ae5d69a1fe58\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-49nlj" Nov 25 15:44:55 crc kubenswrapper[4890]: I1125 15:44:55.289136 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpnl2\" (UniqueName: \"kubernetes.io/projected/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-kube-api-access-qpnl2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-49nlj\" (UID: \"e890bd11-32e9-4d73-9dd0-ae5d69a1fe58\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-49nlj" Nov 25 15:44:55 crc kubenswrapper[4890]: I1125 15:44:55.359113 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-49nlj" Nov 25 15:44:55 crc kubenswrapper[4890]: I1125 15:44:55.968348 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-49nlj"] Nov 25 15:44:55 crc kubenswrapper[4890]: I1125 15:44:55.975090 4890 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 15:44:56 crc kubenswrapper[4890]: I1125 15:44:56.951997 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-49nlj" event={"ID":"e890bd11-32e9-4d73-9dd0-ae5d69a1fe58","Type":"ContainerStarted","Data":"8e79fadd953a14eed3e5a375d82ee1c10b4dd4da678f950428948e536b223ba0"} Nov 25 15:44:56 crc kubenswrapper[4890]: I1125 15:44:56.952625 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-49nlj" event={"ID":"e890bd11-32e9-4d73-9dd0-ae5d69a1fe58","Type":"ContainerStarted","Data":"2cccef917722046c030b3a1affff4121129ec2a46ed5427e2a2f7fec578c60ce"} Nov 25 15:44:56 crc kubenswrapper[4890]: I1125 15:44:56.978980 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-49nlj" podStartSLOduration=1.547131491 podStartE2EDuration="1.978957712s" podCreationTimestamp="2025-11-25 15:44:55 +0000 UTC" firstStartedPulling="2025-11-25 15:44:55.974843806 +0000 UTC m=+2554.417306416" lastFinishedPulling="2025-11-25 15:44:56.406670027 +0000 UTC m=+2554.849132637" observedRunningTime="2025-11-25 15:44:56.97246705 +0000 UTC m=+2555.414929670" watchObservedRunningTime="2025-11-25 15:44:56.978957712 +0000 UTC m=+2555.421420322" Nov 25 15:45:00 crc kubenswrapper[4890]: I1125 15:45:00.144484 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401425-zxwfh"] Nov 25 15:45:00 crc kubenswrapper[4890]: I1125 15:45:00.147085 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401425-zxwfh" Nov 25 15:45:00 crc kubenswrapper[4890]: I1125 15:45:00.155003 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 15:45:00 crc kubenswrapper[4890]: I1125 15:45:00.156147 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 15:45:00 crc kubenswrapper[4890]: I1125 15:45:00.164969 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401425-zxwfh"] Nov 25 15:45:00 crc kubenswrapper[4890]: I1125 15:45:00.286268 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/59d4c133-2584-4c4e-915d-e55e0b997107-secret-volume\") pod \"collect-profiles-29401425-zxwfh\" (UID: \"59d4c133-2584-4c4e-915d-e55e0b997107\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401425-zxwfh" Nov 25 15:45:00 crc kubenswrapper[4890]: I1125 15:45:00.286423 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zwps\" (UniqueName: \"kubernetes.io/projected/59d4c133-2584-4c4e-915d-e55e0b997107-kube-api-access-5zwps\") pod \"collect-profiles-29401425-zxwfh\" (UID: \"59d4c133-2584-4c4e-915d-e55e0b997107\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401425-zxwfh" Nov 25 15:45:00 crc kubenswrapper[4890]: I1125 15:45:00.286519 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/59d4c133-2584-4c4e-915d-e55e0b997107-config-volume\") pod \"collect-profiles-29401425-zxwfh\" (UID: \"59d4c133-2584-4c4e-915d-e55e0b997107\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401425-zxwfh" Nov 25 15:45:00 crc kubenswrapper[4890]: I1125 15:45:00.388420 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/59d4c133-2584-4c4e-915d-e55e0b997107-secret-volume\") pod \"collect-profiles-29401425-zxwfh\" (UID: \"59d4c133-2584-4c4e-915d-e55e0b997107\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401425-zxwfh" Nov 25 15:45:00 crc kubenswrapper[4890]: I1125 15:45:00.388517 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zwps\" (UniqueName: \"kubernetes.io/projected/59d4c133-2584-4c4e-915d-e55e0b997107-kube-api-access-5zwps\") pod \"collect-profiles-29401425-zxwfh\" (UID: \"59d4c133-2584-4c4e-915d-e55e0b997107\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401425-zxwfh" Nov 25 15:45:00 crc kubenswrapper[4890]: I1125 15:45:00.388608 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/59d4c133-2584-4c4e-915d-e55e0b997107-config-volume\") pod \"collect-profiles-29401425-zxwfh\" (UID: \"59d4c133-2584-4c4e-915d-e55e0b997107\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401425-zxwfh" Nov 25 15:45:00 crc kubenswrapper[4890]: I1125 15:45:00.389705 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/59d4c133-2584-4c4e-915d-e55e0b997107-config-volume\") pod \"collect-profiles-29401425-zxwfh\" (UID: \"59d4c133-2584-4c4e-915d-e55e0b997107\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401425-zxwfh" Nov 25 15:45:00 crc kubenswrapper[4890]: I1125 15:45:00.397062 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/59d4c133-2584-4c4e-915d-e55e0b997107-secret-volume\") pod \"collect-profiles-29401425-zxwfh\" (UID: \"59d4c133-2584-4c4e-915d-e55e0b997107\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401425-zxwfh" Nov 25 15:45:00 crc kubenswrapper[4890]: I1125 15:45:00.415593 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zwps\" (UniqueName: \"kubernetes.io/projected/59d4c133-2584-4c4e-915d-e55e0b997107-kube-api-access-5zwps\") pod \"collect-profiles-29401425-zxwfh\" (UID: \"59d4c133-2584-4c4e-915d-e55e0b997107\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401425-zxwfh" Nov 25 15:45:00 crc kubenswrapper[4890]: I1125 15:45:00.494370 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401425-zxwfh" Nov 25 15:45:00 crc kubenswrapper[4890]: I1125 15:45:00.984968 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401425-zxwfh"] Nov 25 15:45:01 crc kubenswrapper[4890]: W1125 15:45:01.003180 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59d4c133_2584_4c4e_915d_e55e0b997107.slice/crio-a00049e4eb59ec9627b72de41fa617e6027642d099d8a15af5a964f38283f2d8 WatchSource:0}: Error finding container a00049e4eb59ec9627b72de41fa617e6027642d099d8a15af5a964f38283f2d8: Status 404 returned error can't find the container with id a00049e4eb59ec9627b72de41fa617e6027642d099d8a15af5a964f38283f2d8 Nov 25 15:45:02 crc kubenswrapper[4890]: I1125 15:45:02.021356 4890 generic.go:334] "Generic (PLEG): container finished" podID="59d4c133-2584-4c4e-915d-e55e0b997107" containerID="343454f78e3dbcc792aeb828d0a0a69fd875d2689a4145276adb65ccbd394ec5" exitCode=0 Nov 25 15:45:02 crc kubenswrapper[4890]: I1125 15:45:02.021397 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401425-zxwfh" event={"ID":"59d4c133-2584-4c4e-915d-e55e0b997107","Type":"ContainerDied","Data":"343454f78e3dbcc792aeb828d0a0a69fd875d2689a4145276adb65ccbd394ec5"} Nov 25 15:45:02 crc kubenswrapper[4890]: I1125 15:45:02.021667 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401425-zxwfh" event={"ID":"59d4c133-2584-4c4e-915d-e55e0b997107","Type":"ContainerStarted","Data":"a00049e4eb59ec9627b72de41fa617e6027642d099d8a15af5a964f38283f2d8"} Nov 25 15:45:03 crc kubenswrapper[4890]: I1125 15:45:03.385848 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401425-zxwfh" Nov 25 15:45:03 crc kubenswrapper[4890]: I1125 15:45:03.553613 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/59d4c133-2584-4c4e-915d-e55e0b997107-config-volume\") pod \"59d4c133-2584-4c4e-915d-e55e0b997107\" (UID: \"59d4c133-2584-4c4e-915d-e55e0b997107\") " Nov 25 15:45:03 crc kubenswrapper[4890]: I1125 15:45:03.553754 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zwps\" (UniqueName: \"kubernetes.io/projected/59d4c133-2584-4c4e-915d-e55e0b997107-kube-api-access-5zwps\") pod \"59d4c133-2584-4c4e-915d-e55e0b997107\" (UID: \"59d4c133-2584-4c4e-915d-e55e0b997107\") " Nov 25 15:45:03 crc kubenswrapper[4890]: I1125 15:45:03.553856 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/59d4c133-2584-4c4e-915d-e55e0b997107-secret-volume\") pod \"59d4c133-2584-4c4e-915d-e55e0b997107\" (UID: \"59d4c133-2584-4c4e-915d-e55e0b997107\") " Nov 25 15:45:03 crc kubenswrapper[4890]: I1125 15:45:03.554565 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59d4c133-2584-4c4e-915d-e55e0b997107-config-volume" (OuterVolumeSpecName: "config-volume") pod "59d4c133-2584-4c4e-915d-e55e0b997107" (UID: "59d4c133-2584-4c4e-915d-e55e0b997107"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:45:03 crc kubenswrapper[4890]: I1125 15:45:03.561028 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59d4c133-2584-4c4e-915d-e55e0b997107-kube-api-access-5zwps" (OuterVolumeSpecName: "kube-api-access-5zwps") pod "59d4c133-2584-4c4e-915d-e55e0b997107" (UID: "59d4c133-2584-4c4e-915d-e55e0b997107"). InnerVolumeSpecName "kube-api-access-5zwps". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:45:03 crc kubenswrapper[4890]: I1125 15:45:03.562471 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59d4c133-2584-4c4e-915d-e55e0b997107-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "59d4c133-2584-4c4e-915d-e55e0b997107" (UID: "59d4c133-2584-4c4e-915d-e55e0b997107"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:45:03 crc kubenswrapper[4890]: I1125 15:45:03.657318 4890 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/59d4c133-2584-4c4e-915d-e55e0b997107-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 15:45:03 crc kubenswrapper[4890]: I1125 15:45:03.657351 4890 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/59d4c133-2584-4c4e-915d-e55e0b997107-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 15:45:03 crc kubenswrapper[4890]: I1125 15:45:03.657361 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zwps\" (UniqueName: \"kubernetes.io/projected/59d4c133-2584-4c4e-915d-e55e0b997107-kube-api-access-5zwps\") on node \"crc\" DevicePath \"\"" Nov 25 15:45:04 crc kubenswrapper[4890]: I1125 15:45:04.043635 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401425-zxwfh" event={"ID":"59d4c133-2584-4c4e-915d-e55e0b997107","Type":"ContainerDied","Data":"a00049e4eb59ec9627b72de41fa617e6027642d099d8a15af5a964f38283f2d8"} Nov 25 15:45:04 crc kubenswrapper[4890]: I1125 15:45:04.043692 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a00049e4eb59ec9627b72de41fa617e6027642d099d8a15af5a964f38283f2d8" Nov 25 15:45:04 crc kubenswrapper[4890]: I1125 15:45:04.043719 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401425-zxwfh" Nov 25 15:45:04 crc kubenswrapper[4890]: I1125 15:45:04.482255 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401380-96zpb"] Nov 25 15:45:04 crc kubenswrapper[4890]: I1125 15:45:04.490854 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401380-96zpb"] Nov 25 15:45:06 crc kubenswrapper[4890]: I1125 15:45:06.189844 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4622b25d-fb03-4e9f-a66c-7f90def9b5f7" path="/var/lib/kubelet/pods/4622b25d-fb03-4e9f-a66c-7f90def9b5f7/volumes" Nov 25 15:45:39 crc kubenswrapper[4890]: I1125 15:45:39.525475 4890 scope.go:117] "RemoveContainer" containerID="d69aed072dd6ab3043fb1ed07aa7e23803f193be0d9017879940b070a3669203" Nov 25 15:46:51 crc kubenswrapper[4890]: I1125 15:46:51.496434 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-25cgx"] Nov 25 15:46:51 crc kubenswrapper[4890]: E1125 15:46:51.497513 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59d4c133-2584-4c4e-915d-e55e0b997107" containerName="collect-profiles" Nov 25 15:46:51 crc kubenswrapper[4890]: I1125 15:46:51.497530 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="59d4c133-2584-4c4e-915d-e55e0b997107" containerName="collect-profiles" Nov 25 15:46:51 crc kubenswrapper[4890]: I1125 15:46:51.497708 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="59d4c133-2584-4c4e-915d-e55e0b997107" containerName="collect-profiles" Nov 25 15:46:51 crc kubenswrapper[4890]: I1125 15:46:51.499084 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-25cgx" Nov 25 15:46:51 crc kubenswrapper[4890]: I1125 15:46:51.515981 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-25cgx"] Nov 25 15:46:51 crc kubenswrapper[4890]: I1125 15:46:51.683131 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8t795\" (UniqueName: \"kubernetes.io/projected/00720827-a9eb-4fe2-859a-792ff122c0ef-kube-api-access-8t795\") pod \"certified-operators-25cgx\" (UID: \"00720827-a9eb-4fe2-859a-792ff122c0ef\") " pod="openshift-marketplace/certified-operators-25cgx" Nov 25 15:46:51 crc kubenswrapper[4890]: I1125 15:46:51.683221 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00720827-a9eb-4fe2-859a-792ff122c0ef-utilities\") pod \"certified-operators-25cgx\" (UID: \"00720827-a9eb-4fe2-859a-792ff122c0ef\") " pod="openshift-marketplace/certified-operators-25cgx" Nov 25 15:46:51 crc kubenswrapper[4890]: I1125 15:46:51.684458 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00720827-a9eb-4fe2-859a-792ff122c0ef-catalog-content\") pod \"certified-operators-25cgx\" (UID: \"00720827-a9eb-4fe2-859a-792ff122c0ef\") " pod="openshift-marketplace/certified-operators-25cgx" Nov 25 15:46:51 crc kubenswrapper[4890]: I1125 15:46:51.786747 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00720827-a9eb-4fe2-859a-792ff122c0ef-catalog-content\") pod \"certified-operators-25cgx\" (UID: \"00720827-a9eb-4fe2-859a-792ff122c0ef\") " pod="openshift-marketplace/certified-operators-25cgx" Nov 25 15:46:51 crc kubenswrapper[4890]: I1125 15:46:51.786989 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8t795\" (UniqueName: \"kubernetes.io/projected/00720827-a9eb-4fe2-859a-792ff122c0ef-kube-api-access-8t795\") pod \"certified-operators-25cgx\" (UID: \"00720827-a9eb-4fe2-859a-792ff122c0ef\") " pod="openshift-marketplace/certified-operators-25cgx" Nov 25 15:46:51 crc kubenswrapper[4890]: I1125 15:46:51.787024 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00720827-a9eb-4fe2-859a-792ff122c0ef-utilities\") pod \"certified-operators-25cgx\" (UID: \"00720827-a9eb-4fe2-859a-792ff122c0ef\") " pod="openshift-marketplace/certified-operators-25cgx" Nov 25 15:46:51 crc kubenswrapper[4890]: I1125 15:46:51.787419 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00720827-a9eb-4fe2-859a-792ff122c0ef-catalog-content\") pod \"certified-operators-25cgx\" (UID: \"00720827-a9eb-4fe2-859a-792ff122c0ef\") " pod="openshift-marketplace/certified-operators-25cgx" Nov 25 15:46:51 crc kubenswrapper[4890]: I1125 15:46:51.788924 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00720827-a9eb-4fe2-859a-792ff122c0ef-utilities\") pod \"certified-operators-25cgx\" (UID: \"00720827-a9eb-4fe2-859a-792ff122c0ef\") " pod="openshift-marketplace/certified-operators-25cgx" Nov 25 15:46:51 crc kubenswrapper[4890]: I1125 15:46:51.814057 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8t795\" (UniqueName: \"kubernetes.io/projected/00720827-a9eb-4fe2-859a-792ff122c0ef-kube-api-access-8t795\") pod \"certified-operators-25cgx\" (UID: \"00720827-a9eb-4fe2-859a-792ff122c0ef\") " pod="openshift-marketplace/certified-operators-25cgx" Nov 25 15:46:51 crc kubenswrapper[4890]: I1125 15:46:51.825295 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-25cgx" Nov 25 15:46:52 crc kubenswrapper[4890]: I1125 15:46:52.436439 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-25cgx"] Nov 25 15:46:53 crc kubenswrapper[4890]: I1125 15:46:53.114066 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-25cgx" event={"ID":"00720827-a9eb-4fe2-859a-792ff122c0ef","Type":"ContainerStarted","Data":"79f468bba9912200308199ae68bbcdd8aa35aca16f2fd3bd09bba697bc426fbf"} Nov 25 15:46:53 crc kubenswrapper[4890]: I1125 15:46:53.114124 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-25cgx" event={"ID":"00720827-a9eb-4fe2-859a-792ff122c0ef","Type":"ContainerStarted","Data":"7576d8b782376543785bbc3e2339f3e1076ce9d45e091a6f52dc68679c9f749a"} Nov 25 15:46:54 crc kubenswrapper[4890]: I1125 15:46:54.132699 4890 generic.go:334] "Generic (PLEG): container finished" podID="00720827-a9eb-4fe2-859a-792ff122c0ef" containerID="79f468bba9912200308199ae68bbcdd8aa35aca16f2fd3bd09bba697bc426fbf" exitCode=0 Nov 25 15:46:54 crc kubenswrapper[4890]: I1125 15:46:54.132786 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-25cgx" event={"ID":"00720827-a9eb-4fe2-859a-792ff122c0ef","Type":"ContainerDied","Data":"79f468bba9912200308199ae68bbcdd8aa35aca16f2fd3bd09bba697bc426fbf"} Nov 25 15:46:56 crc kubenswrapper[4890]: I1125 15:46:56.157422 4890 generic.go:334] "Generic (PLEG): container finished" podID="00720827-a9eb-4fe2-859a-792ff122c0ef" containerID="3052e0c81565aa4a5bd2d86213f0e3237fa2f2d064aeb247b5cc1b217c88f0ef" exitCode=0 Nov 25 15:46:56 crc kubenswrapper[4890]: I1125 15:46:56.157536 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-25cgx" event={"ID":"00720827-a9eb-4fe2-859a-792ff122c0ef","Type":"ContainerDied","Data":"3052e0c81565aa4a5bd2d86213f0e3237fa2f2d064aeb247b5cc1b217c88f0ef"} Nov 25 15:46:56 crc kubenswrapper[4890]: I1125 15:46:56.447501 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 15:46:56 crc kubenswrapper[4890]: I1125 15:46:56.447607 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 15:46:58 crc kubenswrapper[4890]: I1125 15:46:58.186569 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-25cgx" event={"ID":"00720827-a9eb-4fe2-859a-792ff122c0ef","Type":"ContainerStarted","Data":"47b8c53403e63ed3ef82270e4d951c4f015aa63383a40b0014c1d582e2d376b7"} Nov 25 15:46:58 crc kubenswrapper[4890]: I1125 15:46:58.213054 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-25cgx" podStartSLOduration=4.497168658 podStartE2EDuration="7.2130275s" podCreationTimestamp="2025-11-25 15:46:51 +0000 UTC" firstStartedPulling="2025-11-25 15:46:54.137927853 +0000 UTC m=+2672.580390453" lastFinishedPulling="2025-11-25 15:46:56.853786685 +0000 UTC m=+2675.296249295" observedRunningTime="2025-11-25 15:46:58.207452641 +0000 UTC m=+2676.649915261" watchObservedRunningTime="2025-11-25 15:46:58.2130275 +0000 UTC m=+2676.655490110" Nov 25 15:47:01 crc kubenswrapper[4890]: I1125 15:47:01.825742 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-25cgx" Nov 25 15:47:01 crc kubenswrapper[4890]: I1125 15:47:01.826683 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-25cgx" Nov 25 15:47:01 crc kubenswrapper[4890]: I1125 15:47:01.879645 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-25cgx" Nov 25 15:47:02 crc kubenswrapper[4890]: I1125 15:47:02.272254 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-25cgx" Nov 25 15:47:02 crc kubenswrapper[4890]: I1125 15:47:02.320971 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-25cgx"] Nov 25 15:47:04 crc kubenswrapper[4890]: I1125 15:47:04.237842 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-25cgx" podUID="00720827-a9eb-4fe2-859a-792ff122c0ef" containerName="registry-server" containerID="cri-o://47b8c53403e63ed3ef82270e4d951c4f015aa63383a40b0014c1d582e2d376b7" gracePeriod=2 Nov 25 15:47:05 crc kubenswrapper[4890]: I1125 15:47:05.254803 4890 generic.go:334] "Generic (PLEG): container finished" podID="00720827-a9eb-4fe2-859a-792ff122c0ef" containerID="47b8c53403e63ed3ef82270e4d951c4f015aa63383a40b0014c1d582e2d376b7" exitCode=0 Nov 25 15:47:05 crc kubenswrapper[4890]: I1125 15:47:05.254981 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-25cgx" event={"ID":"00720827-a9eb-4fe2-859a-792ff122c0ef","Type":"ContainerDied","Data":"47b8c53403e63ed3ef82270e4d951c4f015aa63383a40b0014c1d582e2d376b7"} Nov 25 15:47:05 crc kubenswrapper[4890]: I1125 15:47:05.255847 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-25cgx" event={"ID":"00720827-a9eb-4fe2-859a-792ff122c0ef","Type":"ContainerDied","Data":"7576d8b782376543785bbc3e2339f3e1076ce9d45e091a6f52dc68679c9f749a"} Nov 25 15:47:05 crc kubenswrapper[4890]: I1125 15:47:05.255877 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7576d8b782376543785bbc3e2339f3e1076ce9d45e091a6f52dc68679c9f749a" Nov 25 15:47:05 crc kubenswrapper[4890]: I1125 15:47:05.318307 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-25cgx" Nov 25 15:47:05 crc kubenswrapper[4890]: I1125 15:47:05.432985 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00720827-a9eb-4fe2-859a-792ff122c0ef-utilities\") pod \"00720827-a9eb-4fe2-859a-792ff122c0ef\" (UID: \"00720827-a9eb-4fe2-859a-792ff122c0ef\") " Nov 25 15:47:05 crc kubenswrapper[4890]: I1125 15:47:05.433230 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00720827-a9eb-4fe2-859a-792ff122c0ef-catalog-content\") pod \"00720827-a9eb-4fe2-859a-792ff122c0ef\" (UID: \"00720827-a9eb-4fe2-859a-792ff122c0ef\") " Nov 25 15:47:05 crc kubenswrapper[4890]: I1125 15:47:05.433364 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8t795\" (UniqueName: \"kubernetes.io/projected/00720827-a9eb-4fe2-859a-792ff122c0ef-kube-api-access-8t795\") pod \"00720827-a9eb-4fe2-859a-792ff122c0ef\" (UID: \"00720827-a9eb-4fe2-859a-792ff122c0ef\") " Nov 25 15:47:05 crc kubenswrapper[4890]: I1125 15:47:05.434812 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00720827-a9eb-4fe2-859a-792ff122c0ef-utilities" (OuterVolumeSpecName: "utilities") pod "00720827-a9eb-4fe2-859a-792ff122c0ef" (UID: "00720827-a9eb-4fe2-859a-792ff122c0ef"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:47:05 crc kubenswrapper[4890]: I1125 15:47:05.447123 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00720827-a9eb-4fe2-859a-792ff122c0ef-kube-api-access-8t795" (OuterVolumeSpecName: "kube-api-access-8t795") pod "00720827-a9eb-4fe2-859a-792ff122c0ef" (UID: "00720827-a9eb-4fe2-859a-792ff122c0ef"). InnerVolumeSpecName "kube-api-access-8t795". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:47:05 crc kubenswrapper[4890]: I1125 15:47:05.506559 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00720827-a9eb-4fe2-859a-792ff122c0ef-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "00720827-a9eb-4fe2-859a-792ff122c0ef" (UID: "00720827-a9eb-4fe2-859a-792ff122c0ef"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:47:05 crc kubenswrapper[4890]: I1125 15:47:05.535425 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00720827-a9eb-4fe2-859a-792ff122c0ef-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 15:47:05 crc kubenswrapper[4890]: I1125 15:47:05.535706 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00720827-a9eb-4fe2-859a-792ff122c0ef-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 15:47:05 crc kubenswrapper[4890]: I1125 15:47:05.535924 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8t795\" (UniqueName: \"kubernetes.io/projected/00720827-a9eb-4fe2-859a-792ff122c0ef-kube-api-access-8t795\") on node \"crc\" DevicePath \"\"" Nov 25 15:47:06 crc kubenswrapper[4890]: I1125 15:47:06.268336 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-25cgx" Nov 25 15:47:06 crc kubenswrapper[4890]: I1125 15:47:06.297306 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-25cgx"] Nov 25 15:47:06 crc kubenswrapper[4890]: I1125 15:47:06.306469 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-25cgx"] Nov 25 15:47:08 crc kubenswrapper[4890]: I1125 15:47:08.182322 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00720827-a9eb-4fe2-859a-792ff122c0ef" path="/var/lib/kubelet/pods/00720827-a9eb-4fe2-859a-792ff122c0ef/volumes" Nov 25 15:47:26 crc kubenswrapper[4890]: I1125 15:47:26.447103 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 15:47:26 crc kubenswrapper[4890]: I1125 15:47:26.447874 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 15:47:31 crc kubenswrapper[4890]: I1125 15:47:31.561981 4890 generic.go:334] "Generic (PLEG): container finished" podID="e890bd11-32e9-4d73-9dd0-ae5d69a1fe58" containerID="8e79fadd953a14eed3e5a375d82ee1c10b4dd4da678f950428948e536b223ba0" exitCode=0 Nov 25 15:47:31 crc kubenswrapper[4890]: I1125 15:47:31.562075 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-49nlj" event={"ID":"e890bd11-32e9-4d73-9dd0-ae5d69a1fe58","Type":"ContainerDied","Data":"8e79fadd953a14eed3e5a375d82ee1c10b4dd4da678f950428948e536b223ba0"} Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.036433 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-49nlj" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.225114 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qpnl2\" (UniqueName: \"kubernetes.io/projected/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-kube-api-access-qpnl2\") pod \"e890bd11-32e9-4d73-9dd0-ae5d69a1fe58\" (UID: \"e890bd11-32e9-4d73-9dd0-ae5d69a1fe58\") " Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.225191 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-nova-cell1-compute-config-1\") pod \"e890bd11-32e9-4d73-9dd0-ae5d69a1fe58\" (UID: \"e890bd11-32e9-4d73-9dd0-ae5d69a1fe58\") " Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.225324 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-nova-cell1-compute-config-0\") pod \"e890bd11-32e9-4d73-9dd0-ae5d69a1fe58\" (UID: \"e890bd11-32e9-4d73-9dd0-ae5d69a1fe58\") " Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.225377 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-nova-combined-ca-bundle\") pod \"e890bd11-32e9-4d73-9dd0-ae5d69a1fe58\" (UID: \"e890bd11-32e9-4d73-9dd0-ae5d69a1fe58\") " Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.225430 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-nova-extra-config-0\") pod \"e890bd11-32e9-4d73-9dd0-ae5d69a1fe58\" (UID: \"e890bd11-32e9-4d73-9dd0-ae5d69a1fe58\") " Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.225482 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-nova-migration-ssh-key-0\") pod \"e890bd11-32e9-4d73-9dd0-ae5d69a1fe58\" (UID: \"e890bd11-32e9-4d73-9dd0-ae5d69a1fe58\") " Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.225513 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-ssh-key\") pod \"e890bd11-32e9-4d73-9dd0-ae5d69a1fe58\" (UID: \"e890bd11-32e9-4d73-9dd0-ae5d69a1fe58\") " Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.225557 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-nova-migration-ssh-key-1\") pod \"e890bd11-32e9-4d73-9dd0-ae5d69a1fe58\" (UID: \"e890bd11-32e9-4d73-9dd0-ae5d69a1fe58\") " Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.225581 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-inventory\") pod \"e890bd11-32e9-4d73-9dd0-ae5d69a1fe58\" (UID: \"e890bd11-32e9-4d73-9dd0-ae5d69a1fe58\") " Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.240128 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-kube-api-access-qpnl2" (OuterVolumeSpecName: "kube-api-access-qpnl2") pod "e890bd11-32e9-4d73-9dd0-ae5d69a1fe58" (UID: "e890bd11-32e9-4d73-9dd0-ae5d69a1fe58"). InnerVolumeSpecName "kube-api-access-qpnl2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.240177 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "e890bd11-32e9-4d73-9dd0-ae5d69a1fe58" (UID: "e890bd11-32e9-4d73-9dd0-ae5d69a1fe58"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.254748 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "e890bd11-32e9-4d73-9dd0-ae5d69a1fe58" (UID: "e890bd11-32e9-4d73-9dd0-ae5d69a1fe58"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.256789 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e890bd11-32e9-4d73-9dd0-ae5d69a1fe58" (UID: "e890bd11-32e9-4d73-9dd0-ae5d69a1fe58"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.260726 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "e890bd11-32e9-4d73-9dd0-ae5d69a1fe58" (UID: "e890bd11-32e9-4d73-9dd0-ae5d69a1fe58"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.267293 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "e890bd11-32e9-4d73-9dd0-ae5d69a1fe58" (UID: "e890bd11-32e9-4d73-9dd0-ae5d69a1fe58"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.276145 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "e890bd11-32e9-4d73-9dd0-ae5d69a1fe58" (UID: "e890bd11-32e9-4d73-9dd0-ae5d69a1fe58"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.278901 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "e890bd11-32e9-4d73-9dd0-ae5d69a1fe58" (UID: "e890bd11-32e9-4d73-9dd0-ae5d69a1fe58"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.285782 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-inventory" (OuterVolumeSpecName: "inventory") pod "e890bd11-32e9-4d73-9dd0-ae5d69a1fe58" (UID: "e890bd11-32e9-4d73-9dd0-ae5d69a1fe58"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.328424 4890 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.328461 4890 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.328473 4890 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.328484 4890 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.328493 4890 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.328502 4890 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.328512 4890 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.328521 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qpnl2\" (UniqueName: \"kubernetes.io/projected/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-kube-api-access-qpnl2\") on node \"crc\" DevicePath \"\"" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.328528 4890 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/e890bd11-32e9-4d73-9dd0-ae5d69a1fe58-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.583040 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-49nlj" event={"ID":"e890bd11-32e9-4d73-9dd0-ae5d69a1fe58","Type":"ContainerDied","Data":"2cccef917722046c030b3a1affff4121129ec2a46ed5427e2a2f7fec578c60ce"} Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.583096 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2cccef917722046c030b3a1affff4121129ec2a46ed5427e2a2f7fec578c60ce" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.583199 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-49nlj" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.684066 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jm99p"] Nov 25 15:47:33 crc kubenswrapper[4890]: E1125 15:47:33.684632 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00720827-a9eb-4fe2-859a-792ff122c0ef" containerName="registry-server" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.684658 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="00720827-a9eb-4fe2-859a-792ff122c0ef" containerName="registry-server" Nov 25 15:47:33 crc kubenswrapper[4890]: E1125 15:47:33.684672 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00720827-a9eb-4fe2-859a-792ff122c0ef" containerName="extract-utilities" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.684681 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="00720827-a9eb-4fe2-859a-792ff122c0ef" containerName="extract-utilities" Nov 25 15:47:33 crc kubenswrapper[4890]: E1125 15:47:33.684734 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e890bd11-32e9-4d73-9dd0-ae5d69a1fe58" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.684744 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="e890bd11-32e9-4d73-9dd0-ae5d69a1fe58" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 25 15:47:33 crc kubenswrapper[4890]: E1125 15:47:33.684770 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00720827-a9eb-4fe2-859a-792ff122c0ef" containerName="extract-content" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.684779 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="00720827-a9eb-4fe2-859a-792ff122c0ef" containerName="extract-content" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.685018 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="e890bd11-32e9-4d73-9dd0-ae5d69a1fe58" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.685056 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="00720827-a9eb-4fe2-859a-792ff122c0ef" containerName="registry-server" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.686003 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jm99p" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.688217 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.688887 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-22zzp" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.689531 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.689573 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.689691 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.701073 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jm99p"] Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.839929 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/588df9b2-9cfb-449a-9dca-17912f88a4c6-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-jm99p\" (UID: \"588df9b2-9cfb-449a-9dca-17912f88a4c6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jm99p" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.839985 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/588df9b2-9cfb-449a-9dca-17912f88a4c6-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-jm99p\" (UID: \"588df9b2-9cfb-449a-9dca-17912f88a4c6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jm99p" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.840007 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/588df9b2-9cfb-449a-9dca-17912f88a4c6-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-jm99p\" (UID: \"588df9b2-9cfb-449a-9dca-17912f88a4c6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jm99p" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.840126 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/588df9b2-9cfb-449a-9dca-17912f88a4c6-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-jm99p\" (UID: \"588df9b2-9cfb-449a-9dca-17912f88a4c6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jm99p" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.840149 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/588df9b2-9cfb-449a-9dca-17912f88a4c6-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-jm99p\" (UID: \"588df9b2-9cfb-449a-9dca-17912f88a4c6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jm99p" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.840202 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/588df9b2-9cfb-449a-9dca-17912f88a4c6-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-jm99p\" (UID: \"588df9b2-9cfb-449a-9dca-17912f88a4c6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jm99p" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.840251 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hjws\" (UniqueName: \"kubernetes.io/projected/588df9b2-9cfb-449a-9dca-17912f88a4c6-kube-api-access-7hjws\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-jm99p\" (UID: \"588df9b2-9cfb-449a-9dca-17912f88a4c6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jm99p" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.942140 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/588df9b2-9cfb-449a-9dca-17912f88a4c6-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-jm99p\" (UID: \"588df9b2-9cfb-449a-9dca-17912f88a4c6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jm99p" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.942232 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hjws\" (UniqueName: \"kubernetes.io/projected/588df9b2-9cfb-449a-9dca-17912f88a4c6-kube-api-access-7hjws\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-jm99p\" (UID: \"588df9b2-9cfb-449a-9dca-17912f88a4c6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jm99p" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.942330 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/588df9b2-9cfb-449a-9dca-17912f88a4c6-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-jm99p\" (UID: \"588df9b2-9cfb-449a-9dca-17912f88a4c6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jm99p" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.942367 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/588df9b2-9cfb-449a-9dca-17912f88a4c6-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-jm99p\" (UID: \"588df9b2-9cfb-449a-9dca-17912f88a4c6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jm99p" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.942392 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/588df9b2-9cfb-449a-9dca-17912f88a4c6-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-jm99p\" (UID: \"588df9b2-9cfb-449a-9dca-17912f88a4c6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jm99p" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.942438 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/588df9b2-9cfb-449a-9dca-17912f88a4c6-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-jm99p\" (UID: \"588df9b2-9cfb-449a-9dca-17912f88a4c6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jm99p" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.942789 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/588df9b2-9cfb-449a-9dca-17912f88a4c6-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-jm99p\" (UID: \"588df9b2-9cfb-449a-9dca-17912f88a4c6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jm99p" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.948034 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/588df9b2-9cfb-449a-9dca-17912f88a4c6-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-jm99p\" (UID: \"588df9b2-9cfb-449a-9dca-17912f88a4c6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jm99p" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.948296 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/588df9b2-9cfb-449a-9dca-17912f88a4c6-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-jm99p\" (UID: \"588df9b2-9cfb-449a-9dca-17912f88a4c6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jm99p" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.948336 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/588df9b2-9cfb-449a-9dca-17912f88a4c6-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-jm99p\" (UID: \"588df9b2-9cfb-449a-9dca-17912f88a4c6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jm99p" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.948368 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/588df9b2-9cfb-449a-9dca-17912f88a4c6-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-jm99p\" (UID: \"588df9b2-9cfb-449a-9dca-17912f88a4c6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jm99p" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.948753 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/588df9b2-9cfb-449a-9dca-17912f88a4c6-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-jm99p\" (UID: \"588df9b2-9cfb-449a-9dca-17912f88a4c6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jm99p" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.949074 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/588df9b2-9cfb-449a-9dca-17912f88a4c6-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-jm99p\" (UID: \"588df9b2-9cfb-449a-9dca-17912f88a4c6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jm99p" Nov 25 15:47:33 crc kubenswrapper[4890]: I1125 15:47:33.959475 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hjws\" (UniqueName: \"kubernetes.io/projected/588df9b2-9cfb-449a-9dca-17912f88a4c6-kube-api-access-7hjws\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-jm99p\" (UID: \"588df9b2-9cfb-449a-9dca-17912f88a4c6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jm99p" Nov 25 15:47:34 crc kubenswrapper[4890]: I1125 15:47:34.010228 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jm99p" Nov 25 15:47:34 crc kubenswrapper[4890]: I1125 15:47:34.604022 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jm99p"] Nov 25 15:47:34 crc kubenswrapper[4890]: I1125 15:47:34.610213 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jm99p" event={"ID":"588df9b2-9cfb-449a-9dca-17912f88a4c6","Type":"ContainerStarted","Data":"f397a631fe6c7432f11e02914e49c914480a065970dab2596ec24f50cc07cfda"} Nov 25 15:47:35 crc kubenswrapper[4890]: I1125 15:47:35.623060 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jm99p" event={"ID":"588df9b2-9cfb-449a-9dca-17912f88a4c6","Type":"ContainerStarted","Data":"31b48008cfb1224efade9a7153cde8a26abfd8d9a9e40c470aeffde54afcd9e8"} Nov 25 15:47:35 crc kubenswrapper[4890]: I1125 15:47:35.649250 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jm99p" podStartSLOduration=2.13316995 podStartE2EDuration="2.649226292s" podCreationTimestamp="2025-11-25 15:47:33 +0000 UTC" firstStartedPulling="2025-11-25 15:47:34.594092794 +0000 UTC m=+2713.036555404" lastFinishedPulling="2025-11-25 15:47:35.110149136 +0000 UTC m=+2713.552611746" observedRunningTime="2025-11-25 15:47:35.642569436 +0000 UTC m=+2714.085032056" watchObservedRunningTime="2025-11-25 15:47:35.649226292 +0000 UTC m=+2714.091688902" Nov 25 15:47:56 crc kubenswrapper[4890]: I1125 15:47:56.447977 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 15:47:56 crc kubenswrapper[4890]: I1125 15:47:56.448540 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 15:47:56 crc kubenswrapper[4890]: I1125 15:47:56.448585 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" Nov 25 15:47:56 crc kubenswrapper[4890]: I1125 15:47:56.449277 4890 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3add3ce883ec44a5429bf3e4879e91159a451c26e6f8d63ebfd5527dc360e6b8"} pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 15:47:56 crc kubenswrapper[4890]: I1125 15:47:56.449327 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" containerID="cri-o://3add3ce883ec44a5429bf3e4879e91159a451c26e6f8d63ebfd5527dc360e6b8" gracePeriod=600 Nov 25 15:47:56 crc kubenswrapper[4890]: I1125 15:47:56.846869 4890 generic.go:334] "Generic (PLEG): container finished" podID="4e4f849d-f239-4727-a73e-18327856929a" containerID="3add3ce883ec44a5429bf3e4879e91159a451c26e6f8d63ebfd5527dc360e6b8" exitCode=0 Nov 25 15:47:56 crc kubenswrapper[4890]: I1125 15:47:56.846937 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" event={"ID":"4e4f849d-f239-4727-a73e-18327856929a","Type":"ContainerDied","Data":"3add3ce883ec44a5429bf3e4879e91159a451c26e6f8d63ebfd5527dc360e6b8"} Nov 25 15:47:56 crc kubenswrapper[4890]: I1125 15:47:56.847270 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" event={"ID":"4e4f849d-f239-4727-a73e-18327856929a","Type":"ContainerStarted","Data":"aa8a6c3d508388bee84226737d06d0443dd3e6d511ad22e6e9e1c1f48362f983"} Nov 25 15:47:56 crc kubenswrapper[4890]: I1125 15:47:56.847301 4890 scope.go:117] "RemoveContainer" containerID="67ca0de5cd1ef0f73ff70ee2df49e62ca714111ad20a0de1fef88f6f60fdf262" Nov 25 15:49:42 crc kubenswrapper[4890]: I1125 15:49:42.009099 4890 generic.go:334] "Generic (PLEG): container finished" podID="588df9b2-9cfb-449a-9dca-17912f88a4c6" containerID="31b48008cfb1224efade9a7153cde8a26abfd8d9a9e40c470aeffde54afcd9e8" exitCode=0 Nov 25 15:49:42 crc kubenswrapper[4890]: I1125 15:49:42.009208 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jm99p" event={"ID":"588df9b2-9cfb-449a-9dca-17912f88a4c6","Type":"ContainerDied","Data":"31b48008cfb1224efade9a7153cde8a26abfd8d9a9e40c470aeffde54afcd9e8"} Nov 25 15:49:44 crc kubenswrapper[4890]: I1125 15:49:43.536545 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jm99p" Nov 25 15:49:44 crc kubenswrapper[4890]: I1125 15:49:43.618698 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/588df9b2-9cfb-449a-9dca-17912f88a4c6-ceilometer-compute-config-data-0\") pod \"588df9b2-9cfb-449a-9dca-17912f88a4c6\" (UID: \"588df9b2-9cfb-449a-9dca-17912f88a4c6\") " Nov 25 15:49:44 crc kubenswrapper[4890]: I1125 15:49:43.618846 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7hjws\" (UniqueName: \"kubernetes.io/projected/588df9b2-9cfb-449a-9dca-17912f88a4c6-kube-api-access-7hjws\") pod \"588df9b2-9cfb-449a-9dca-17912f88a4c6\" (UID: \"588df9b2-9cfb-449a-9dca-17912f88a4c6\") " Nov 25 15:49:44 crc kubenswrapper[4890]: I1125 15:49:43.618882 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/588df9b2-9cfb-449a-9dca-17912f88a4c6-inventory\") pod \"588df9b2-9cfb-449a-9dca-17912f88a4c6\" (UID: \"588df9b2-9cfb-449a-9dca-17912f88a4c6\") " Nov 25 15:49:44 crc kubenswrapper[4890]: I1125 15:49:43.619016 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/588df9b2-9cfb-449a-9dca-17912f88a4c6-ceilometer-compute-config-data-1\") pod \"588df9b2-9cfb-449a-9dca-17912f88a4c6\" (UID: \"588df9b2-9cfb-449a-9dca-17912f88a4c6\") " Nov 25 15:49:44 crc kubenswrapper[4890]: I1125 15:49:43.619135 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/588df9b2-9cfb-449a-9dca-17912f88a4c6-ssh-key\") pod \"588df9b2-9cfb-449a-9dca-17912f88a4c6\" (UID: \"588df9b2-9cfb-449a-9dca-17912f88a4c6\") " Nov 25 15:49:44 crc kubenswrapper[4890]: I1125 15:49:43.619257 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/588df9b2-9cfb-449a-9dca-17912f88a4c6-ceilometer-compute-config-data-2\") pod \"588df9b2-9cfb-449a-9dca-17912f88a4c6\" (UID: \"588df9b2-9cfb-449a-9dca-17912f88a4c6\") " Nov 25 15:49:44 crc kubenswrapper[4890]: I1125 15:49:43.619334 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/588df9b2-9cfb-449a-9dca-17912f88a4c6-telemetry-combined-ca-bundle\") pod \"588df9b2-9cfb-449a-9dca-17912f88a4c6\" (UID: \"588df9b2-9cfb-449a-9dca-17912f88a4c6\") " Nov 25 15:49:44 crc kubenswrapper[4890]: I1125 15:49:43.626818 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/588df9b2-9cfb-449a-9dca-17912f88a4c6-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "588df9b2-9cfb-449a-9dca-17912f88a4c6" (UID: "588df9b2-9cfb-449a-9dca-17912f88a4c6"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:49:44 crc kubenswrapper[4890]: I1125 15:49:43.626785 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/588df9b2-9cfb-449a-9dca-17912f88a4c6-kube-api-access-7hjws" (OuterVolumeSpecName: "kube-api-access-7hjws") pod "588df9b2-9cfb-449a-9dca-17912f88a4c6" (UID: "588df9b2-9cfb-449a-9dca-17912f88a4c6"). InnerVolumeSpecName "kube-api-access-7hjws". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:49:44 crc kubenswrapper[4890]: I1125 15:49:43.654974 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/588df9b2-9cfb-449a-9dca-17912f88a4c6-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "588df9b2-9cfb-449a-9dca-17912f88a4c6" (UID: "588df9b2-9cfb-449a-9dca-17912f88a4c6"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:49:44 crc kubenswrapper[4890]: I1125 15:49:43.666950 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/588df9b2-9cfb-449a-9dca-17912f88a4c6-inventory" (OuterVolumeSpecName: "inventory") pod "588df9b2-9cfb-449a-9dca-17912f88a4c6" (UID: "588df9b2-9cfb-449a-9dca-17912f88a4c6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:49:44 crc kubenswrapper[4890]: I1125 15:49:43.667334 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/588df9b2-9cfb-449a-9dca-17912f88a4c6-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "588df9b2-9cfb-449a-9dca-17912f88a4c6" (UID: "588df9b2-9cfb-449a-9dca-17912f88a4c6"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:49:44 crc kubenswrapper[4890]: I1125 15:49:43.670043 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/588df9b2-9cfb-449a-9dca-17912f88a4c6-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "588df9b2-9cfb-449a-9dca-17912f88a4c6" (UID: "588df9b2-9cfb-449a-9dca-17912f88a4c6"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:49:44 crc kubenswrapper[4890]: I1125 15:49:43.679999 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/588df9b2-9cfb-449a-9dca-17912f88a4c6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "588df9b2-9cfb-449a-9dca-17912f88a4c6" (UID: "588df9b2-9cfb-449a-9dca-17912f88a4c6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 15:49:44 crc kubenswrapper[4890]: I1125 15:49:43.721784 4890 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/588df9b2-9cfb-449a-9dca-17912f88a4c6-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Nov 25 15:49:44 crc kubenswrapper[4890]: I1125 15:49:43.721816 4890 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/588df9b2-9cfb-449a-9dca-17912f88a4c6-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 15:49:44 crc kubenswrapper[4890]: I1125 15:49:43.721825 4890 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/588df9b2-9cfb-449a-9dca-17912f88a4c6-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Nov 25 15:49:44 crc kubenswrapper[4890]: I1125 15:49:43.721836 4890 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/588df9b2-9cfb-449a-9dca-17912f88a4c6-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 15:49:44 crc kubenswrapper[4890]: I1125 15:49:43.721848 4890 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/588df9b2-9cfb-449a-9dca-17912f88a4c6-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Nov 25 15:49:44 crc kubenswrapper[4890]: I1125 15:49:43.721859 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7hjws\" (UniqueName: \"kubernetes.io/projected/588df9b2-9cfb-449a-9dca-17912f88a4c6-kube-api-access-7hjws\") on node \"crc\" DevicePath \"\"" Nov 25 15:49:44 crc kubenswrapper[4890]: I1125 15:49:43.721868 4890 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/588df9b2-9cfb-449a-9dca-17912f88a4c6-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 15:49:44 crc kubenswrapper[4890]: I1125 15:49:44.036534 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jm99p" event={"ID":"588df9b2-9cfb-449a-9dca-17912f88a4c6","Type":"ContainerDied","Data":"f397a631fe6c7432f11e02914e49c914480a065970dab2596ec24f50cc07cfda"} Nov 25 15:49:44 crc kubenswrapper[4890]: I1125 15:49:44.036602 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f397a631fe6c7432f11e02914e49c914480a065970dab2596ec24f50cc07cfda" Nov 25 15:49:44 crc kubenswrapper[4890]: I1125 15:49:44.036665 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-jm99p" Nov 25 15:49:56 crc kubenswrapper[4890]: I1125 15:49:56.446926 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 15:49:56 crc kubenswrapper[4890]: I1125 15:49:56.447595 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 15:50:08 crc kubenswrapper[4890]: I1125 15:50:08.902103 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ww2d5"] Nov 25 15:50:08 crc kubenswrapper[4890]: E1125 15:50:08.903139 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="588df9b2-9cfb-449a-9dca-17912f88a4c6" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 25 15:50:08 crc kubenswrapper[4890]: I1125 15:50:08.903179 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="588df9b2-9cfb-449a-9dca-17912f88a4c6" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 25 15:50:08 crc kubenswrapper[4890]: I1125 15:50:08.903342 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="588df9b2-9cfb-449a-9dca-17912f88a4c6" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 25 15:50:08 crc kubenswrapper[4890]: I1125 15:50:08.904639 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ww2d5" Nov 25 15:50:08 crc kubenswrapper[4890]: I1125 15:50:08.933654 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ww2d5"] Nov 25 15:50:08 crc kubenswrapper[4890]: I1125 15:50:08.954317 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/276e9712-d4a0-46f9-b667-4f1adc2213d8-utilities\") pod \"community-operators-ww2d5\" (UID: \"276e9712-d4a0-46f9-b667-4f1adc2213d8\") " pod="openshift-marketplace/community-operators-ww2d5" Nov 25 15:50:08 crc kubenswrapper[4890]: I1125 15:50:08.954444 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhfnh\" (UniqueName: \"kubernetes.io/projected/276e9712-d4a0-46f9-b667-4f1adc2213d8-kube-api-access-jhfnh\") pod \"community-operators-ww2d5\" (UID: \"276e9712-d4a0-46f9-b667-4f1adc2213d8\") " pod="openshift-marketplace/community-operators-ww2d5" Nov 25 15:50:08 crc kubenswrapper[4890]: I1125 15:50:08.954470 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/276e9712-d4a0-46f9-b667-4f1adc2213d8-catalog-content\") pod \"community-operators-ww2d5\" (UID: \"276e9712-d4a0-46f9-b667-4f1adc2213d8\") " pod="openshift-marketplace/community-operators-ww2d5" Nov 25 15:50:09 crc kubenswrapper[4890]: I1125 15:50:09.055300 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/276e9712-d4a0-46f9-b667-4f1adc2213d8-utilities\") pod \"community-operators-ww2d5\" (UID: \"276e9712-d4a0-46f9-b667-4f1adc2213d8\") " pod="openshift-marketplace/community-operators-ww2d5" Nov 25 15:50:09 crc kubenswrapper[4890]: I1125 15:50:09.055380 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhfnh\" (UniqueName: \"kubernetes.io/projected/276e9712-d4a0-46f9-b667-4f1adc2213d8-kube-api-access-jhfnh\") pod \"community-operators-ww2d5\" (UID: \"276e9712-d4a0-46f9-b667-4f1adc2213d8\") " pod="openshift-marketplace/community-operators-ww2d5" Nov 25 15:50:09 crc kubenswrapper[4890]: I1125 15:50:09.055402 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/276e9712-d4a0-46f9-b667-4f1adc2213d8-catalog-content\") pod \"community-operators-ww2d5\" (UID: \"276e9712-d4a0-46f9-b667-4f1adc2213d8\") " pod="openshift-marketplace/community-operators-ww2d5" Nov 25 15:50:09 crc kubenswrapper[4890]: I1125 15:50:09.055826 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/276e9712-d4a0-46f9-b667-4f1adc2213d8-catalog-content\") pod \"community-operators-ww2d5\" (UID: \"276e9712-d4a0-46f9-b667-4f1adc2213d8\") " pod="openshift-marketplace/community-operators-ww2d5" Nov 25 15:50:09 crc kubenswrapper[4890]: I1125 15:50:09.056034 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/276e9712-d4a0-46f9-b667-4f1adc2213d8-utilities\") pod \"community-operators-ww2d5\" (UID: \"276e9712-d4a0-46f9-b667-4f1adc2213d8\") " pod="openshift-marketplace/community-operators-ww2d5" Nov 25 15:50:09 crc kubenswrapper[4890]: I1125 15:50:09.090367 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhfnh\" (UniqueName: \"kubernetes.io/projected/276e9712-d4a0-46f9-b667-4f1adc2213d8-kube-api-access-jhfnh\") pod \"community-operators-ww2d5\" (UID: \"276e9712-d4a0-46f9-b667-4f1adc2213d8\") " pod="openshift-marketplace/community-operators-ww2d5" Nov 25 15:50:09 crc kubenswrapper[4890]: I1125 15:50:09.224833 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ww2d5" Nov 25 15:50:09 crc kubenswrapper[4890]: I1125 15:50:09.709682 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ww2d5"] Nov 25 15:50:10 crc kubenswrapper[4890]: I1125 15:50:10.299877 4890 generic.go:334] "Generic (PLEG): container finished" podID="276e9712-d4a0-46f9-b667-4f1adc2213d8" containerID="d1b24a26467b39c7528ecf258cf97d5f3749453bf635e1196270a886b3d43691" exitCode=0 Nov 25 15:50:10 crc kubenswrapper[4890]: I1125 15:50:10.299972 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ww2d5" event={"ID":"276e9712-d4a0-46f9-b667-4f1adc2213d8","Type":"ContainerDied","Data":"d1b24a26467b39c7528ecf258cf97d5f3749453bf635e1196270a886b3d43691"} Nov 25 15:50:10 crc kubenswrapper[4890]: I1125 15:50:10.300207 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ww2d5" event={"ID":"276e9712-d4a0-46f9-b667-4f1adc2213d8","Type":"ContainerStarted","Data":"e8fdb3d6fcce42165d39c37b144fc0d08fdc5511e326a701d33a3496a9366ef6"} Nov 25 15:50:10 crc kubenswrapper[4890]: I1125 15:50:10.303851 4890 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 15:50:11 crc kubenswrapper[4890]: I1125 15:50:11.315578 4890 generic.go:334] "Generic (PLEG): container finished" podID="276e9712-d4a0-46f9-b667-4f1adc2213d8" containerID="48c7a8f6ec57bc561cdb5e3d38ba40d535e34b64f3268d482c7101a3a468b96d" exitCode=0 Nov 25 15:50:11 crc kubenswrapper[4890]: I1125 15:50:11.315724 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ww2d5" event={"ID":"276e9712-d4a0-46f9-b667-4f1adc2213d8","Type":"ContainerDied","Data":"48c7a8f6ec57bc561cdb5e3d38ba40d535e34b64f3268d482c7101a3a468b96d"} Nov 25 15:50:12 crc kubenswrapper[4890]: I1125 15:50:12.331389 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ww2d5" event={"ID":"276e9712-d4a0-46f9-b667-4f1adc2213d8","Type":"ContainerStarted","Data":"ce7f71ff20f6f50895ac77ae2564e86c1357fea8745492b09033c4ecb9aa55a6"} Nov 25 15:50:12 crc kubenswrapper[4890]: I1125 15:50:12.357959 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ww2d5" podStartSLOduration=2.88024511 podStartE2EDuration="4.357940338s" podCreationTimestamp="2025-11-25 15:50:08 +0000 UTC" firstStartedPulling="2025-11-25 15:50:10.303642667 +0000 UTC m=+2868.746105277" lastFinishedPulling="2025-11-25 15:50:11.781337895 +0000 UTC m=+2870.223800505" observedRunningTime="2025-11-25 15:50:12.357571279 +0000 UTC m=+2870.800033889" watchObservedRunningTime="2025-11-25 15:50:12.357940338 +0000 UTC m=+2870.800402968" Nov 25 15:50:19 crc kubenswrapper[4890]: I1125 15:50:19.226000 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ww2d5" Nov 25 15:50:19 crc kubenswrapper[4890]: I1125 15:50:19.227094 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ww2d5" Nov 25 15:50:19 crc kubenswrapper[4890]: I1125 15:50:19.294941 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ww2d5" Nov 25 15:50:19 crc kubenswrapper[4890]: I1125 15:50:19.488085 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ww2d5" Nov 25 15:50:19 crc kubenswrapper[4890]: I1125 15:50:19.551822 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ww2d5"] Nov 25 15:50:21 crc kubenswrapper[4890]: I1125 15:50:21.434853 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ww2d5" podUID="276e9712-d4a0-46f9-b667-4f1adc2213d8" containerName="registry-server" containerID="cri-o://ce7f71ff20f6f50895ac77ae2564e86c1357fea8745492b09033c4ecb9aa55a6" gracePeriod=2 Nov 25 15:50:21 crc kubenswrapper[4890]: I1125 15:50:21.983833 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ww2d5" Nov 25 15:50:22 crc kubenswrapper[4890]: I1125 15:50:22.165812 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/276e9712-d4a0-46f9-b667-4f1adc2213d8-utilities\") pod \"276e9712-d4a0-46f9-b667-4f1adc2213d8\" (UID: \"276e9712-d4a0-46f9-b667-4f1adc2213d8\") " Nov 25 15:50:22 crc kubenswrapper[4890]: I1125 15:50:22.166368 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/276e9712-d4a0-46f9-b667-4f1adc2213d8-catalog-content\") pod \"276e9712-d4a0-46f9-b667-4f1adc2213d8\" (UID: \"276e9712-d4a0-46f9-b667-4f1adc2213d8\") " Nov 25 15:50:22 crc kubenswrapper[4890]: I1125 15:50:22.166749 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhfnh\" (UniqueName: \"kubernetes.io/projected/276e9712-d4a0-46f9-b667-4f1adc2213d8-kube-api-access-jhfnh\") pod \"276e9712-d4a0-46f9-b667-4f1adc2213d8\" (UID: \"276e9712-d4a0-46f9-b667-4f1adc2213d8\") " Nov 25 15:50:22 crc kubenswrapper[4890]: I1125 15:50:22.167742 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/276e9712-d4a0-46f9-b667-4f1adc2213d8-utilities" (OuterVolumeSpecName: "utilities") pod "276e9712-d4a0-46f9-b667-4f1adc2213d8" (UID: "276e9712-d4a0-46f9-b667-4f1adc2213d8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:50:22 crc kubenswrapper[4890]: I1125 15:50:22.174542 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/276e9712-d4a0-46f9-b667-4f1adc2213d8-kube-api-access-jhfnh" (OuterVolumeSpecName: "kube-api-access-jhfnh") pod "276e9712-d4a0-46f9-b667-4f1adc2213d8" (UID: "276e9712-d4a0-46f9-b667-4f1adc2213d8"). InnerVolumeSpecName "kube-api-access-jhfnh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:50:22 crc kubenswrapper[4890]: I1125 15:50:22.270224 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhfnh\" (UniqueName: \"kubernetes.io/projected/276e9712-d4a0-46f9-b667-4f1adc2213d8-kube-api-access-jhfnh\") on node \"crc\" DevicePath \"\"" Nov 25 15:50:22 crc kubenswrapper[4890]: I1125 15:50:22.270780 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/276e9712-d4a0-46f9-b667-4f1adc2213d8-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 15:50:22 crc kubenswrapper[4890]: I1125 15:50:22.446022 4890 generic.go:334] "Generic (PLEG): container finished" podID="276e9712-d4a0-46f9-b667-4f1adc2213d8" containerID="ce7f71ff20f6f50895ac77ae2564e86c1357fea8745492b09033c4ecb9aa55a6" exitCode=0 Nov 25 15:50:22 crc kubenswrapper[4890]: I1125 15:50:22.446076 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ww2d5" event={"ID":"276e9712-d4a0-46f9-b667-4f1adc2213d8","Type":"ContainerDied","Data":"ce7f71ff20f6f50895ac77ae2564e86c1357fea8745492b09033c4ecb9aa55a6"} Nov 25 15:50:22 crc kubenswrapper[4890]: I1125 15:50:22.446117 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ww2d5" event={"ID":"276e9712-d4a0-46f9-b667-4f1adc2213d8","Type":"ContainerDied","Data":"e8fdb3d6fcce42165d39c37b144fc0d08fdc5511e326a701d33a3496a9366ef6"} Nov 25 15:50:22 crc kubenswrapper[4890]: I1125 15:50:22.446130 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ww2d5" Nov 25 15:50:22 crc kubenswrapper[4890]: I1125 15:50:22.446143 4890 scope.go:117] "RemoveContainer" containerID="ce7f71ff20f6f50895ac77ae2564e86c1357fea8745492b09033c4ecb9aa55a6" Nov 25 15:50:22 crc kubenswrapper[4890]: I1125 15:50:22.487843 4890 scope.go:117] "RemoveContainer" containerID="48c7a8f6ec57bc561cdb5e3d38ba40d535e34b64f3268d482c7101a3a468b96d" Nov 25 15:50:22 crc kubenswrapper[4890]: I1125 15:50:22.520850 4890 scope.go:117] "RemoveContainer" containerID="d1b24a26467b39c7528ecf258cf97d5f3749453bf635e1196270a886b3d43691" Nov 25 15:50:22 crc kubenswrapper[4890]: I1125 15:50:22.572711 4890 scope.go:117] "RemoveContainer" containerID="ce7f71ff20f6f50895ac77ae2564e86c1357fea8745492b09033c4ecb9aa55a6" Nov 25 15:50:22 crc kubenswrapper[4890]: E1125 15:50:22.573580 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce7f71ff20f6f50895ac77ae2564e86c1357fea8745492b09033c4ecb9aa55a6\": container with ID starting with ce7f71ff20f6f50895ac77ae2564e86c1357fea8745492b09033c4ecb9aa55a6 not found: ID does not exist" containerID="ce7f71ff20f6f50895ac77ae2564e86c1357fea8745492b09033c4ecb9aa55a6" Nov 25 15:50:22 crc kubenswrapper[4890]: I1125 15:50:22.573636 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce7f71ff20f6f50895ac77ae2564e86c1357fea8745492b09033c4ecb9aa55a6"} err="failed to get container status \"ce7f71ff20f6f50895ac77ae2564e86c1357fea8745492b09033c4ecb9aa55a6\": rpc error: code = NotFound desc = could not find container \"ce7f71ff20f6f50895ac77ae2564e86c1357fea8745492b09033c4ecb9aa55a6\": container with ID starting with ce7f71ff20f6f50895ac77ae2564e86c1357fea8745492b09033c4ecb9aa55a6 not found: ID does not exist" Nov 25 15:50:22 crc kubenswrapper[4890]: I1125 15:50:22.573670 4890 scope.go:117] "RemoveContainer" containerID="48c7a8f6ec57bc561cdb5e3d38ba40d535e34b64f3268d482c7101a3a468b96d" Nov 25 15:50:22 crc kubenswrapper[4890]: E1125 15:50:22.574238 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48c7a8f6ec57bc561cdb5e3d38ba40d535e34b64f3268d482c7101a3a468b96d\": container with ID starting with 48c7a8f6ec57bc561cdb5e3d38ba40d535e34b64f3268d482c7101a3a468b96d not found: ID does not exist" containerID="48c7a8f6ec57bc561cdb5e3d38ba40d535e34b64f3268d482c7101a3a468b96d" Nov 25 15:50:22 crc kubenswrapper[4890]: I1125 15:50:22.574288 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48c7a8f6ec57bc561cdb5e3d38ba40d535e34b64f3268d482c7101a3a468b96d"} err="failed to get container status \"48c7a8f6ec57bc561cdb5e3d38ba40d535e34b64f3268d482c7101a3a468b96d\": rpc error: code = NotFound desc = could not find container \"48c7a8f6ec57bc561cdb5e3d38ba40d535e34b64f3268d482c7101a3a468b96d\": container with ID starting with 48c7a8f6ec57bc561cdb5e3d38ba40d535e34b64f3268d482c7101a3a468b96d not found: ID does not exist" Nov 25 15:50:22 crc kubenswrapper[4890]: I1125 15:50:22.574316 4890 scope.go:117] "RemoveContainer" containerID="d1b24a26467b39c7528ecf258cf97d5f3749453bf635e1196270a886b3d43691" Nov 25 15:50:22 crc kubenswrapper[4890]: E1125 15:50:22.574674 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1b24a26467b39c7528ecf258cf97d5f3749453bf635e1196270a886b3d43691\": container with ID starting with d1b24a26467b39c7528ecf258cf97d5f3749453bf635e1196270a886b3d43691 not found: ID does not exist" containerID="d1b24a26467b39c7528ecf258cf97d5f3749453bf635e1196270a886b3d43691" Nov 25 15:50:22 crc kubenswrapper[4890]: I1125 15:50:22.574703 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1b24a26467b39c7528ecf258cf97d5f3749453bf635e1196270a886b3d43691"} err="failed to get container status \"d1b24a26467b39c7528ecf258cf97d5f3749453bf635e1196270a886b3d43691\": rpc error: code = NotFound desc = could not find container \"d1b24a26467b39c7528ecf258cf97d5f3749453bf635e1196270a886b3d43691\": container with ID starting with d1b24a26467b39c7528ecf258cf97d5f3749453bf635e1196270a886b3d43691 not found: ID does not exist" Nov 25 15:50:22 crc kubenswrapper[4890]: I1125 15:50:22.646481 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/276e9712-d4a0-46f9-b667-4f1adc2213d8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "276e9712-d4a0-46f9-b667-4f1adc2213d8" (UID: "276e9712-d4a0-46f9-b667-4f1adc2213d8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:50:22 crc kubenswrapper[4890]: I1125 15:50:22.679684 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/276e9712-d4a0-46f9-b667-4f1adc2213d8-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 15:50:22 crc kubenswrapper[4890]: I1125 15:50:22.798096 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ww2d5"] Nov 25 15:50:22 crc kubenswrapper[4890]: I1125 15:50:22.810505 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ww2d5"] Nov 25 15:50:24 crc kubenswrapper[4890]: I1125 15:50:24.191659 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="276e9712-d4a0-46f9-b667-4f1adc2213d8" path="/var/lib/kubelet/pods/276e9712-d4a0-46f9-b667-4f1adc2213d8/volumes" Nov 25 15:50:26 crc kubenswrapper[4890]: I1125 15:50:26.448315 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 15:50:26 crc kubenswrapper[4890]: I1125 15:50:26.449360 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 15:50:41 crc kubenswrapper[4890]: I1125 15:50:41.895025 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Nov 25 15:50:41 crc kubenswrapper[4890]: E1125 15:50:41.896362 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="276e9712-d4a0-46f9-b667-4f1adc2213d8" containerName="extract-utilities" Nov 25 15:50:41 crc kubenswrapper[4890]: I1125 15:50:41.896378 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="276e9712-d4a0-46f9-b667-4f1adc2213d8" containerName="extract-utilities" Nov 25 15:50:41 crc kubenswrapper[4890]: E1125 15:50:41.896391 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="276e9712-d4a0-46f9-b667-4f1adc2213d8" containerName="extract-content" Nov 25 15:50:41 crc kubenswrapper[4890]: I1125 15:50:41.896403 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="276e9712-d4a0-46f9-b667-4f1adc2213d8" containerName="extract-content" Nov 25 15:50:41 crc kubenswrapper[4890]: E1125 15:50:41.896420 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="276e9712-d4a0-46f9-b667-4f1adc2213d8" containerName="registry-server" Nov 25 15:50:41 crc kubenswrapper[4890]: I1125 15:50:41.896426 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="276e9712-d4a0-46f9-b667-4f1adc2213d8" containerName="registry-server" Nov 25 15:50:41 crc kubenswrapper[4890]: I1125 15:50:41.896612 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="276e9712-d4a0-46f9-b667-4f1adc2213d8" containerName="registry-server" Nov 25 15:50:41 crc kubenswrapper[4890]: I1125 15:50:41.897268 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 25 15:50:41 crc kubenswrapper[4890]: I1125 15:50:41.900983 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Nov 25 15:50:41 crc kubenswrapper[4890]: I1125 15:50:41.902294 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-wpdrd" Nov 25 15:50:41 crc kubenswrapper[4890]: I1125 15:50:41.902538 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Nov 25 15:50:41 crc kubenswrapper[4890]: I1125 15:50:41.903214 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Nov 25 15:50:41 crc kubenswrapper[4890]: I1125 15:50:41.911661 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Nov 25 15:50:42 crc kubenswrapper[4890]: I1125 15:50:42.025754 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8b652448-2962-42f9-b0b5-e551b50edc1e-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"8b652448-2962-42f9-b0b5-e551b50edc1e\") " pod="openstack/tempest-tests-tempest" Nov 25 15:50:42 crc kubenswrapper[4890]: I1125 15:50:42.025909 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/8b652448-2962-42f9-b0b5-e551b50edc1e-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"8b652448-2962-42f9-b0b5-e551b50edc1e\") " pod="openstack/tempest-tests-tempest" Nov 25 15:50:42 crc kubenswrapper[4890]: I1125 15:50:42.025965 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgmvt\" (UniqueName: \"kubernetes.io/projected/8b652448-2962-42f9-b0b5-e551b50edc1e-kube-api-access-fgmvt\") pod \"tempest-tests-tempest\" (UID: \"8b652448-2962-42f9-b0b5-e551b50edc1e\") " pod="openstack/tempest-tests-tempest" Nov 25 15:50:42 crc kubenswrapper[4890]: I1125 15:50:42.026025 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"tempest-tests-tempest\" (UID: \"8b652448-2962-42f9-b0b5-e551b50edc1e\") " pod="openstack/tempest-tests-tempest" Nov 25 15:50:42 crc kubenswrapper[4890]: I1125 15:50:42.026113 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/8b652448-2962-42f9-b0b5-e551b50edc1e-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"8b652448-2962-42f9-b0b5-e551b50edc1e\") " pod="openstack/tempest-tests-tempest" Nov 25 15:50:42 crc kubenswrapper[4890]: I1125 15:50:42.026151 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8b652448-2962-42f9-b0b5-e551b50edc1e-config-data\") pod \"tempest-tests-tempest\" (UID: \"8b652448-2962-42f9-b0b5-e551b50edc1e\") " pod="openstack/tempest-tests-tempest" Nov 25 15:50:42 crc kubenswrapper[4890]: I1125 15:50:42.026200 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8b652448-2962-42f9-b0b5-e551b50edc1e-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"8b652448-2962-42f9-b0b5-e551b50edc1e\") " pod="openstack/tempest-tests-tempest" Nov 25 15:50:42 crc kubenswrapper[4890]: I1125 15:50:42.026224 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8b652448-2962-42f9-b0b5-e551b50edc1e-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"8b652448-2962-42f9-b0b5-e551b50edc1e\") " pod="openstack/tempest-tests-tempest" Nov 25 15:50:42 crc kubenswrapper[4890]: I1125 15:50:42.026422 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/8b652448-2962-42f9-b0b5-e551b50edc1e-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"8b652448-2962-42f9-b0b5-e551b50edc1e\") " pod="openstack/tempest-tests-tempest" Nov 25 15:50:42 crc kubenswrapper[4890]: I1125 15:50:42.129029 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8b652448-2962-42f9-b0b5-e551b50edc1e-config-data\") pod \"tempest-tests-tempest\" (UID: \"8b652448-2962-42f9-b0b5-e551b50edc1e\") " pod="openstack/tempest-tests-tempest" Nov 25 15:50:42 crc kubenswrapper[4890]: I1125 15:50:42.129121 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8b652448-2962-42f9-b0b5-e551b50edc1e-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"8b652448-2962-42f9-b0b5-e551b50edc1e\") " pod="openstack/tempest-tests-tempest" Nov 25 15:50:42 crc kubenswrapper[4890]: I1125 15:50:42.129154 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8b652448-2962-42f9-b0b5-e551b50edc1e-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"8b652448-2962-42f9-b0b5-e551b50edc1e\") " pod="openstack/tempest-tests-tempest" Nov 25 15:50:42 crc kubenswrapper[4890]: I1125 15:50:42.129229 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/8b652448-2962-42f9-b0b5-e551b50edc1e-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"8b652448-2962-42f9-b0b5-e551b50edc1e\") " pod="openstack/tempest-tests-tempest" Nov 25 15:50:42 crc kubenswrapper[4890]: I1125 15:50:42.129301 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8b652448-2962-42f9-b0b5-e551b50edc1e-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"8b652448-2962-42f9-b0b5-e551b50edc1e\") " pod="openstack/tempest-tests-tempest" Nov 25 15:50:42 crc kubenswrapper[4890]: I1125 15:50:42.129400 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/8b652448-2962-42f9-b0b5-e551b50edc1e-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"8b652448-2962-42f9-b0b5-e551b50edc1e\") " pod="openstack/tempest-tests-tempest" Nov 25 15:50:42 crc kubenswrapper[4890]: I1125 15:50:42.129442 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgmvt\" (UniqueName: \"kubernetes.io/projected/8b652448-2962-42f9-b0b5-e551b50edc1e-kube-api-access-fgmvt\") pod \"tempest-tests-tempest\" (UID: \"8b652448-2962-42f9-b0b5-e551b50edc1e\") " pod="openstack/tempest-tests-tempest" Nov 25 15:50:42 crc kubenswrapper[4890]: I1125 15:50:42.129472 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"tempest-tests-tempest\" (UID: \"8b652448-2962-42f9-b0b5-e551b50edc1e\") " pod="openstack/tempest-tests-tempest" Nov 25 15:50:42 crc kubenswrapper[4890]: I1125 15:50:42.129549 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/8b652448-2962-42f9-b0b5-e551b50edc1e-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"8b652448-2962-42f9-b0b5-e551b50edc1e\") " pod="openstack/tempest-tests-tempest" Nov 25 15:50:42 crc kubenswrapper[4890]: I1125 15:50:42.130278 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/8b652448-2962-42f9-b0b5-e551b50edc1e-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"8b652448-2962-42f9-b0b5-e551b50edc1e\") " pod="openstack/tempest-tests-tempest" Nov 25 15:50:42 crc kubenswrapper[4890]: I1125 15:50:42.130394 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8b652448-2962-42f9-b0b5-e551b50edc1e-config-data\") pod \"tempest-tests-tempest\" (UID: \"8b652448-2962-42f9-b0b5-e551b50edc1e\") " pod="openstack/tempest-tests-tempest" Nov 25 15:50:42 crc kubenswrapper[4890]: I1125 15:50:42.130987 4890 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"tempest-tests-tempest\" (UID: \"8b652448-2962-42f9-b0b5-e551b50edc1e\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/tempest-tests-tempest" Nov 25 15:50:42 crc kubenswrapper[4890]: I1125 15:50:42.131231 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/8b652448-2962-42f9-b0b5-e551b50edc1e-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"8b652448-2962-42f9-b0b5-e551b50edc1e\") " pod="openstack/tempest-tests-tempest" Nov 25 15:50:42 crc kubenswrapper[4890]: I1125 15:50:42.131893 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8b652448-2962-42f9-b0b5-e551b50edc1e-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"8b652448-2962-42f9-b0b5-e551b50edc1e\") " pod="openstack/tempest-tests-tempest" Nov 25 15:50:42 crc kubenswrapper[4890]: I1125 15:50:42.138776 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/8b652448-2962-42f9-b0b5-e551b50edc1e-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"8b652448-2962-42f9-b0b5-e551b50edc1e\") " pod="openstack/tempest-tests-tempest" Nov 25 15:50:42 crc kubenswrapper[4890]: I1125 15:50:42.138938 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8b652448-2962-42f9-b0b5-e551b50edc1e-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"8b652448-2962-42f9-b0b5-e551b50edc1e\") " pod="openstack/tempest-tests-tempest" Nov 25 15:50:42 crc kubenswrapper[4890]: I1125 15:50:42.143054 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8b652448-2962-42f9-b0b5-e551b50edc1e-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"8b652448-2962-42f9-b0b5-e551b50edc1e\") " pod="openstack/tempest-tests-tempest" Nov 25 15:50:42 crc kubenswrapper[4890]: I1125 15:50:42.150405 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgmvt\" (UniqueName: \"kubernetes.io/projected/8b652448-2962-42f9-b0b5-e551b50edc1e-kube-api-access-fgmvt\") pod \"tempest-tests-tempest\" (UID: \"8b652448-2962-42f9-b0b5-e551b50edc1e\") " pod="openstack/tempest-tests-tempest" Nov 25 15:50:42 crc kubenswrapper[4890]: I1125 15:50:42.162727 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"tempest-tests-tempest\" (UID: \"8b652448-2962-42f9-b0b5-e551b50edc1e\") " pod="openstack/tempest-tests-tempest" Nov 25 15:50:42 crc kubenswrapper[4890]: I1125 15:50:42.225819 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 25 15:50:42 crc kubenswrapper[4890]: I1125 15:50:42.735669 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Nov 25 15:50:43 crc kubenswrapper[4890]: I1125 15:50:43.684595 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"8b652448-2962-42f9-b0b5-e551b50edc1e","Type":"ContainerStarted","Data":"10fafdacefa12996d94beffa374682d9bccca4cd4e5bc1dca33d1f626c9893e5"} Nov 25 15:50:56 crc kubenswrapper[4890]: I1125 15:50:56.447024 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 15:50:56 crc kubenswrapper[4890]: I1125 15:50:56.447670 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 15:50:56 crc kubenswrapper[4890]: I1125 15:50:56.447746 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" Nov 25 15:50:56 crc kubenswrapper[4890]: I1125 15:50:56.448557 4890 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"aa8a6c3d508388bee84226737d06d0443dd3e6d511ad22e6e9e1c1f48362f983"} pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 15:50:56 crc kubenswrapper[4890]: I1125 15:50:56.448641 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" containerID="cri-o://aa8a6c3d508388bee84226737d06d0443dd3e6d511ad22e6e9e1c1f48362f983" gracePeriod=600 Nov 25 15:50:56 crc kubenswrapper[4890]: I1125 15:50:56.881054 4890 generic.go:334] "Generic (PLEG): container finished" podID="4e4f849d-f239-4727-a73e-18327856929a" containerID="aa8a6c3d508388bee84226737d06d0443dd3e6d511ad22e6e9e1c1f48362f983" exitCode=0 Nov 25 15:50:56 crc kubenswrapper[4890]: I1125 15:50:56.881111 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" event={"ID":"4e4f849d-f239-4727-a73e-18327856929a","Type":"ContainerDied","Data":"aa8a6c3d508388bee84226737d06d0443dd3e6d511ad22e6e9e1c1f48362f983"} Nov 25 15:50:56 crc kubenswrapper[4890]: I1125 15:50:56.881213 4890 scope.go:117] "RemoveContainer" containerID="3add3ce883ec44a5429bf3e4879e91159a451c26e6f8d63ebfd5527dc360e6b8" Nov 25 15:51:02 crc kubenswrapper[4890]: I1125 15:51:02.193551 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-f9bwc"] Nov 25 15:51:02 crc kubenswrapper[4890]: I1125 15:51:02.203453 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f9bwc"] Nov 25 15:51:02 crc kubenswrapper[4890]: I1125 15:51:02.203611 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f9bwc" Nov 25 15:51:02 crc kubenswrapper[4890]: I1125 15:51:02.402173 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cf2f167-8a28-4bcf-891a-1f7cf91d433b-utilities\") pod \"redhat-operators-f9bwc\" (UID: \"5cf2f167-8a28-4bcf-891a-1f7cf91d433b\") " pod="openshift-marketplace/redhat-operators-f9bwc" Nov 25 15:51:02 crc kubenswrapper[4890]: I1125 15:51:02.402344 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wpbg\" (UniqueName: \"kubernetes.io/projected/5cf2f167-8a28-4bcf-891a-1f7cf91d433b-kube-api-access-8wpbg\") pod \"redhat-operators-f9bwc\" (UID: \"5cf2f167-8a28-4bcf-891a-1f7cf91d433b\") " pod="openshift-marketplace/redhat-operators-f9bwc" Nov 25 15:51:02 crc kubenswrapper[4890]: I1125 15:51:02.402403 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cf2f167-8a28-4bcf-891a-1f7cf91d433b-catalog-content\") pod \"redhat-operators-f9bwc\" (UID: \"5cf2f167-8a28-4bcf-891a-1f7cf91d433b\") " pod="openshift-marketplace/redhat-operators-f9bwc" Nov 25 15:51:02 crc kubenswrapper[4890]: I1125 15:51:02.503823 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cf2f167-8a28-4bcf-891a-1f7cf91d433b-utilities\") pod \"redhat-operators-f9bwc\" (UID: \"5cf2f167-8a28-4bcf-891a-1f7cf91d433b\") " pod="openshift-marketplace/redhat-operators-f9bwc" Nov 25 15:51:02 crc kubenswrapper[4890]: I1125 15:51:02.504248 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wpbg\" (UniqueName: \"kubernetes.io/projected/5cf2f167-8a28-4bcf-891a-1f7cf91d433b-kube-api-access-8wpbg\") pod \"redhat-operators-f9bwc\" (UID: \"5cf2f167-8a28-4bcf-891a-1f7cf91d433b\") " pod="openshift-marketplace/redhat-operators-f9bwc" Nov 25 15:51:02 crc kubenswrapper[4890]: I1125 15:51:02.504292 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cf2f167-8a28-4bcf-891a-1f7cf91d433b-catalog-content\") pod \"redhat-operators-f9bwc\" (UID: \"5cf2f167-8a28-4bcf-891a-1f7cf91d433b\") " pod="openshift-marketplace/redhat-operators-f9bwc" Nov 25 15:51:02 crc kubenswrapper[4890]: I1125 15:51:02.504668 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cf2f167-8a28-4bcf-891a-1f7cf91d433b-utilities\") pod \"redhat-operators-f9bwc\" (UID: \"5cf2f167-8a28-4bcf-891a-1f7cf91d433b\") " pod="openshift-marketplace/redhat-operators-f9bwc" Nov 25 15:51:02 crc kubenswrapper[4890]: I1125 15:51:02.504795 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cf2f167-8a28-4bcf-891a-1f7cf91d433b-catalog-content\") pod \"redhat-operators-f9bwc\" (UID: \"5cf2f167-8a28-4bcf-891a-1f7cf91d433b\") " pod="openshift-marketplace/redhat-operators-f9bwc" Nov 25 15:51:02 crc kubenswrapper[4890]: I1125 15:51:02.547811 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wpbg\" (UniqueName: \"kubernetes.io/projected/5cf2f167-8a28-4bcf-891a-1f7cf91d433b-kube-api-access-8wpbg\") pod \"redhat-operators-f9bwc\" (UID: \"5cf2f167-8a28-4bcf-891a-1f7cf91d433b\") " pod="openshift-marketplace/redhat-operators-f9bwc" Nov 25 15:51:02 crc kubenswrapper[4890]: I1125 15:51:02.841607 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f9bwc" Nov 25 15:51:06 crc kubenswrapper[4890]: E1125 15:51:06.809937 4890 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/system.slice/rpm-ostreed.service\": RecentStats: unable to find data in memory cache]" Nov 25 15:51:13 crc kubenswrapper[4890]: E1125 15:51:13.894076 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:51:14 crc kubenswrapper[4890]: E1125 15:51:14.067455 4890 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Nov 25 15:51:14 crc kubenswrapper[4890]: E1125 15:51:14.067973 4890 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fgmvt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(8b652448-2962-42f9-b0b5-e551b50edc1e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 15:51:14 crc kubenswrapper[4890]: E1125 15:51:14.069314 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="8b652448-2962-42f9-b0b5-e551b50edc1e" Nov 25 15:51:14 crc kubenswrapper[4890]: I1125 15:51:14.170278 4890 scope.go:117] "RemoveContainer" containerID="aa8a6c3d508388bee84226737d06d0443dd3e6d511ad22e6e9e1c1f48362f983" Nov 25 15:51:14 crc kubenswrapper[4890]: E1125 15:51:14.170680 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:51:14 crc kubenswrapper[4890]: E1125 15:51:14.171539 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="8b652448-2962-42f9-b0b5-e551b50edc1e" Nov 25 15:51:14 crc kubenswrapper[4890]: I1125 15:51:14.378046 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f9bwc"] Nov 25 15:51:15 crc kubenswrapper[4890]: I1125 15:51:15.185201 4890 generic.go:334] "Generic (PLEG): container finished" podID="5cf2f167-8a28-4bcf-891a-1f7cf91d433b" containerID="0185c639fb8a4401a43960c6c8c9962227adaeb83baf1ab7afd9a5099f90d8be" exitCode=0 Nov 25 15:51:15 crc kubenswrapper[4890]: I1125 15:51:15.185282 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f9bwc" event={"ID":"5cf2f167-8a28-4bcf-891a-1f7cf91d433b","Type":"ContainerDied","Data":"0185c639fb8a4401a43960c6c8c9962227adaeb83baf1ab7afd9a5099f90d8be"} Nov 25 15:51:15 crc kubenswrapper[4890]: I1125 15:51:15.185711 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f9bwc" event={"ID":"5cf2f167-8a28-4bcf-891a-1f7cf91d433b","Type":"ContainerStarted","Data":"f53f4cbda163c61c89739d2ffc1ac8f1848690bd57942940579720fa4e486033"} Nov 25 15:51:17 crc kubenswrapper[4890]: I1125 15:51:17.225091 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f9bwc" event={"ID":"5cf2f167-8a28-4bcf-891a-1f7cf91d433b","Type":"ContainerStarted","Data":"d2b45b3c99fde6d7fd14ba10971416462c939286af9b27ebca974fafe1b2db74"} Nov 25 15:51:18 crc kubenswrapper[4890]: I1125 15:51:18.250791 4890 generic.go:334] "Generic (PLEG): container finished" podID="5cf2f167-8a28-4bcf-891a-1f7cf91d433b" containerID="d2b45b3c99fde6d7fd14ba10971416462c939286af9b27ebca974fafe1b2db74" exitCode=0 Nov 25 15:51:18 crc kubenswrapper[4890]: I1125 15:51:18.250890 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f9bwc" event={"ID":"5cf2f167-8a28-4bcf-891a-1f7cf91d433b","Type":"ContainerDied","Data":"d2b45b3c99fde6d7fd14ba10971416462c939286af9b27ebca974fafe1b2db74"} Nov 25 15:51:21 crc kubenswrapper[4890]: I1125 15:51:21.300559 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f9bwc" event={"ID":"5cf2f167-8a28-4bcf-891a-1f7cf91d433b","Type":"ContainerStarted","Data":"23a9b8fa2451bcbb91ef90a4f685c7fdabf86312bf265ec14af4e6c34828d3a1"} Nov 25 15:51:21 crc kubenswrapper[4890]: I1125 15:51:21.331729 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-f9bwc" podStartSLOduration=14.282430043 podStartE2EDuration="19.331702408s" podCreationTimestamp="2025-11-25 15:51:02 +0000 UTC" firstStartedPulling="2025-11-25 15:51:15.191144563 +0000 UTC m=+2933.633607173" lastFinishedPulling="2025-11-25 15:51:20.240416918 +0000 UTC m=+2938.682879538" observedRunningTime="2025-11-25 15:51:21.326119859 +0000 UTC m=+2939.768582509" watchObservedRunningTime="2025-11-25 15:51:21.331702408 +0000 UTC m=+2939.774165018" Nov 25 15:51:22 crc kubenswrapper[4890]: I1125 15:51:22.843671 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-f9bwc" Nov 25 15:51:22 crc kubenswrapper[4890]: I1125 15:51:22.844100 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-f9bwc" Nov 25 15:51:23 crc kubenswrapper[4890]: I1125 15:51:23.896783 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-f9bwc" podUID="5cf2f167-8a28-4bcf-891a-1f7cf91d433b" containerName="registry-server" probeResult="failure" output=< Nov 25 15:51:23 crc kubenswrapper[4890]: timeout: failed to connect service ":50051" within 1s Nov 25 15:51:23 crc kubenswrapper[4890]: > Nov 25 15:51:27 crc kubenswrapper[4890]: I1125 15:51:27.174141 4890 scope.go:117] "RemoveContainer" containerID="aa8a6c3d508388bee84226737d06d0443dd3e6d511ad22e6e9e1c1f48362f983" Nov 25 15:51:27 crc kubenswrapper[4890]: E1125 15:51:27.175731 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:51:29 crc kubenswrapper[4890]: I1125 15:51:29.826639 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Nov 25 15:51:31 crc kubenswrapper[4890]: I1125 15:51:31.435236 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"8b652448-2962-42f9-b0b5-e551b50edc1e","Type":"ContainerStarted","Data":"ba77aee782ce8f9ea592a0a8f000c7a382df11999dec90fd8e463c93097cbf2b"} Nov 25 15:51:31 crc kubenswrapper[4890]: I1125 15:51:31.460377 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.375129042 podStartE2EDuration="51.46035779s" podCreationTimestamp="2025-11-25 15:50:40 +0000 UTC" firstStartedPulling="2025-11-25 15:50:42.737292729 +0000 UTC m=+2901.179755369" lastFinishedPulling="2025-11-25 15:51:29.822521507 +0000 UTC m=+2948.264984117" observedRunningTime="2025-11-25 15:51:31.456470173 +0000 UTC m=+2949.898932783" watchObservedRunningTime="2025-11-25 15:51:31.46035779 +0000 UTC m=+2949.902820400" Nov 25 15:51:32 crc kubenswrapper[4890]: I1125 15:51:32.912832 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-f9bwc" Nov 25 15:51:32 crc kubenswrapper[4890]: I1125 15:51:32.971504 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-f9bwc" Nov 25 15:51:33 crc kubenswrapper[4890]: I1125 15:51:33.372781 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f9bwc"] Nov 25 15:51:34 crc kubenswrapper[4890]: I1125 15:51:34.466245 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-f9bwc" podUID="5cf2f167-8a28-4bcf-891a-1f7cf91d433b" containerName="registry-server" containerID="cri-o://23a9b8fa2451bcbb91ef90a4f685c7fdabf86312bf265ec14af4e6c34828d3a1" gracePeriod=2 Nov 25 15:51:35 crc kubenswrapper[4890]: I1125 15:51:35.044586 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f9bwc" Nov 25 15:51:35 crc kubenswrapper[4890]: I1125 15:51:35.116429 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8wpbg\" (UniqueName: \"kubernetes.io/projected/5cf2f167-8a28-4bcf-891a-1f7cf91d433b-kube-api-access-8wpbg\") pod \"5cf2f167-8a28-4bcf-891a-1f7cf91d433b\" (UID: \"5cf2f167-8a28-4bcf-891a-1f7cf91d433b\") " Nov 25 15:51:35 crc kubenswrapper[4890]: I1125 15:51:35.116983 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cf2f167-8a28-4bcf-891a-1f7cf91d433b-utilities\") pod \"5cf2f167-8a28-4bcf-891a-1f7cf91d433b\" (UID: \"5cf2f167-8a28-4bcf-891a-1f7cf91d433b\") " Nov 25 15:51:35 crc kubenswrapper[4890]: I1125 15:51:35.117010 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cf2f167-8a28-4bcf-891a-1f7cf91d433b-catalog-content\") pod \"5cf2f167-8a28-4bcf-891a-1f7cf91d433b\" (UID: \"5cf2f167-8a28-4bcf-891a-1f7cf91d433b\") " Nov 25 15:51:35 crc kubenswrapper[4890]: I1125 15:51:35.117817 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cf2f167-8a28-4bcf-891a-1f7cf91d433b-utilities" (OuterVolumeSpecName: "utilities") pod "5cf2f167-8a28-4bcf-891a-1f7cf91d433b" (UID: "5cf2f167-8a28-4bcf-891a-1f7cf91d433b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:51:35 crc kubenswrapper[4890]: I1125 15:51:35.124427 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cf2f167-8a28-4bcf-891a-1f7cf91d433b-kube-api-access-8wpbg" (OuterVolumeSpecName: "kube-api-access-8wpbg") pod "5cf2f167-8a28-4bcf-891a-1f7cf91d433b" (UID: "5cf2f167-8a28-4bcf-891a-1f7cf91d433b"). InnerVolumeSpecName "kube-api-access-8wpbg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:51:35 crc kubenswrapper[4890]: I1125 15:51:35.220879 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8wpbg\" (UniqueName: \"kubernetes.io/projected/5cf2f167-8a28-4bcf-891a-1f7cf91d433b-kube-api-access-8wpbg\") on node \"crc\" DevicePath \"\"" Nov 25 15:51:35 crc kubenswrapper[4890]: I1125 15:51:35.220957 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cf2f167-8a28-4bcf-891a-1f7cf91d433b-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 15:51:35 crc kubenswrapper[4890]: I1125 15:51:35.229082 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cf2f167-8a28-4bcf-891a-1f7cf91d433b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5cf2f167-8a28-4bcf-891a-1f7cf91d433b" (UID: "5cf2f167-8a28-4bcf-891a-1f7cf91d433b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:51:35 crc kubenswrapper[4890]: I1125 15:51:35.323122 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cf2f167-8a28-4bcf-891a-1f7cf91d433b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 15:51:35 crc kubenswrapper[4890]: I1125 15:51:35.483232 4890 generic.go:334] "Generic (PLEG): container finished" podID="5cf2f167-8a28-4bcf-891a-1f7cf91d433b" containerID="23a9b8fa2451bcbb91ef90a4f685c7fdabf86312bf265ec14af4e6c34828d3a1" exitCode=0 Nov 25 15:51:35 crc kubenswrapper[4890]: I1125 15:51:35.483301 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f9bwc" Nov 25 15:51:35 crc kubenswrapper[4890]: I1125 15:51:35.483299 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f9bwc" event={"ID":"5cf2f167-8a28-4bcf-891a-1f7cf91d433b","Type":"ContainerDied","Data":"23a9b8fa2451bcbb91ef90a4f685c7fdabf86312bf265ec14af4e6c34828d3a1"} Nov 25 15:51:35 crc kubenswrapper[4890]: I1125 15:51:35.483379 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f9bwc" event={"ID":"5cf2f167-8a28-4bcf-891a-1f7cf91d433b","Type":"ContainerDied","Data":"f53f4cbda163c61c89739d2ffc1ac8f1848690bd57942940579720fa4e486033"} Nov 25 15:51:35 crc kubenswrapper[4890]: I1125 15:51:35.483405 4890 scope.go:117] "RemoveContainer" containerID="23a9b8fa2451bcbb91ef90a4f685c7fdabf86312bf265ec14af4e6c34828d3a1" Nov 25 15:51:35 crc kubenswrapper[4890]: I1125 15:51:35.517218 4890 scope.go:117] "RemoveContainer" containerID="d2b45b3c99fde6d7fd14ba10971416462c939286af9b27ebca974fafe1b2db74" Nov 25 15:51:35 crc kubenswrapper[4890]: I1125 15:51:35.546668 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f9bwc"] Nov 25 15:51:35 crc kubenswrapper[4890]: I1125 15:51:35.558900 4890 scope.go:117] "RemoveContainer" containerID="0185c639fb8a4401a43960c6c8c9962227adaeb83baf1ab7afd9a5099f90d8be" Nov 25 15:51:35 crc kubenswrapper[4890]: I1125 15:51:35.567600 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-f9bwc"] Nov 25 15:51:35 crc kubenswrapper[4890]: I1125 15:51:35.602259 4890 scope.go:117] "RemoveContainer" containerID="23a9b8fa2451bcbb91ef90a4f685c7fdabf86312bf265ec14af4e6c34828d3a1" Nov 25 15:51:35 crc kubenswrapper[4890]: E1125 15:51:35.603688 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23a9b8fa2451bcbb91ef90a4f685c7fdabf86312bf265ec14af4e6c34828d3a1\": container with ID starting with 23a9b8fa2451bcbb91ef90a4f685c7fdabf86312bf265ec14af4e6c34828d3a1 not found: ID does not exist" containerID="23a9b8fa2451bcbb91ef90a4f685c7fdabf86312bf265ec14af4e6c34828d3a1" Nov 25 15:51:35 crc kubenswrapper[4890]: I1125 15:51:35.603786 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23a9b8fa2451bcbb91ef90a4f685c7fdabf86312bf265ec14af4e6c34828d3a1"} err="failed to get container status \"23a9b8fa2451bcbb91ef90a4f685c7fdabf86312bf265ec14af4e6c34828d3a1\": rpc error: code = NotFound desc = could not find container \"23a9b8fa2451bcbb91ef90a4f685c7fdabf86312bf265ec14af4e6c34828d3a1\": container with ID starting with 23a9b8fa2451bcbb91ef90a4f685c7fdabf86312bf265ec14af4e6c34828d3a1 not found: ID does not exist" Nov 25 15:51:35 crc kubenswrapper[4890]: I1125 15:51:35.603851 4890 scope.go:117] "RemoveContainer" containerID="d2b45b3c99fde6d7fd14ba10971416462c939286af9b27ebca974fafe1b2db74" Nov 25 15:51:35 crc kubenswrapper[4890]: E1125 15:51:35.606984 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2b45b3c99fde6d7fd14ba10971416462c939286af9b27ebca974fafe1b2db74\": container with ID starting with d2b45b3c99fde6d7fd14ba10971416462c939286af9b27ebca974fafe1b2db74 not found: ID does not exist" containerID="d2b45b3c99fde6d7fd14ba10971416462c939286af9b27ebca974fafe1b2db74" Nov 25 15:51:35 crc kubenswrapper[4890]: I1125 15:51:35.607039 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2b45b3c99fde6d7fd14ba10971416462c939286af9b27ebca974fafe1b2db74"} err="failed to get container status \"d2b45b3c99fde6d7fd14ba10971416462c939286af9b27ebca974fafe1b2db74\": rpc error: code = NotFound desc = could not find container \"d2b45b3c99fde6d7fd14ba10971416462c939286af9b27ebca974fafe1b2db74\": container with ID starting with d2b45b3c99fde6d7fd14ba10971416462c939286af9b27ebca974fafe1b2db74 not found: ID does not exist" Nov 25 15:51:35 crc kubenswrapper[4890]: I1125 15:51:35.607078 4890 scope.go:117] "RemoveContainer" containerID="0185c639fb8a4401a43960c6c8c9962227adaeb83baf1ab7afd9a5099f90d8be" Nov 25 15:51:35 crc kubenswrapper[4890]: E1125 15:51:35.608114 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0185c639fb8a4401a43960c6c8c9962227adaeb83baf1ab7afd9a5099f90d8be\": container with ID starting with 0185c639fb8a4401a43960c6c8c9962227adaeb83baf1ab7afd9a5099f90d8be not found: ID does not exist" containerID="0185c639fb8a4401a43960c6c8c9962227adaeb83baf1ab7afd9a5099f90d8be" Nov 25 15:51:35 crc kubenswrapper[4890]: I1125 15:51:35.608182 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0185c639fb8a4401a43960c6c8c9962227adaeb83baf1ab7afd9a5099f90d8be"} err="failed to get container status \"0185c639fb8a4401a43960c6c8c9962227adaeb83baf1ab7afd9a5099f90d8be\": rpc error: code = NotFound desc = could not find container \"0185c639fb8a4401a43960c6c8c9962227adaeb83baf1ab7afd9a5099f90d8be\": container with ID starting with 0185c639fb8a4401a43960c6c8c9962227adaeb83baf1ab7afd9a5099f90d8be not found: ID does not exist" Nov 25 15:51:36 crc kubenswrapper[4890]: I1125 15:51:36.188698 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cf2f167-8a28-4bcf-891a-1f7cf91d433b" path="/var/lib/kubelet/pods/5cf2f167-8a28-4bcf-891a-1f7cf91d433b/volumes" Nov 25 15:51:38 crc kubenswrapper[4890]: I1125 15:51:38.174015 4890 scope.go:117] "RemoveContainer" containerID="aa8a6c3d508388bee84226737d06d0443dd3e6d511ad22e6e9e1c1f48362f983" Nov 25 15:51:38 crc kubenswrapper[4890]: E1125 15:51:38.174916 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:51:53 crc kubenswrapper[4890]: I1125 15:51:53.175033 4890 scope.go:117] "RemoveContainer" containerID="aa8a6c3d508388bee84226737d06d0443dd3e6d511ad22e6e9e1c1f48362f983" Nov 25 15:51:53 crc kubenswrapper[4890]: E1125 15:51:53.175918 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:52:08 crc kubenswrapper[4890]: I1125 15:52:08.173818 4890 scope.go:117] "RemoveContainer" containerID="aa8a6c3d508388bee84226737d06d0443dd3e6d511ad22e6e9e1c1f48362f983" Nov 25 15:52:08 crc kubenswrapper[4890]: E1125 15:52:08.174984 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:52:20 crc kubenswrapper[4890]: I1125 15:52:20.173958 4890 scope.go:117] "RemoveContainer" containerID="aa8a6c3d508388bee84226737d06d0443dd3e6d511ad22e6e9e1c1f48362f983" Nov 25 15:52:20 crc kubenswrapper[4890]: E1125 15:52:20.175060 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:52:34 crc kubenswrapper[4890]: I1125 15:52:34.173772 4890 scope.go:117] "RemoveContainer" containerID="aa8a6c3d508388bee84226737d06d0443dd3e6d511ad22e6e9e1c1f48362f983" Nov 25 15:52:34 crc kubenswrapper[4890]: E1125 15:52:34.174861 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:52:47 crc kubenswrapper[4890]: I1125 15:52:47.173702 4890 scope.go:117] "RemoveContainer" containerID="aa8a6c3d508388bee84226737d06d0443dd3e6d511ad22e6e9e1c1f48362f983" Nov 25 15:52:47 crc kubenswrapper[4890]: E1125 15:52:47.174647 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:52:59 crc kubenswrapper[4890]: I1125 15:52:59.173858 4890 scope.go:117] "RemoveContainer" containerID="aa8a6c3d508388bee84226737d06d0443dd3e6d511ad22e6e9e1c1f48362f983" Nov 25 15:52:59 crc kubenswrapper[4890]: E1125 15:52:59.175015 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:53:12 crc kubenswrapper[4890]: I1125 15:53:12.183856 4890 scope.go:117] "RemoveContainer" containerID="aa8a6c3d508388bee84226737d06d0443dd3e6d511ad22e6e9e1c1f48362f983" Nov 25 15:53:12 crc kubenswrapper[4890]: E1125 15:53:12.184940 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:53:23 crc kubenswrapper[4890]: I1125 15:53:23.173462 4890 scope.go:117] "RemoveContainer" containerID="aa8a6c3d508388bee84226737d06d0443dd3e6d511ad22e6e9e1c1f48362f983" Nov 25 15:53:23 crc kubenswrapper[4890]: E1125 15:53:23.174593 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:53:37 crc kubenswrapper[4890]: I1125 15:53:37.174133 4890 scope.go:117] "RemoveContainer" containerID="aa8a6c3d508388bee84226737d06d0443dd3e6d511ad22e6e9e1c1f48362f983" Nov 25 15:53:37 crc kubenswrapper[4890]: E1125 15:53:37.175793 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:53:39 crc kubenswrapper[4890]: I1125 15:53:39.894825 4890 scope.go:117] "RemoveContainer" containerID="47b8c53403e63ed3ef82270e4d951c4f015aa63383a40b0014c1d582e2d376b7" Nov 25 15:53:39 crc kubenswrapper[4890]: I1125 15:53:39.932746 4890 scope.go:117] "RemoveContainer" containerID="3052e0c81565aa4a5bd2d86213f0e3237fa2f2d064aeb247b5cc1b217c88f0ef" Nov 25 15:53:39 crc kubenswrapper[4890]: I1125 15:53:39.962983 4890 scope.go:117] "RemoveContainer" containerID="79f468bba9912200308199ae68bbcdd8aa35aca16f2fd3bd09bba697bc426fbf" Nov 25 15:53:48 crc kubenswrapper[4890]: I1125 15:53:48.174187 4890 scope.go:117] "RemoveContainer" containerID="aa8a6c3d508388bee84226737d06d0443dd3e6d511ad22e6e9e1c1f48362f983" Nov 25 15:53:48 crc kubenswrapper[4890]: E1125 15:53:48.175179 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:54:01 crc kubenswrapper[4890]: I1125 15:54:01.173654 4890 scope.go:117] "RemoveContainer" containerID="aa8a6c3d508388bee84226737d06d0443dd3e6d511ad22e6e9e1c1f48362f983" Nov 25 15:54:01 crc kubenswrapper[4890]: E1125 15:54:01.174462 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:54:15 crc kubenswrapper[4890]: I1125 15:54:15.173473 4890 scope.go:117] "RemoveContainer" containerID="aa8a6c3d508388bee84226737d06d0443dd3e6d511ad22e6e9e1c1f48362f983" Nov 25 15:54:15 crc kubenswrapper[4890]: E1125 15:54:15.174543 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:54:29 crc kubenswrapper[4890]: I1125 15:54:29.172956 4890 scope.go:117] "RemoveContainer" containerID="aa8a6c3d508388bee84226737d06d0443dd3e6d511ad22e6e9e1c1f48362f983" Nov 25 15:54:29 crc kubenswrapper[4890]: E1125 15:54:29.173835 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:54:43 crc kubenswrapper[4890]: I1125 15:54:43.174070 4890 scope.go:117] "RemoveContainer" containerID="aa8a6c3d508388bee84226737d06d0443dd3e6d511ad22e6e9e1c1f48362f983" Nov 25 15:54:43 crc kubenswrapper[4890]: E1125 15:54:43.175732 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:54:58 crc kubenswrapper[4890]: I1125 15:54:58.174480 4890 scope.go:117] "RemoveContainer" containerID="aa8a6c3d508388bee84226737d06d0443dd3e6d511ad22e6e9e1c1f48362f983" Nov 25 15:54:58 crc kubenswrapper[4890]: E1125 15:54:58.175385 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:55:12 crc kubenswrapper[4890]: I1125 15:55:12.182578 4890 scope.go:117] "RemoveContainer" containerID="aa8a6c3d508388bee84226737d06d0443dd3e6d511ad22e6e9e1c1f48362f983" Nov 25 15:55:12 crc kubenswrapper[4890]: E1125 15:55:12.183538 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:55:25 crc kubenswrapper[4890]: I1125 15:55:25.174426 4890 scope.go:117] "RemoveContainer" containerID="aa8a6c3d508388bee84226737d06d0443dd3e6d511ad22e6e9e1c1f48362f983" Nov 25 15:55:25 crc kubenswrapper[4890]: E1125 15:55:25.175373 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:55:39 crc kubenswrapper[4890]: I1125 15:55:39.173506 4890 scope.go:117] "RemoveContainer" containerID="aa8a6c3d508388bee84226737d06d0443dd3e6d511ad22e6e9e1c1f48362f983" Nov 25 15:55:39 crc kubenswrapper[4890]: E1125 15:55:39.174532 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:55:54 crc kubenswrapper[4890]: I1125 15:55:54.173066 4890 scope.go:117] "RemoveContainer" containerID="aa8a6c3d508388bee84226737d06d0443dd3e6d511ad22e6e9e1c1f48362f983" Nov 25 15:55:54 crc kubenswrapper[4890]: E1125 15:55:54.174373 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 15:56:07 crc kubenswrapper[4890]: I1125 15:56:07.174297 4890 scope.go:117] "RemoveContainer" containerID="aa8a6c3d508388bee84226737d06d0443dd3e6d511ad22e6e9e1c1f48362f983" Nov 25 15:56:07 crc kubenswrapper[4890]: I1125 15:56:07.635003 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" event={"ID":"4e4f849d-f239-4727-a73e-18327856929a","Type":"ContainerStarted","Data":"54ca4ab07f0e1f73cd740dc4474b2b043d2b98ae0c3d04e251edd62a255d4192"} Nov 25 15:56:39 crc kubenswrapper[4890]: I1125 15:56:39.043063 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-z4mtk"] Nov 25 15:56:39 crc kubenswrapper[4890]: E1125 15:56:39.044466 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cf2f167-8a28-4bcf-891a-1f7cf91d433b" containerName="registry-server" Nov 25 15:56:39 crc kubenswrapper[4890]: I1125 15:56:39.044486 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cf2f167-8a28-4bcf-891a-1f7cf91d433b" containerName="registry-server" Nov 25 15:56:39 crc kubenswrapper[4890]: E1125 15:56:39.044509 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cf2f167-8a28-4bcf-891a-1f7cf91d433b" containerName="extract-utilities" Nov 25 15:56:39 crc kubenswrapper[4890]: I1125 15:56:39.044517 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cf2f167-8a28-4bcf-891a-1f7cf91d433b" containerName="extract-utilities" Nov 25 15:56:39 crc kubenswrapper[4890]: E1125 15:56:39.044589 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cf2f167-8a28-4bcf-891a-1f7cf91d433b" containerName="extract-content" Nov 25 15:56:39 crc kubenswrapper[4890]: I1125 15:56:39.044597 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cf2f167-8a28-4bcf-891a-1f7cf91d433b" containerName="extract-content" Nov 25 15:56:39 crc kubenswrapper[4890]: I1125 15:56:39.044850 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cf2f167-8a28-4bcf-891a-1f7cf91d433b" containerName="registry-server" Nov 25 15:56:39 crc kubenswrapper[4890]: I1125 15:56:39.047147 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z4mtk" Nov 25 15:56:39 crc kubenswrapper[4890]: I1125 15:56:39.078224 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-z4mtk"] Nov 25 15:56:39 crc kubenswrapper[4890]: I1125 15:56:39.151900 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f3e81ce-242c-4dd0-be0e-507f98ec5bcc-catalog-content\") pod \"redhat-marketplace-z4mtk\" (UID: \"7f3e81ce-242c-4dd0-be0e-507f98ec5bcc\") " pod="openshift-marketplace/redhat-marketplace-z4mtk" Nov 25 15:56:39 crc kubenswrapper[4890]: I1125 15:56:39.151962 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7pdd\" (UniqueName: \"kubernetes.io/projected/7f3e81ce-242c-4dd0-be0e-507f98ec5bcc-kube-api-access-v7pdd\") pod \"redhat-marketplace-z4mtk\" (UID: \"7f3e81ce-242c-4dd0-be0e-507f98ec5bcc\") " pod="openshift-marketplace/redhat-marketplace-z4mtk" Nov 25 15:56:39 crc kubenswrapper[4890]: I1125 15:56:39.152066 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f3e81ce-242c-4dd0-be0e-507f98ec5bcc-utilities\") pod \"redhat-marketplace-z4mtk\" (UID: \"7f3e81ce-242c-4dd0-be0e-507f98ec5bcc\") " pod="openshift-marketplace/redhat-marketplace-z4mtk" Nov 25 15:56:39 crc kubenswrapper[4890]: I1125 15:56:39.254874 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f3e81ce-242c-4dd0-be0e-507f98ec5bcc-catalog-content\") pod \"redhat-marketplace-z4mtk\" (UID: \"7f3e81ce-242c-4dd0-be0e-507f98ec5bcc\") " pod="openshift-marketplace/redhat-marketplace-z4mtk" Nov 25 15:56:39 crc kubenswrapper[4890]: I1125 15:56:39.254973 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7pdd\" (UniqueName: \"kubernetes.io/projected/7f3e81ce-242c-4dd0-be0e-507f98ec5bcc-kube-api-access-v7pdd\") pod \"redhat-marketplace-z4mtk\" (UID: \"7f3e81ce-242c-4dd0-be0e-507f98ec5bcc\") " pod="openshift-marketplace/redhat-marketplace-z4mtk" Nov 25 15:56:39 crc kubenswrapper[4890]: I1125 15:56:39.255075 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f3e81ce-242c-4dd0-be0e-507f98ec5bcc-utilities\") pod \"redhat-marketplace-z4mtk\" (UID: \"7f3e81ce-242c-4dd0-be0e-507f98ec5bcc\") " pod="openshift-marketplace/redhat-marketplace-z4mtk" Nov 25 15:56:39 crc kubenswrapper[4890]: I1125 15:56:39.255664 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f3e81ce-242c-4dd0-be0e-507f98ec5bcc-catalog-content\") pod \"redhat-marketplace-z4mtk\" (UID: \"7f3e81ce-242c-4dd0-be0e-507f98ec5bcc\") " pod="openshift-marketplace/redhat-marketplace-z4mtk" Nov 25 15:56:39 crc kubenswrapper[4890]: I1125 15:56:39.255710 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f3e81ce-242c-4dd0-be0e-507f98ec5bcc-utilities\") pod \"redhat-marketplace-z4mtk\" (UID: \"7f3e81ce-242c-4dd0-be0e-507f98ec5bcc\") " pod="openshift-marketplace/redhat-marketplace-z4mtk" Nov 25 15:56:39 crc kubenswrapper[4890]: I1125 15:56:39.287391 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7pdd\" (UniqueName: \"kubernetes.io/projected/7f3e81ce-242c-4dd0-be0e-507f98ec5bcc-kube-api-access-v7pdd\") pod \"redhat-marketplace-z4mtk\" (UID: \"7f3e81ce-242c-4dd0-be0e-507f98ec5bcc\") " pod="openshift-marketplace/redhat-marketplace-z4mtk" Nov 25 15:56:39 crc kubenswrapper[4890]: I1125 15:56:39.375032 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z4mtk" Nov 25 15:56:39 crc kubenswrapper[4890]: I1125 15:56:39.826051 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-z4mtk"] Nov 25 15:56:40 crc kubenswrapper[4890]: I1125 15:56:40.061479 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z4mtk" event={"ID":"7f3e81ce-242c-4dd0-be0e-507f98ec5bcc","Type":"ContainerStarted","Data":"73375011bb0fb07d35ef176ccee8664e2f53ae9f988fdbebf076aa738a68b1ec"} Nov 25 15:56:41 crc kubenswrapper[4890]: I1125 15:56:41.077583 4890 generic.go:334] "Generic (PLEG): container finished" podID="7f3e81ce-242c-4dd0-be0e-507f98ec5bcc" containerID="ba21d5b44994ccab997c2978b912ab7e9b34b0844f58fe4fbbba2a68331294cb" exitCode=0 Nov 25 15:56:41 crc kubenswrapper[4890]: I1125 15:56:41.077665 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z4mtk" event={"ID":"7f3e81ce-242c-4dd0-be0e-507f98ec5bcc","Type":"ContainerDied","Data":"ba21d5b44994ccab997c2978b912ab7e9b34b0844f58fe4fbbba2a68331294cb"} Nov 25 15:56:41 crc kubenswrapper[4890]: I1125 15:56:41.080727 4890 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 15:56:43 crc kubenswrapper[4890]: I1125 15:56:43.133722 4890 generic.go:334] "Generic (PLEG): container finished" podID="7f3e81ce-242c-4dd0-be0e-507f98ec5bcc" containerID="0023a08ab05b1eac3435d82d1786a51f93cb09bd9baf2f8f5aab4cfa8909aa53" exitCode=0 Nov 25 15:56:43 crc kubenswrapper[4890]: I1125 15:56:43.133791 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z4mtk" event={"ID":"7f3e81ce-242c-4dd0-be0e-507f98ec5bcc","Type":"ContainerDied","Data":"0023a08ab05b1eac3435d82d1786a51f93cb09bd9baf2f8f5aab4cfa8909aa53"} Nov 25 15:56:44 crc kubenswrapper[4890]: I1125 15:56:44.154717 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z4mtk" event={"ID":"7f3e81ce-242c-4dd0-be0e-507f98ec5bcc","Type":"ContainerStarted","Data":"5d060f3d484589c72590906f0f73297145d974450ffac08d2082ba5fa31c5389"} Nov 25 15:56:44 crc kubenswrapper[4890]: I1125 15:56:44.189373 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-z4mtk" podStartSLOduration=2.695187567 podStartE2EDuration="5.189340662s" podCreationTimestamp="2025-11-25 15:56:39 +0000 UTC" firstStartedPulling="2025-11-25 15:56:41.08040347 +0000 UTC m=+3259.522866080" lastFinishedPulling="2025-11-25 15:56:43.574556565 +0000 UTC m=+3262.017019175" observedRunningTime="2025-11-25 15:56:44.18041905 +0000 UTC m=+3262.622881660" watchObservedRunningTime="2025-11-25 15:56:44.189340662 +0000 UTC m=+3262.631803292" Nov 25 15:56:49 crc kubenswrapper[4890]: I1125 15:56:49.375805 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-z4mtk" Nov 25 15:56:49 crc kubenswrapper[4890]: I1125 15:56:49.377853 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-z4mtk" Nov 25 15:56:49 crc kubenswrapper[4890]: I1125 15:56:49.432658 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-z4mtk" Nov 25 15:56:50 crc kubenswrapper[4890]: I1125 15:56:50.299286 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-z4mtk" Nov 25 15:56:50 crc kubenswrapper[4890]: I1125 15:56:50.406026 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-z4mtk"] Nov 25 15:56:52 crc kubenswrapper[4890]: I1125 15:56:52.269665 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-z4mtk" podUID="7f3e81ce-242c-4dd0-be0e-507f98ec5bcc" containerName="registry-server" containerID="cri-o://5d060f3d484589c72590906f0f73297145d974450ffac08d2082ba5fa31c5389" gracePeriod=2 Nov 25 15:56:53 crc kubenswrapper[4890]: I1125 15:56:53.289922 4890 generic.go:334] "Generic (PLEG): container finished" podID="7f3e81ce-242c-4dd0-be0e-507f98ec5bcc" containerID="5d060f3d484589c72590906f0f73297145d974450ffac08d2082ba5fa31c5389" exitCode=0 Nov 25 15:56:53 crc kubenswrapper[4890]: I1125 15:56:53.289977 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z4mtk" event={"ID":"7f3e81ce-242c-4dd0-be0e-507f98ec5bcc","Type":"ContainerDied","Data":"5d060f3d484589c72590906f0f73297145d974450ffac08d2082ba5fa31c5389"} Nov 25 15:56:53 crc kubenswrapper[4890]: I1125 15:56:53.508674 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z4mtk" Nov 25 15:56:53 crc kubenswrapper[4890]: I1125 15:56:53.611832 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f3e81ce-242c-4dd0-be0e-507f98ec5bcc-catalog-content\") pod \"7f3e81ce-242c-4dd0-be0e-507f98ec5bcc\" (UID: \"7f3e81ce-242c-4dd0-be0e-507f98ec5bcc\") " Nov 25 15:56:53 crc kubenswrapper[4890]: I1125 15:56:53.611928 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7pdd\" (UniqueName: \"kubernetes.io/projected/7f3e81ce-242c-4dd0-be0e-507f98ec5bcc-kube-api-access-v7pdd\") pod \"7f3e81ce-242c-4dd0-be0e-507f98ec5bcc\" (UID: \"7f3e81ce-242c-4dd0-be0e-507f98ec5bcc\") " Nov 25 15:56:53 crc kubenswrapper[4890]: I1125 15:56:53.612072 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f3e81ce-242c-4dd0-be0e-507f98ec5bcc-utilities\") pod \"7f3e81ce-242c-4dd0-be0e-507f98ec5bcc\" (UID: \"7f3e81ce-242c-4dd0-be0e-507f98ec5bcc\") " Nov 25 15:56:53 crc kubenswrapper[4890]: I1125 15:56:53.614750 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f3e81ce-242c-4dd0-be0e-507f98ec5bcc-utilities" (OuterVolumeSpecName: "utilities") pod "7f3e81ce-242c-4dd0-be0e-507f98ec5bcc" (UID: "7f3e81ce-242c-4dd0-be0e-507f98ec5bcc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:56:53 crc kubenswrapper[4890]: I1125 15:56:53.621653 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f3e81ce-242c-4dd0-be0e-507f98ec5bcc-kube-api-access-v7pdd" (OuterVolumeSpecName: "kube-api-access-v7pdd") pod "7f3e81ce-242c-4dd0-be0e-507f98ec5bcc" (UID: "7f3e81ce-242c-4dd0-be0e-507f98ec5bcc"). InnerVolumeSpecName "kube-api-access-v7pdd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:56:53 crc kubenswrapper[4890]: I1125 15:56:53.714951 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7pdd\" (UniqueName: \"kubernetes.io/projected/7f3e81ce-242c-4dd0-be0e-507f98ec5bcc-kube-api-access-v7pdd\") on node \"crc\" DevicePath \"\"" Nov 25 15:56:53 crc kubenswrapper[4890]: I1125 15:56:53.715004 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f3e81ce-242c-4dd0-be0e-507f98ec5bcc-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 15:56:53 crc kubenswrapper[4890]: I1125 15:56:53.856294 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f3e81ce-242c-4dd0-be0e-507f98ec5bcc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7f3e81ce-242c-4dd0-be0e-507f98ec5bcc" (UID: "7f3e81ce-242c-4dd0-be0e-507f98ec5bcc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:56:53 crc kubenswrapper[4890]: I1125 15:56:53.920198 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f3e81ce-242c-4dd0-be0e-507f98ec5bcc-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 15:56:54 crc kubenswrapper[4890]: I1125 15:56:54.305463 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z4mtk" event={"ID":"7f3e81ce-242c-4dd0-be0e-507f98ec5bcc","Type":"ContainerDied","Data":"73375011bb0fb07d35ef176ccee8664e2f53ae9f988fdbebf076aa738a68b1ec"} Nov 25 15:56:54 crc kubenswrapper[4890]: I1125 15:56:54.305540 4890 scope.go:117] "RemoveContainer" containerID="5d060f3d484589c72590906f0f73297145d974450ffac08d2082ba5fa31c5389" Nov 25 15:56:54 crc kubenswrapper[4890]: I1125 15:56:54.305543 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z4mtk" Nov 25 15:56:54 crc kubenswrapper[4890]: I1125 15:56:54.343369 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-z4mtk"] Nov 25 15:56:54 crc kubenswrapper[4890]: I1125 15:56:54.346305 4890 scope.go:117] "RemoveContainer" containerID="0023a08ab05b1eac3435d82d1786a51f93cb09bd9baf2f8f5aab4cfa8909aa53" Nov 25 15:56:54 crc kubenswrapper[4890]: I1125 15:56:54.358034 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-z4mtk"] Nov 25 15:56:54 crc kubenswrapper[4890]: I1125 15:56:54.375698 4890 scope.go:117] "RemoveContainer" containerID="ba21d5b44994ccab997c2978b912ab7e9b34b0844f58fe4fbbba2a68331294cb" Nov 25 15:56:56 crc kubenswrapper[4890]: I1125 15:56:56.187013 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f3e81ce-242c-4dd0-be0e-507f98ec5bcc" path="/var/lib/kubelet/pods/7f3e81ce-242c-4dd0-be0e-507f98ec5bcc/volumes" Nov 25 15:57:39 crc kubenswrapper[4890]: I1125 15:57:39.524322 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hm69q"] Nov 25 15:57:39 crc kubenswrapper[4890]: E1125 15:57:39.525445 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f3e81ce-242c-4dd0-be0e-507f98ec5bcc" containerName="registry-server" Nov 25 15:57:39 crc kubenswrapper[4890]: I1125 15:57:39.525464 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f3e81ce-242c-4dd0-be0e-507f98ec5bcc" containerName="registry-server" Nov 25 15:57:39 crc kubenswrapper[4890]: E1125 15:57:39.525484 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f3e81ce-242c-4dd0-be0e-507f98ec5bcc" containerName="extract-utilities" Nov 25 15:57:39 crc kubenswrapper[4890]: I1125 15:57:39.525493 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f3e81ce-242c-4dd0-be0e-507f98ec5bcc" containerName="extract-utilities" Nov 25 15:57:39 crc kubenswrapper[4890]: E1125 15:57:39.525523 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f3e81ce-242c-4dd0-be0e-507f98ec5bcc" containerName="extract-content" Nov 25 15:57:39 crc kubenswrapper[4890]: I1125 15:57:39.525532 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f3e81ce-242c-4dd0-be0e-507f98ec5bcc" containerName="extract-content" Nov 25 15:57:39 crc kubenswrapper[4890]: I1125 15:57:39.525849 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f3e81ce-242c-4dd0-be0e-507f98ec5bcc" containerName="registry-server" Nov 25 15:57:39 crc kubenswrapper[4890]: I1125 15:57:39.528378 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hm69q" Nov 25 15:57:39 crc kubenswrapper[4890]: I1125 15:57:39.549123 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hm69q"] Nov 25 15:57:39 crc kubenswrapper[4890]: I1125 15:57:39.650659 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19677448-ff37-442f-8316-1dd3fb856119-catalog-content\") pod \"certified-operators-hm69q\" (UID: \"19677448-ff37-442f-8316-1dd3fb856119\") " pod="openshift-marketplace/certified-operators-hm69q" Nov 25 15:57:39 crc kubenswrapper[4890]: I1125 15:57:39.650755 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2spg\" (UniqueName: \"kubernetes.io/projected/19677448-ff37-442f-8316-1dd3fb856119-kube-api-access-d2spg\") pod \"certified-operators-hm69q\" (UID: \"19677448-ff37-442f-8316-1dd3fb856119\") " pod="openshift-marketplace/certified-operators-hm69q" Nov 25 15:57:39 crc kubenswrapper[4890]: I1125 15:57:39.650820 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19677448-ff37-442f-8316-1dd3fb856119-utilities\") pod \"certified-operators-hm69q\" (UID: \"19677448-ff37-442f-8316-1dd3fb856119\") " pod="openshift-marketplace/certified-operators-hm69q" Nov 25 15:57:39 crc kubenswrapper[4890]: I1125 15:57:39.753924 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19677448-ff37-442f-8316-1dd3fb856119-catalog-content\") pod \"certified-operators-hm69q\" (UID: \"19677448-ff37-442f-8316-1dd3fb856119\") " pod="openshift-marketplace/certified-operators-hm69q" Nov 25 15:57:39 crc kubenswrapper[4890]: I1125 15:57:39.754038 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2spg\" (UniqueName: \"kubernetes.io/projected/19677448-ff37-442f-8316-1dd3fb856119-kube-api-access-d2spg\") pod \"certified-operators-hm69q\" (UID: \"19677448-ff37-442f-8316-1dd3fb856119\") " pod="openshift-marketplace/certified-operators-hm69q" Nov 25 15:57:39 crc kubenswrapper[4890]: I1125 15:57:39.754123 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19677448-ff37-442f-8316-1dd3fb856119-utilities\") pod \"certified-operators-hm69q\" (UID: \"19677448-ff37-442f-8316-1dd3fb856119\") " pod="openshift-marketplace/certified-operators-hm69q" Nov 25 15:57:39 crc kubenswrapper[4890]: I1125 15:57:39.754780 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19677448-ff37-442f-8316-1dd3fb856119-utilities\") pod \"certified-operators-hm69q\" (UID: \"19677448-ff37-442f-8316-1dd3fb856119\") " pod="openshift-marketplace/certified-operators-hm69q" Nov 25 15:57:39 crc kubenswrapper[4890]: I1125 15:57:39.755132 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19677448-ff37-442f-8316-1dd3fb856119-catalog-content\") pod \"certified-operators-hm69q\" (UID: \"19677448-ff37-442f-8316-1dd3fb856119\") " pod="openshift-marketplace/certified-operators-hm69q" Nov 25 15:57:39 crc kubenswrapper[4890]: I1125 15:57:39.783346 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2spg\" (UniqueName: \"kubernetes.io/projected/19677448-ff37-442f-8316-1dd3fb856119-kube-api-access-d2spg\") pod \"certified-operators-hm69q\" (UID: \"19677448-ff37-442f-8316-1dd3fb856119\") " pod="openshift-marketplace/certified-operators-hm69q" Nov 25 15:57:39 crc kubenswrapper[4890]: I1125 15:57:39.859922 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hm69q" Nov 25 15:57:40 crc kubenswrapper[4890]: I1125 15:57:40.420736 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hm69q"] Nov 25 15:57:40 crc kubenswrapper[4890]: I1125 15:57:40.825178 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hm69q" event={"ID":"19677448-ff37-442f-8316-1dd3fb856119","Type":"ContainerStarted","Data":"c231d0b721cbc49b9d5e35da56f132b1b516c0d3e44e1f13abdd576bf227da25"} Nov 25 15:57:41 crc kubenswrapper[4890]: I1125 15:57:41.839812 4890 generic.go:334] "Generic (PLEG): container finished" podID="19677448-ff37-442f-8316-1dd3fb856119" containerID="8d66f0014e53294e5b1ead5eea14c213951571cbdf077c364e36f4d0671a9b60" exitCode=0 Nov 25 15:57:41 crc kubenswrapper[4890]: I1125 15:57:41.839931 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hm69q" event={"ID":"19677448-ff37-442f-8316-1dd3fb856119","Type":"ContainerDied","Data":"8d66f0014e53294e5b1ead5eea14c213951571cbdf077c364e36f4d0671a9b60"} Nov 25 15:57:45 crc kubenswrapper[4890]: I1125 15:57:45.888552 4890 generic.go:334] "Generic (PLEG): container finished" podID="19677448-ff37-442f-8316-1dd3fb856119" containerID="1f0daf341f4decb4e1b77e8a186c2922b6a8258917ab0d9023dc8eaac8cec002" exitCode=0 Nov 25 15:57:45 crc kubenswrapper[4890]: I1125 15:57:45.888649 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hm69q" event={"ID":"19677448-ff37-442f-8316-1dd3fb856119","Type":"ContainerDied","Data":"1f0daf341f4decb4e1b77e8a186c2922b6a8258917ab0d9023dc8eaac8cec002"} Nov 25 15:57:50 crc kubenswrapper[4890]: I1125 15:57:50.965672 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hm69q" event={"ID":"19677448-ff37-442f-8316-1dd3fb856119","Type":"ContainerStarted","Data":"b58e4a73fc10b09ba37a543369af9b2621d8e4a2e5ff0ecd7f352a91f823a1f0"} Nov 25 15:57:52 crc kubenswrapper[4890]: I1125 15:57:52.009719 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hm69q" podStartSLOduration=4.454794734 podStartE2EDuration="13.009688386s" podCreationTimestamp="2025-11-25 15:57:39 +0000 UTC" firstStartedPulling="2025-11-25 15:57:41.841976994 +0000 UTC m=+3320.284439604" lastFinishedPulling="2025-11-25 15:57:50.396870656 +0000 UTC m=+3328.839333256" observedRunningTime="2025-11-25 15:57:51.999857532 +0000 UTC m=+3330.442320142" watchObservedRunningTime="2025-11-25 15:57:52.009688386 +0000 UTC m=+3330.452150996" Nov 25 15:57:59 crc kubenswrapper[4890]: I1125 15:57:59.860292 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hm69q" Nov 25 15:57:59 crc kubenswrapper[4890]: I1125 15:57:59.862413 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hm69q" Nov 25 15:57:59 crc kubenswrapper[4890]: I1125 15:57:59.911760 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hm69q" Nov 25 15:58:00 crc kubenswrapper[4890]: I1125 15:58:00.123969 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hm69q" Nov 25 15:58:00 crc kubenswrapper[4890]: I1125 15:58:00.186315 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hm69q"] Nov 25 15:58:02 crc kubenswrapper[4890]: I1125 15:58:02.088787 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hm69q" podUID="19677448-ff37-442f-8316-1dd3fb856119" containerName="registry-server" containerID="cri-o://b58e4a73fc10b09ba37a543369af9b2621d8e4a2e5ff0ecd7f352a91f823a1f0" gracePeriod=2 Nov 25 15:58:03 crc kubenswrapper[4890]: I1125 15:58:03.106042 4890 generic.go:334] "Generic (PLEG): container finished" podID="19677448-ff37-442f-8316-1dd3fb856119" containerID="b58e4a73fc10b09ba37a543369af9b2621d8e4a2e5ff0ecd7f352a91f823a1f0" exitCode=0 Nov 25 15:58:03 crc kubenswrapper[4890]: I1125 15:58:03.106128 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hm69q" event={"ID":"19677448-ff37-442f-8316-1dd3fb856119","Type":"ContainerDied","Data":"b58e4a73fc10b09ba37a543369af9b2621d8e4a2e5ff0ecd7f352a91f823a1f0"} Nov 25 15:58:03 crc kubenswrapper[4890]: I1125 15:58:03.279950 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hm69q" Nov 25 15:58:03 crc kubenswrapper[4890]: I1125 15:58:03.410095 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2spg\" (UniqueName: \"kubernetes.io/projected/19677448-ff37-442f-8316-1dd3fb856119-kube-api-access-d2spg\") pod \"19677448-ff37-442f-8316-1dd3fb856119\" (UID: \"19677448-ff37-442f-8316-1dd3fb856119\") " Nov 25 15:58:03 crc kubenswrapper[4890]: I1125 15:58:03.410571 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19677448-ff37-442f-8316-1dd3fb856119-utilities\") pod \"19677448-ff37-442f-8316-1dd3fb856119\" (UID: \"19677448-ff37-442f-8316-1dd3fb856119\") " Nov 25 15:58:03 crc kubenswrapper[4890]: I1125 15:58:03.410753 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19677448-ff37-442f-8316-1dd3fb856119-catalog-content\") pod \"19677448-ff37-442f-8316-1dd3fb856119\" (UID: \"19677448-ff37-442f-8316-1dd3fb856119\") " Nov 25 15:58:03 crc kubenswrapper[4890]: I1125 15:58:03.411540 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19677448-ff37-442f-8316-1dd3fb856119-utilities" (OuterVolumeSpecName: "utilities") pod "19677448-ff37-442f-8316-1dd3fb856119" (UID: "19677448-ff37-442f-8316-1dd3fb856119"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:58:03 crc kubenswrapper[4890]: I1125 15:58:03.418484 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19677448-ff37-442f-8316-1dd3fb856119-kube-api-access-d2spg" (OuterVolumeSpecName: "kube-api-access-d2spg") pod "19677448-ff37-442f-8316-1dd3fb856119" (UID: "19677448-ff37-442f-8316-1dd3fb856119"). InnerVolumeSpecName "kube-api-access-d2spg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 15:58:03 crc kubenswrapper[4890]: I1125 15:58:03.457896 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19677448-ff37-442f-8316-1dd3fb856119-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "19677448-ff37-442f-8316-1dd3fb856119" (UID: "19677448-ff37-442f-8316-1dd3fb856119"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 15:58:03 crc kubenswrapper[4890]: I1125 15:58:03.513995 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2spg\" (UniqueName: \"kubernetes.io/projected/19677448-ff37-442f-8316-1dd3fb856119-kube-api-access-d2spg\") on node \"crc\" DevicePath \"\"" Nov 25 15:58:03 crc kubenswrapper[4890]: I1125 15:58:03.514047 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19677448-ff37-442f-8316-1dd3fb856119-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 15:58:03 crc kubenswrapper[4890]: I1125 15:58:03.514061 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19677448-ff37-442f-8316-1dd3fb856119-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 15:58:04 crc kubenswrapper[4890]: I1125 15:58:04.129699 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hm69q" event={"ID":"19677448-ff37-442f-8316-1dd3fb856119","Type":"ContainerDied","Data":"c231d0b721cbc49b9d5e35da56f132b1b516c0d3e44e1f13abdd576bf227da25"} Nov 25 15:58:04 crc kubenswrapper[4890]: I1125 15:58:04.129816 4890 scope.go:117] "RemoveContainer" containerID="b58e4a73fc10b09ba37a543369af9b2621d8e4a2e5ff0ecd7f352a91f823a1f0" Nov 25 15:58:04 crc kubenswrapper[4890]: I1125 15:58:04.129815 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hm69q" Nov 25 15:58:04 crc kubenswrapper[4890]: I1125 15:58:04.163435 4890 scope.go:117] "RemoveContainer" containerID="1f0daf341f4decb4e1b77e8a186c2922b6a8258917ab0d9023dc8eaac8cec002" Nov 25 15:58:04 crc kubenswrapper[4890]: I1125 15:58:04.190218 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hm69q"] Nov 25 15:58:04 crc kubenswrapper[4890]: I1125 15:58:04.190298 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hm69q"] Nov 25 15:58:04 crc kubenswrapper[4890]: I1125 15:58:04.210799 4890 scope.go:117] "RemoveContainer" containerID="8d66f0014e53294e5b1ead5eea14c213951571cbdf077c364e36f4d0671a9b60" Nov 25 15:58:06 crc kubenswrapper[4890]: I1125 15:58:06.191839 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19677448-ff37-442f-8316-1dd3fb856119" path="/var/lib/kubelet/pods/19677448-ff37-442f-8316-1dd3fb856119/volumes" Nov 25 15:58:26 crc kubenswrapper[4890]: I1125 15:58:26.447714 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 15:58:26 crc kubenswrapper[4890]: I1125 15:58:26.448455 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 15:58:56 crc kubenswrapper[4890]: I1125 15:58:56.448319 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 15:58:56 crc kubenswrapper[4890]: I1125 15:58:56.449226 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 15:59:26 crc kubenswrapper[4890]: I1125 15:59:26.447316 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 15:59:26 crc kubenswrapper[4890]: I1125 15:59:26.449438 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 15:59:26 crc kubenswrapper[4890]: I1125 15:59:26.449529 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" Nov 25 15:59:26 crc kubenswrapper[4890]: I1125 15:59:26.450571 4890 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"54ca4ab07f0e1f73cd740dc4474b2b043d2b98ae0c3d04e251edd62a255d4192"} pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 15:59:26 crc kubenswrapper[4890]: I1125 15:59:26.450641 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" containerID="cri-o://54ca4ab07f0e1f73cd740dc4474b2b043d2b98ae0c3d04e251edd62a255d4192" gracePeriod=600 Nov 25 15:59:27 crc kubenswrapper[4890]: I1125 15:59:27.060877 4890 generic.go:334] "Generic (PLEG): container finished" podID="4e4f849d-f239-4727-a73e-18327856929a" containerID="54ca4ab07f0e1f73cd740dc4474b2b043d2b98ae0c3d04e251edd62a255d4192" exitCode=0 Nov 25 15:59:27 crc kubenswrapper[4890]: I1125 15:59:27.060972 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" event={"ID":"4e4f849d-f239-4727-a73e-18327856929a","Type":"ContainerDied","Data":"54ca4ab07f0e1f73cd740dc4474b2b043d2b98ae0c3d04e251edd62a255d4192"} Nov 25 15:59:27 crc kubenswrapper[4890]: I1125 15:59:27.061400 4890 scope.go:117] "RemoveContainer" containerID="aa8a6c3d508388bee84226737d06d0443dd3e6d511ad22e6e9e1c1f48362f983" Nov 25 15:59:28 crc kubenswrapper[4890]: I1125 15:59:28.083260 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" event={"ID":"4e4f849d-f239-4727-a73e-18327856929a","Type":"ContainerStarted","Data":"a0bac43b712e038c240449718f3b53c0d88f1d20e0f5e92b7dfb0385a9de4f68"} Nov 25 16:00:00 crc kubenswrapper[4890]: I1125 16:00:00.152238 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401440-mgqhf"] Nov 25 16:00:00 crc kubenswrapper[4890]: E1125 16:00:00.153428 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19677448-ff37-442f-8316-1dd3fb856119" containerName="extract-utilities" Nov 25 16:00:00 crc kubenswrapper[4890]: I1125 16:00:00.153448 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="19677448-ff37-442f-8316-1dd3fb856119" containerName="extract-utilities" Nov 25 16:00:00 crc kubenswrapper[4890]: E1125 16:00:00.153486 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19677448-ff37-442f-8316-1dd3fb856119" containerName="extract-content" Nov 25 16:00:00 crc kubenswrapper[4890]: I1125 16:00:00.153495 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="19677448-ff37-442f-8316-1dd3fb856119" containerName="extract-content" Nov 25 16:00:00 crc kubenswrapper[4890]: E1125 16:00:00.153546 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19677448-ff37-442f-8316-1dd3fb856119" containerName="registry-server" Nov 25 16:00:00 crc kubenswrapper[4890]: I1125 16:00:00.153552 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="19677448-ff37-442f-8316-1dd3fb856119" containerName="registry-server" Nov 25 16:00:00 crc kubenswrapper[4890]: I1125 16:00:00.153810 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="19677448-ff37-442f-8316-1dd3fb856119" containerName="registry-server" Nov 25 16:00:00 crc kubenswrapper[4890]: I1125 16:00:00.154766 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401440-mgqhf" Nov 25 16:00:00 crc kubenswrapper[4890]: I1125 16:00:00.157717 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 16:00:00 crc kubenswrapper[4890]: I1125 16:00:00.157922 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 16:00:00 crc kubenswrapper[4890]: I1125 16:00:00.207525 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401440-mgqhf"] Nov 25 16:00:00 crc kubenswrapper[4890]: I1125 16:00:00.274626 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6vdg\" (UniqueName: \"kubernetes.io/projected/90d634b8-cc51-4165-95e3-85cf39a6be96-kube-api-access-g6vdg\") pod \"collect-profiles-29401440-mgqhf\" (UID: \"90d634b8-cc51-4165-95e3-85cf39a6be96\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401440-mgqhf" Nov 25 16:00:00 crc kubenswrapper[4890]: I1125 16:00:00.275321 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/90d634b8-cc51-4165-95e3-85cf39a6be96-secret-volume\") pod \"collect-profiles-29401440-mgqhf\" (UID: \"90d634b8-cc51-4165-95e3-85cf39a6be96\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401440-mgqhf" Nov 25 16:00:00 crc kubenswrapper[4890]: I1125 16:00:00.275386 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/90d634b8-cc51-4165-95e3-85cf39a6be96-config-volume\") pod \"collect-profiles-29401440-mgqhf\" (UID: \"90d634b8-cc51-4165-95e3-85cf39a6be96\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401440-mgqhf" Nov 25 16:00:00 crc kubenswrapper[4890]: I1125 16:00:00.377900 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/90d634b8-cc51-4165-95e3-85cf39a6be96-secret-volume\") pod \"collect-profiles-29401440-mgqhf\" (UID: \"90d634b8-cc51-4165-95e3-85cf39a6be96\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401440-mgqhf" Nov 25 16:00:00 crc kubenswrapper[4890]: I1125 16:00:00.377990 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/90d634b8-cc51-4165-95e3-85cf39a6be96-config-volume\") pod \"collect-profiles-29401440-mgqhf\" (UID: \"90d634b8-cc51-4165-95e3-85cf39a6be96\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401440-mgqhf" Nov 25 16:00:00 crc kubenswrapper[4890]: I1125 16:00:00.378059 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6vdg\" (UniqueName: \"kubernetes.io/projected/90d634b8-cc51-4165-95e3-85cf39a6be96-kube-api-access-g6vdg\") pod \"collect-profiles-29401440-mgqhf\" (UID: \"90d634b8-cc51-4165-95e3-85cf39a6be96\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401440-mgqhf" Nov 25 16:00:00 crc kubenswrapper[4890]: I1125 16:00:00.379260 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/90d634b8-cc51-4165-95e3-85cf39a6be96-config-volume\") pod \"collect-profiles-29401440-mgqhf\" (UID: \"90d634b8-cc51-4165-95e3-85cf39a6be96\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401440-mgqhf" Nov 25 16:00:00 crc kubenswrapper[4890]: I1125 16:00:00.388125 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/90d634b8-cc51-4165-95e3-85cf39a6be96-secret-volume\") pod \"collect-profiles-29401440-mgqhf\" (UID: \"90d634b8-cc51-4165-95e3-85cf39a6be96\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401440-mgqhf" Nov 25 16:00:00 crc kubenswrapper[4890]: I1125 16:00:00.403595 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6vdg\" (UniqueName: \"kubernetes.io/projected/90d634b8-cc51-4165-95e3-85cf39a6be96-kube-api-access-g6vdg\") pod \"collect-profiles-29401440-mgqhf\" (UID: \"90d634b8-cc51-4165-95e3-85cf39a6be96\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401440-mgqhf" Nov 25 16:00:00 crc kubenswrapper[4890]: I1125 16:00:00.497022 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401440-mgqhf" Nov 25 16:00:01 crc kubenswrapper[4890]: I1125 16:00:01.024299 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401440-mgqhf"] Nov 25 16:00:01 crc kubenswrapper[4890]: I1125 16:00:01.460552 4890 generic.go:334] "Generic (PLEG): container finished" podID="90d634b8-cc51-4165-95e3-85cf39a6be96" containerID="b74b8f554c1062e03087e65da7db5a5d36ba47b67f9f1b60c395ad76a37ab657" exitCode=0 Nov 25 16:00:01 crc kubenswrapper[4890]: I1125 16:00:01.460640 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401440-mgqhf" event={"ID":"90d634b8-cc51-4165-95e3-85cf39a6be96","Type":"ContainerDied","Data":"b74b8f554c1062e03087e65da7db5a5d36ba47b67f9f1b60c395ad76a37ab657"} Nov 25 16:00:01 crc kubenswrapper[4890]: I1125 16:00:01.460930 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401440-mgqhf" event={"ID":"90d634b8-cc51-4165-95e3-85cf39a6be96","Type":"ContainerStarted","Data":"fa16f987d2cbad3db78f799ff069e752f95ed0e8ef9021831dc5e4a204b8145b"} Nov 25 16:00:02 crc kubenswrapper[4890]: I1125 16:00:02.860381 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401440-mgqhf" Nov 25 16:00:02 crc kubenswrapper[4890]: I1125 16:00:02.944952 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g6vdg\" (UniqueName: \"kubernetes.io/projected/90d634b8-cc51-4165-95e3-85cf39a6be96-kube-api-access-g6vdg\") pod \"90d634b8-cc51-4165-95e3-85cf39a6be96\" (UID: \"90d634b8-cc51-4165-95e3-85cf39a6be96\") " Nov 25 16:00:02 crc kubenswrapper[4890]: I1125 16:00:02.945044 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/90d634b8-cc51-4165-95e3-85cf39a6be96-secret-volume\") pod \"90d634b8-cc51-4165-95e3-85cf39a6be96\" (UID: \"90d634b8-cc51-4165-95e3-85cf39a6be96\") " Nov 25 16:00:02 crc kubenswrapper[4890]: I1125 16:00:02.945397 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/90d634b8-cc51-4165-95e3-85cf39a6be96-config-volume\") pod \"90d634b8-cc51-4165-95e3-85cf39a6be96\" (UID: \"90d634b8-cc51-4165-95e3-85cf39a6be96\") " Nov 25 16:00:02 crc kubenswrapper[4890]: I1125 16:00:02.946668 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90d634b8-cc51-4165-95e3-85cf39a6be96-config-volume" (OuterVolumeSpecName: "config-volume") pod "90d634b8-cc51-4165-95e3-85cf39a6be96" (UID: "90d634b8-cc51-4165-95e3-85cf39a6be96"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:00:02 crc kubenswrapper[4890]: I1125 16:00:02.953758 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90d634b8-cc51-4165-95e3-85cf39a6be96-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "90d634b8-cc51-4165-95e3-85cf39a6be96" (UID: "90d634b8-cc51-4165-95e3-85cf39a6be96"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:00:02 crc kubenswrapper[4890]: I1125 16:00:02.954464 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90d634b8-cc51-4165-95e3-85cf39a6be96-kube-api-access-g6vdg" (OuterVolumeSpecName: "kube-api-access-g6vdg") pod "90d634b8-cc51-4165-95e3-85cf39a6be96" (UID: "90d634b8-cc51-4165-95e3-85cf39a6be96"). InnerVolumeSpecName "kube-api-access-g6vdg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:00:03 crc kubenswrapper[4890]: I1125 16:00:03.048138 4890 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/90d634b8-cc51-4165-95e3-85cf39a6be96-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 16:00:03 crc kubenswrapper[4890]: I1125 16:00:03.048438 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g6vdg\" (UniqueName: \"kubernetes.io/projected/90d634b8-cc51-4165-95e3-85cf39a6be96-kube-api-access-g6vdg\") on node \"crc\" DevicePath \"\"" Nov 25 16:00:03 crc kubenswrapper[4890]: I1125 16:00:03.048632 4890 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/90d634b8-cc51-4165-95e3-85cf39a6be96-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 16:00:03 crc kubenswrapper[4890]: I1125 16:00:03.488142 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401440-mgqhf" event={"ID":"90d634b8-cc51-4165-95e3-85cf39a6be96","Type":"ContainerDied","Data":"fa16f987d2cbad3db78f799ff069e752f95ed0e8ef9021831dc5e4a204b8145b"} Nov 25 16:00:03 crc kubenswrapper[4890]: I1125 16:00:03.488237 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa16f987d2cbad3db78f799ff069e752f95ed0e8ef9021831dc5e4a204b8145b" Nov 25 16:00:03 crc kubenswrapper[4890]: I1125 16:00:03.490319 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401440-mgqhf" Nov 25 16:00:03 crc kubenswrapper[4890]: I1125 16:00:03.939474 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401395-hslqf"] Nov 25 16:00:03 crc kubenswrapper[4890]: I1125 16:00:03.947893 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401395-hslqf"] Nov 25 16:00:04 crc kubenswrapper[4890]: I1125 16:00:04.530763 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed35cb6c-f2b1-4d8a-9b43-782ca62594c7" path="/var/lib/kubelet/pods/ed35cb6c-f2b1-4d8a-9b43-782ca62594c7/volumes" Nov 25 16:00:38 crc kubenswrapper[4890]: I1125 16:00:38.063557 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9fzln"] Nov 25 16:00:38 crc kubenswrapper[4890]: E1125 16:00:38.065465 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90d634b8-cc51-4165-95e3-85cf39a6be96" containerName="collect-profiles" Nov 25 16:00:38 crc kubenswrapper[4890]: I1125 16:00:38.065487 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="90d634b8-cc51-4165-95e3-85cf39a6be96" containerName="collect-profiles" Nov 25 16:00:38 crc kubenswrapper[4890]: I1125 16:00:38.065759 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="90d634b8-cc51-4165-95e3-85cf39a6be96" containerName="collect-profiles" Nov 25 16:00:38 crc kubenswrapper[4890]: I1125 16:00:38.067795 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9fzln" Nov 25 16:00:38 crc kubenswrapper[4890]: I1125 16:00:38.081294 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9fzln"] Nov 25 16:00:38 crc kubenswrapper[4890]: I1125 16:00:38.175922 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b596a32-948b-4c02-aa48-1125c0e9b2fb-catalog-content\") pod \"community-operators-9fzln\" (UID: \"2b596a32-948b-4c02-aa48-1125c0e9b2fb\") " pod="openshift-marketplace/community-operators-9fzln" Nov 25 16:00:38 crc kubenswrapper[4890]: I1125 16:00:38.176179 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b596a32-948b-4c02-aa48-1125c0e9b2fb-utilities\") pod \"community-operators-9fzln\" (UID: \"2b596a32-948b-4c02-aa48-1125c0e9b2fb\") " pod="openshift-marketplace/community-operators-9fzln" Nov 25 16:00:38 crc kubenswrapper[4890]: I1125 16:00:38.176491 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swxcr\" (UniqueName: \"kubernetes.io/projected/2b596a32-948b-4c02-aa48-1125c0e9b2fb-kube-api-access-swxcr\") pod \"community-operators-9fzln\" (UID: \"2b596a32-948b-4c02-aa48-1125c0e9b2fb\") " pod="openshift-marketplace/community-operators-9fzln" Nov 25 16:00:38 crc kubenswrapper[4890]: I1125 16:00:38.278696 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swxcr\" (UniqueName: \"kubernetes.io/projected/2b596a32-948b-4c02-aa48-1125c0e9b2fb-kube-api-access-swxcr\") pod \"community-operators-9fzln\" (UID: \"2b596a32-948b-4c02-aa48-1125c0e9b2fb\") " pod="openshift-marketplace/community-operators-9fzln" Nov 25 16:00:38 crc kubenswrapper[4890]: I1125 16:00:38.278812 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b596a32-948b-4c02-aa48-1125c0e9b2fb-catalog-content\") pod \"community-operators-9fzln\" (UID: \"2b596a32-948b-4c02-aa48-1125c0e9b2fb\") " pod="openshift-marketplace/community-operators-9fzln" Nov 25 16:00:38 crc kubenswrapper[4890]: I1125 16:00:38.278926 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b596a32-948b-4c02-aa48-1125c0e9b2fb-utilities\") pod \"community-operators-9fzln\" (UID: \"2b596a32-948b-4c02-aa48-1125c0e9b2fb\") " pod="openshift-marketplace/community-operators-9fzln" Nov 25 16:00:38 crc kubenswrapper[4890]: I1125 16:00:38.279450 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b596a32-948b-4c02-aa48-1125c0e9b2fb-catalog-content\") pod \"community-operators-9fzln\" (UID: \"2b596a32-948b-4c02-aa48-1125c0e9b2fb\") " pod="openshift-marketplace/community-operators-9fzln" Nov 25 16:00:38 crc kubenswrapper[4890]: I1125 16:00:38.279508 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b596a32-948b-4c02-aa48-1125c0e9b2fb-utilities\") pod \"community-operators-9fzln\" (UID: \"2b596a32-948b-4c02-aa48-1125c0e9b2fb\") " pod="openshift-marketplace/community-operators-9fzln" Nov 25 16:00:38 crc kubenswrapper[4890]: I1125 16:00:38.312124 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swxcr\" (UniqueName: \"kubernetes.io/projected/2b596a32-948b-4c02-aa48-1125c0e9b2fb-kube-api-access-swxcr\") pod \"community-operators-9fzln\" (UID: \"2b596a32-948b-4c02-aa48-1125c0e9b2fb\") " pod="openshift-marketplace/community-operators-9fzln" Nov 25 16:00:38 crc kubenswrapper[4890]: I1125 16:00:38.397894 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9fzln" Nov 25 16:00:38 crc kubenswrapper[4890]: I1125 16:00:38.990232 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9fzln"] Nov 25 16:00:39 crc kubenswrapper[4890]: I1125 16:00:39.966603 4890 generic.go:334] "Generic (PLEG): container finished" podID="2b596a32-948b-4c02-aa48-1125c0e9b2fb" containerID="7d378d902f48b858de4ea20ce87c7cd569ad5f804db666ec276cebe75ed278e2" exitCode=0 Nov 25 16:00:39 crc kubenswrapper[4890]: I1125 16:00:39.967212 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9fzln" event={"ID":"2b596a32-948b-4c02-aa48-1125c0e9b2fb","Type":"ContainerDied","Data":"7d378d902f48b858de4ea20ce87c7cd569ad5f804db666ec276cebe75ed278e2"} Nov 25 16:00:39 crc kubenswrapper[4890]: I1125 16:00:39.967266 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9fzln" event={"ID":"2b596a32-948b-4c02-aa48-1125c0e9b2fb","Type":"ContainerStarted","Data":"f4bde364198d5454a0edc207a928f38b4b6aca8c7654c590e5e5ef8ab6a7b72f"} Nov 25 16:00:40 crc kubenswrapper[4890]: I1125 16:00:40.270202 4890 scope.go:117] "RemoveContainer" containerID="2229f386779c716886cbedd7df7fc219b411ba788eae258f76d6836ac9dc6d8f" Nov 25 16:00:41 crc kubenswrapper[4890]: I1125 16:00:41.989747 4890 generic.go:334] "Generic (PLEG): container finished" podID="2b596a32-948b-4c02-aa48-1125c0e9b2fb" containerID="c998bbb00494acc38800b7420b74b96c3e0323ca91f67ec96d3c979ec2933f0b" exitCode=0 Nov 25 16:00:41 crc kubenswrapper[4890]: I1125 16:00:41.989825 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9fzln" event={"ID":"2b596a32-948b-4c02-aa48-1125c0e9b2fb","Type":"ContainerDied","Data":"c998bbb00494acc38800b7420b74b96c3e0323ca91f67ec96d3c979ec2933f0b"} Nov 25 16:00:43 crc kubenswrapper[4890]: I1125 16:00:43.004055 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9fzln" event={"ID":"2b596a32-948b-4c02-aa48-1125c0e9b2fb","Type":"ContainerStarted","Data":"b2c8d5aec867dc313755b82e5c8dccbb5a09789422d65455ae9133403580811e"} Nov 25 16:00:48 crc kubenswrapper[4890]: I1125 16:00:48.398365 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9fzln" Nov 25 16:00:48 crc kubenswrapper[4890]: I1125 16:00:48.400611 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9fzln" Nov 25 16:00:48 crc kubenswrapper[4890]: I1125 16:00:48.447132 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9fzln" Nov 25 16:00:48 crc kubenswrapper[4890]: I1125 16:00:48.480493 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9fzln" podStartSLOduration=7.987561532 podStartE2EDuration="10.480463461s" podCreationTimestamp="2025-11-25 16:00:38 +0000 UTC" firstStartedPulling="2025-11-25 16:00:39.974578901 +0000 UTC m=+3498.417041511" lastFinishedPulling="2025-11-25 16:00:42.46748083 +0000 UTC m=+3500.909943440" observedRunningTime="2025-11-25 16:00:43.027299253 +0000 UTC m=+3501.469761873" watchObservedRunningTime="2025-11-25 16:00:48.480463461 +0000 UTC m=+3506.922926081" Nov 25 16:00:49 crc kubenswrapper[4890]: I1125 16:00:49.119765 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9fzln" Nov 25 16:00:49 crc kubenswrapper[4890]: I1125 16:00:49.178238 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9fzln"] Nov 25 16:00:51 crc kubenswrapper[4890]: I1125 16:00:51.083201 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9fzln" podUID="2b596a32-948b-4c02-aa48-1125c0e9b2fb" containerName="registry-server" containerID="cri-o://b2c8d5aec867dc313755b82e5c8dccbb5a09789422d65455ae9133403580811e" gracePeriod=2 Nov 25 16:00:51 crc kubenswrapper[4890]: I1125 16:00:51.608356 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9fzln" Nov 25 16:00:51 crc kubenswrapper[4890]: I1125 16:00:51.728883 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swxcr\" (UniqueName: \"kubernetes.io/projected/2b596a32-948b-4c02-aa48-1125c0e9b2fb-kube-api-access-swxcr\") pod \"2b596a32-948b-4c02-aa48-1125c0e9b2fb\" (UID: \"2b596a32-948b-4c02-aa48-1125c0e9b2fb\") " Nov 25 16:00:51 crc kubenswrapper[4890]: I1125 16:00:51.730879 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b596a32-948b-4c02-aa48-1125c0e9b2fb-catalog-content\") pod \"2b596a32-948b-4c02-aa48-1125c0e9b2fb\" (UID: \"2b596a32-948b-4c02-aa48-1125c0e9b2fb\") " Nov 25 16:00:51 crc kubenswrapper[4890]: I1125 16:00:51.731062 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b596a32-948b-4c02-aa48-1125c0e9b2fb-utilities\") pod \"2b596a32-948b-4c02-aa48-1125c0e9b2fb\" (UID: \"2b596a32-948b-4c02-aa48-1125c0e9b2fb\") " Nov 25 16:00:51 crc kubenswrapper[4890]: I1125 16:00:51.732074 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b596a32-948b-4c02-aa48-1125c0e9b2fb-utilities" (OuterVolumeSpecName: "utilities") pod "2b596a32-948b-4c02-aa48-1125c0e9b2fb" (UID: "2b596a32-948b-4c02-aa48-1125c0e9b2fb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:00:51 crc kubenswrapper[4890]: I1125 16:00:51.736953 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b596a32-948b-4c02-aa48-1125c0e9b2fb-kube-api-access-swxcr" (OuterVolumeSpecName: "kube-api-access-swxcr") pod "2b596a32-948b-4c02-aa48-1125c0e9b2fb" (UID: "2b596a32-948b-4c02-aa48-1125c0e9b2fb"). InnerVolumeSpecName "kube-api-access-swxcr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:00:51 crc kubenswrapper[4890]: I1125 16:00:51.796672 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b596a32-948b-4c02-aa48-1125c0e9b2fb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2b596a32-948b-4c02-aa48-1125c0e9b2fb" (UID: "2b596a32-948b-4c02-aa48-1125c0e9b2fb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:00:51 crc kubenswrapper[4890]: I1125 16:00:51.833312 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b596a32-948b-4c02-aa48-1125c0e9b2fb-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 16:00:51 crc kubenswrapper[4890]: I1125 16:00:51.833357 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b596a32-948b-4c02-aa48-1125c0e9b2fb-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 16:00:51 crc kubenswrapper[4890]: I1125 16:00:51.833371 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swxcr\" (UniqueName: \"kubernetes.io/projected/2b596a32-948b-4c02-aa48-1125c0e9b2fb-kube-api-access-swxcr\") on node \"crc\" DevicePath \"\"" Nov 25 16:00:52 crc kubenswrapper[4890]: I1125 16:00:52.095818 4890 generic.go:334] "Generic (PLEG): container finished" podID="2b596a32-948b-4c02-aa48-1125c0e9b2fb" containerID="b2c8d5aec867dc313755b82e5c8dccbb5a09789422d65455ae9133403580811e" exitCode=0 Nov 25 16:00:52 crc kubenswrapper[4890]: I1125 16:00:52.095884 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9fzln" event={"ID":"2b596a32-948b-4c02-aa48-1125c0e9b2fb","Type":"ContainerDied","Data":"b2c8d5aec867dc313755b82e5c8dccbb5a09789422d65455ae9133403580811e"} Nov 25 16:00:52 crc kubenswrapper[4890]: I1125 16:00:52.095904 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9fzln" Nov 25 16:00:52 crc kubenswrapper[4890]: I1125 16:00:52.095941 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9fzln" event={"ID":"2b596a32-948b-4c02-aa48-1125c0e9b2fb","Type":"ContainerDied","Data":"f4bde364198d5454a0edc207a928f38b4b6aca8c7654c590e5e5ef8ab6a7b72f"} Nov 25 16:00:52 crc kubenswrapper[4890]: I1125 16:00:52.095972 4890 scope.go:117] "RemoveContainer" containerID="b2c8d5aec867dc313755b82e5c8dccbb5a09789422d65455ae9133403580811e" Nov 25 16:00:52 crc kubenswrapper[4890]: I1125 16:00:52.137282 4890 scope.go:117] "RemoveContainer" containerID="c998bbb00494acc38800b7420b74b96c3e0323ca91f67ec96d3c979ec2933f0b" Nov 25 16:00:52 crc kubenswrapper[4890]: I1125 16:00:52.142740 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9fzln"] Nov 25 16:00:52 crc kubenswrapper[4890]: I1125 16:00:52.153770 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9fzln"] Nov 25 16:00:52 crc kubenswrapper[4890]: I1125 16:00:52.165106 4890 scope.go:117] "RemoveContainer" containerID="7d378d902f48b858de4ea20ce87c7cd569ad5f804db666ec276cebe75ed278e2" Nov 25 16:00:52 crc kubenswrapper[4890]: I1125 16:00:52.185408 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b596a32-948b-4c02-aa48-1125c0e9b2fb" path="/var/lib/kubelet/pods/2b596a32-948b-4c02-aa48-1125c0e9b2fb/volumes" Nov 25 16:00:52 crc kubenswrapper[4890]: I1125 16:00:52.213860 4890 scope.go:117] "RemoveContainer" containerID="b2c8d5aec867dc313755b82e5c8dccbb5a09789422d65455ae9133403580811e" Nov 25 16:00:52 crc kubenswrapper[4890]: E1125 16:00:52.214366 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2c8d5aec867dc313755b82e5c8dccbb5a09789422d65455ae9133403580811e\": container with ID starting with b2c8d5aec867dc313755b82e5c8dccbb5a09789422d65455ae9133403580811e not found: ID does not exist" containerID="b2c8d5aec867dc313755b82e5c8dccbb5a09789422d65455ae9133403580811e" Nov 25 16:00:52 crc kubenswrapper[4890]: I1125 16:00:52.214407 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2c8d5aec867dc313755b82e5c8dccbb5a09789422d65455ae9133403580811e"} err="failed to get container status \"b2c8d5aec867dc313755b82e5c8dccbb5a09789422d65455ae9133403580811e\": rpc error: code = NotFound desc = could not find container \"b2c8d5aec867dc313755b82e5c8dccbb5a09789422d65455ae9133403580811e\": container with ID starting with b2c8d5aec867dc313755b82e5c8dccbb5a09789422d65455ae9133403580811e not found: ID does not exist" Nov 25 16:00:52 crc kubenswrapper[4890]: I1125 16:00:52.214439 4890 scope.go:117] "RemoveContainer" containerID="c998bbb00494acc38800b7420b74b96c3e0323ca91f67ec96d3c979ec2933f0b" Nov 25 16:00:52 crc kubenswrapper[4890]: E1125 16:00:52.214889 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c998bbb00494acc38800b7420b74b96c3e0323ca91f67ec96d3c979ec2933f0b\": container with ID starting with c998bbb00494acc38800b7420b74b96c3e0323ca91f67ec96d3c979ec2933f0b not found: ID does not exist" containerID="c998bbb00494acc38800b7420b74b96c3e0323ca91f67ec96d3c979ec2933f0b" Nov 25 16:00:52 crc kubenswrapper[4890]: I1125 16:00:52.214925 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c998bbb00494acc38800b7420b74b96c3e0323ca91f67ec96d3c979ec2933f0b"} err="failed to get container status \"c998bbb00494acc38800b7420b74b96c3e0323ca91f67ec96d3c979ec2933f0b\": rpc error: code = NotFound desc = could not find container \"c998bbb00494acc38800b7420b74b96c3e0323ca91f67ec96d3c979ec2933f0b\": container with ID starting with c998bbb00494acc38800b7420b74b96c3e0323ca91f67ec96d3c979ec2933f0b not found: ID does not exist" Nov 25 16:00:52 crc kubenswrapper[4890]: I1125 16:00:52.214949 4890 scope.go:117] "RemoveContainer" containerID="7d378d902f48b858de4ea20ce87c7cd569ad5f804db666ec276cebe75ed278e2" Nov 25 16:00:52 crc kubenswrapper[4890]: E1125 16:00:52.215279 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d378d902f48b858de4ea20ce87c7cd569ad5f804db666ec276cebe75ed278e2\": container with ID starting with 7d378d902f48b858de4ea20ce87c7cd569ad5f804db666ec276cebe75ed278e2 not found: ID does not exist" containerID="7d378d902f48b858de4ea20ce87c7cd569ad5f804db666ec276cebe75ed278e2" Nov 25 16:00:52 crc kubenswrapper[4890]: I1125 16:00:52.215311 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d378d902f48b858de4ea20ce87c7cd569ad5f804db666ec276cebe75ed278e2"} err="failed to get container status \"7d378d902f48b858de4ea20ce87c7cd569ad5f804db666ec276cebe75ed278e2\": rpc error: code = NotFound desc = could not find container \"7d378d902f48b858de4ea20ce87c7cd569ad5f804db666ec276cebe75ed278e2\": container with ID starting with 7d378d902f48b858de4ea20ce87c7cd569ad5f804db666ec276cebe75ed278e2 not found: ID does not exist" Nov 25 16:01:00 crc kubenswrapper[4890]: I1125 16:01:00.169041 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29401441-rfltq"] Nov 25 16:01:00 crc kubenswrapper[4890]: E1125 16:01:00.170476 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b596a32-948b-4c02-aa48-1125c0e9b2fb" containerName="extract-utilities" Nov 25 16:01:00 crc kubenswrapper[4890]: I1125 16:01:00.170494 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b596a32-948b-4c02-aa48-1125c0e9b2fb" containerName="extract-utilities" Nov 25 16:01:00 crc kubenswrapper[4890]: E1125 16:01:00.170556 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b596a32-948b-4c02-aa48-1125c0e9b2fb" containerName="registry-server" Nov 25 16:01:00 crc kubenswrapper[4890]: I1125 16:01:00.170569 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b596a32-948b-4c02-aa48-1125c0e9b2fb" containerName="registry-server" Nov 25 16:01:00 crc kubenswrapper[4890]: E1125 16:01:00.170587 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b596a32-948b-4c02-aa48-1125c0e9b2fb" containerName="extract-content" Nov 25 16:01:00 crc kubenswrapper[4890]: I1125 16:01:00.170594 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b596a32-948b-4c02-aa48-1125c0e9b2fb" containerName="extract-content" Nov 25 16:01:00 crc kubenswrapper[4890]: I1125 16:01:00.170849 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b596a32-948b-4c02-aa48-1125c0e9b2fb" containerName="registry-server" Nov 25 16:01:00 crc kubenswrapper[4890]: I1125 16:01:00.171826 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29401441-rfltq" Nov 25 16:01:00 crc kubenswrapper[4890]: I1125 16:01:00.198054 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29401441-rfltq"] Nov 25 16:01:00 crc kubenswrapper[4890]: I1125 16:01:00.233627 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f146ff30-8207-40d9-a579-ef8f314c41e4-combined-ca-bundle\") pod \"keystone-cron-29401441-rfltq\" (UID: \"f146ff30-8207-40d9-a579-ef8f314c41e4\") " pod="openstack/keystone-cron-29401441-rfltq" Nov 25 16:01:00 crc kubenswrapper[4890]: I1125 16:01:00.233769 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7shvr\" (UniqueName: \"kubernetes.io/projected/f146ff30-8207-40d9-a579-ef8f314c41e4-kube-api-access-7shvr\") pod \"keystone-cron-29401441-rfltq\" (UID: \"f146ff30-8207-40d9-a579-ef8f314c41e4\") " pod="openstack/keystone-cron-29401441-rfltq" Nov 25 16:01:00 crc kubenswrapper[4890]: I1125 16:01:00.233827 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f146ff30-8207-40d9-a579-ef8f314c41e4-config-data\") pod \"keystone-cron-29401441-rfltq\" (UID: \"f146ff30-8207-40d9-a579-ef8f314c41e4\") " pod="openstack/keystone-cron-29401441-rfltq" Nov 25 16:01:00 crc kubenswrapper[4890]: I1125 16:01:00.233867 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f146ff30-8207-40d9-a579-ef8f314c41e4-fernet-keys\") pod \"keystone-cron-29401441-rfltq\" (UID: \"f146ff30-8207-40d9-a579-ef8f314c41e4\") " pod="openstack/keystone-cron-29401441-rfltq" Nov 25 16:01:00 crc kubenswrapper[4890]: I1125 16:01:00.336359 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7shvr\" (UniqueName: \"kubernetes.io/projected/f146ff30-8207-40d9-a579-ef8f314c41e4-kube-api-access-7shvr\") pod \"keystone-cron-29401441-rfltq\" (UID: \"f146ff30-8207-40d9-a579-ef8f314c41e4\") " pod="openstack/keystone-cron-29401441-rfltq" Nov 25 16:01:00 crc kubenswrapper[4890]: I1125 16:01:00.336561 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f146ff30-8207-40d9-a579-ef8f314c41e4-config-data\") pod \"keystone-cron-29401441-rfltq\" (UID: \"f146ff30-8207-40d9-a579-ef8f314c41e4\") " pod="openstack/keystone-cron-29401441-rfltq" Nov 25 16:01:00 crc kubenswrapper[4890]: I1125 16:01:00.336670 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f146ff30-8207-40d9-a579-ef8f314c41e4-fernet-keys\") pod \"keystone-cron-29401441-rfltq\" (UID: \"f146ff30-8207-40d9-a579-ef8f314c41e4\") " pod="openstack/keystone-cron-29401441-rfltq" Nov 25 16:01:00 crc kubenswrapper[4890]: I1125 16:01:00.337958 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f146ff30-8207-40d9-a579-ef8f314c41e4-combined-ca-bundle\") pod \"keystone-cron-29401441-rfltq\" (UID: \"f146ff30-8207-40d9-a579-ef8f314c41e4\") " pod="openstack/keystone-cron-29401441-rfltq" Nov 25 16:01:00 crc kubenswrapper[4890]: I1125 16:01:00.345495 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f146ff30-8207-40d9-a579-ef8f314c41e4-fernet-keys\") pod \"keystone-cron-29401441-rfltq\" (UID: \"f146ff30-8207-40d9-a579-ef8f314c41e4\") " pod="openstack/keystone-cron-29401441-rfltq" Nov 25 16:01:00 crc kubenswrapper[4890]: I1125 16:01:00.359434 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f146ff30-8207-40d9-a579-ef8f314c41e4-config-data\") pod \"keystone-cron-29401441-rfltq\" (UID: \"f146ff30-8207-40d9-a579-ef8f314c41e4\") " pod="openstack/keystone-cron-29401441-rfltq" Nov 25 16:01:00 crc kubenswrapper[4890]: I1125 16:01:00.359603 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f146ff30-8207-40d9-a579-ef8f314c41e4-combined-ca-bundle\") pod \"keystone-cron-29401441-rfltq\" (UID: \"f146ff30-8207-40d9-a579-ef8f314c41e4\") " pod="openstack/keystone-cron-29401441-rfltq" Nov 25 16:01:00 crc kubenswrapper[4890]: I1125 16:01:00.365292 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7shvr\" (UniqueName: \"kubernetes.io/projected/f146ff30-8207-40d9-a579-ef8f314c41e4-kube-api-access-7shvr\") pod \"keystone-cron-29401441-rfltq\" (UID: \"f146ff30-8207-40d9-a579-ef8f314c41e4\") " pod="openstack/keystone-cron-29401441-rfltq" Nov 25 16:01:00 crc kubenswrapper[4890]: I1125 16:01:00.506975 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29401441-rfltq" Nov 25 16:01:01 crc kubenswrapper[4890]: I1125 16:01:01.007093 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29401441-rfltq"] Nov 25 16:01:01 crc kubenswrapper[4890]: I1125 16:01:01.206634 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29401441-rfltq" event={"ID":"f146ff30-8207-40d9-a579-ef8f314c41e4","Type":"ContainerStarted","Data":"245f07380c5c8ae6b18fff39d7f1dfc3cf1fc9dac5389dc942491860af5d120f"} Nov 25 16:01:02 crc kubenswrapper[4890]: I1125 16:01:02.230049 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29401441-rfltq" event={"ID":"f146ff30-8207-40d9-a579-ef8f314c41e4","Type":"ContainerStarted","Data":"7e6bdacd5cee19270eaae1598269a8fa1d893257a1c1675c5df0af78d3671003"} Nov 25 16:01:02 crc kubenswrapper[4890]: I1125 16:01:02.258563 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29401441-rfltq" podStartSLOduration=2.258543205 podStartE2EDuration="2.258543205s" podCreationTimestamp="2025-11-25 16:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:01:02.250060144 +0000 UTC m=+3520.692522754" watchObservedRunningTime="2025-11-25 16:01:02.258543205 +0000 UTC m=+3520.701005805" Nov 25 16:01:03 crc kubenswrapper[4890]: I1125 16:01:03.316969 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-x7gbs"] Nov 25 16:01:03 crc kubenswrapper[4890]: I1125 16:01:03.319770 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x7gbs" Nov 25 16:01:03 crc kubenswrapper[4890]: I1125 16:01:03.335361 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x7gbs"] Nov 25 16:01:03 crc kubenswrapper[4890]: I1125 16:01:03.422665 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ll2k\" (UniqueName: \"kubernetes.io/projected/74d71c6f-0f1b-4de1-b608-91aa7b420464-kube-api-access-6ll2k\") pod \"redhat-operators-x7gbs\" (UID: \"74d71c6f-0f1b-4de1-b608-91aa7b420464\") " pod="openshift-marketplace/redhat-operators-x7gbs" Nov 25 16:01:03 crc kubenswrapper[4890]: I1125 16:01:03.422769 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74d71c6f-0f1b-4de1-b608-91aa7b420464-utilities\") pod \"redhat-operators-x7gbs\" (UID: \"74d71c6f-0f1b-4de1-b608-91aa7b420464\") " pod="openshift-marketplace/redhat-operators-x7gbs" Nov 25 16:01:03 crc kubenswrapper[4890]: I1125 16:01:03.422795 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74d71c6f-0f1b-4de1-b608-91aa7b420464-catalog-content\") pod \"redhat-operators-x7gbs\" (UID: \"74d71c6f-0f1b-4de1-b608-91aa7b420464\") " pod="openshift-marketplace/redhat-operators-x7gbs" Nov 25 16:01:03 crc kubenswrapper[4890]: I1125 16:01:03.525761 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ll2k\" (UniqueName: \"kubernetes.io/projected/74d71c6f-0f1b-4de1-b608-91aa7b420464-kube-api-access-6ll2k\") pod \"redhat-operators-x7gbs\" (UID: \"74d71c6f-0f1b-4de1-b608-91aa7b420464\") " pod="openshift-marketplace/redhat-operators-x7gbs" Nov 25 16:01:03 crc kubenswrapper[4890]: I1125 16:01:03.525886 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74d71c6f-0f1b-4de1-b608-91aa7b420464-utilities\") pod \"redhat-operators-x7gbs\" (UID: \"74d71c6f-0f1b-4de1-b608-91aa7b420464\") " pod="openshift-marketplace/redhat-operators-x7gbs" Nov 25 16:01:03 crc kubenswrapper[4890]: I1125 16:01:03.525917 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74d71c6f-0f1b-4de1-b608-91aa7b420464-catalog-content\") pod \"redhat-operators-x7gbs\" (UID: \"74d71c6f-0f1b-4de1-b608-91aa7b420464\") " pod="openshift-marketplace/redhat-operators-x7gbs" Nov 25 16:01:03 crc kubenswrapper[4890]: I1125 16:01:03.526720 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74d71c6f-0f1b-4de1-b608-91aa7b420464-catalog-content\") pod \"redhat-operators-x7gbs\" (UID: \"74d71c6f-0f1b-4de1-b608-91aa7b420464\") " pod="openshift-marketplace/redhat-operators-x7gbs" Nov 25 16:01:03 crc kubenswrapper[4890]: I1125 16:01:03.526778 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74d71c6f-0f1b-4de1-b608-91aa7b420464-utilities\") pod \"redhat-operators-x7gbs\" (UID: \"74d71c6f-0f1b-4de1-b608-91aa7b420464\") " pod="openshift-marketplace/redhat-operators-x7gbs" Nov 25 16:01:03 crc kubenswrapper[4890]: I1125 16:01:03.556583 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ll2k\" (UniqueName: \"kubernetes.io/projected/74d71c6f-0f1b-4de1-b608-91aa7b420464-kube-api-access-6ll2k\") pod \"redhat-operators-x7gbs\" (UID: \"74d71c6f-0f1b-4de1-b608-91aa7b420464\") " pod="openshift-marketplace/redhat-operators-x7gbs" Nov 25 16:01:03 crc kubenswrapper[4890]: I1125 16:01:03.661535 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x7gbs" Nov 25 16:01:04 crc kubenswrapper[4890]: I1125 16:01:04.195441 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x7gbs"] Nov 25 16:01:04 crc kubenswrapper[4890]: W1125 16:01:04.202390 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod74d71c6f_0f1b_4de1_b608_91aa7b420464.slice/crio-c0f73d8a0c0233018bfa3b688ddafa6bc1139b07bb1b2cc53e827e5f55dfdcb7 WatchSource:0}: Error finding container c0f73d8a0c0233018bfa3b688ddafa6bc1139b07bb1b2cc53e827e5f55dfdcb7: Status 404 returned error can't find the container with id c0f73d8a0c0233018bfa3b688ddafa6bc1139b07bb1b2cc53e827e5f55dfdcb7 Nov 25 16:01:04 crc kubenswrapper[4890]: I1125 16:01:04.258933 4890 generic.go:334] "Generic (PLEG): container finished" podID="f146ff30-8207-40d9-a579-ef8f314c41e4" containerID="7e6bdacd5cee19270eaae1598269a8fa1d893257a1c1675c5df0af78d3671003" exitCode=0 Nov 25 16:01:04 crc kubenswrapper[4890]: I1125 16:01:04.259004 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29401441-rfltq" event={"ID":"f146ff30-8207-40d9-a579-ef8f314c41e4","Type":"ContainerDied","Data":"7e6bdacd5cee19270eaae1598269a8fa1d893257a1c1675c5df0af78d3671003"} Nov 25 16:01:04 crc kubenswrapper[4890]: I1125 16:01:04.267231 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7gbs" event={"ID":"74d71c6f-0f1b-4de1-b608-91aa7b420464","Type":"ContainerStarted","Data":"c0f73d8a0c0233018bfa3b688ddafa6bc1139b07bb1b2cc53e827e5f55dfdcb7"} Nov 25 16:01:05 crc kubenswrapper[4890]: I1125 16:01:05.282570 4890 generic.go:334] "Generic (PLEG): container finished" podID="74d71c6f-0f1b-4de1-b608-91aa7b420464" containerID="607e132da773d8e4b6fd0e2f304ba09f13c86e617b9f6acf7ab88b13acdc6d64" exitCode=0 Nov 25 16:01:05 crc kubenswrapper[4890]: I1125 16:01:05.283498 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7gbs" event={"ID":"74d71c6f-0f1b-4de1-b608-91aa7b420464","Type":"ContainerDied","Data":"607e132da773d8e4b6fd0e2f304ba09f13c86e617b9f6acf7ab88b13acdc6d64"} Nov 25 16:01:05 crc kubenswrapper[4890]: I1125 16:01:05.737785 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29401441-rfltq" Nov 25 16:01:05 crc kubenswrapper[4890]: I1125 16:01:05.895390 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f146ff30-8207-40d9-a579-ef8f314c41e4-config-data\") pod \"f146ff30-8207-40d9-a579-ef8f314c41e4\" (UID: \"f146ff30-8207-40d9-a579-ef8f314c41e4\") " Nov 25 16:01:05 crc kubenswrapper[4890]: I1125 16:01:05.895914 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f146ff30-8207-40d9-a579-ef8f314c41e4-combined-ca-bundle\") pod \"f146ff30-8207-40d9-a579-ef8f314c41e4\" (UID: \"f146ff30-8207-40d9-a579-ef8f314c41e4\") " Nov 25 16:01:05 crc kubenswrapper[4890]: I1125 16:01:05.896110 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f146ff30-8207-40d9-a579-ef8f314c41e4-fernet-keys\") pod \"f146ff30-8207-40d9-a579-ef8f314c41e4\" (UID: \"f146ff30-8207-40d9-a579-ef8f314c41e4\") " Nov 25 16:01:05 crc kubenswrapper[4890]: I1125 16:01:05.896303 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7shvr\" (UniqueName: \"kubernetes.io/projected/f146ff30-8207-40d9-a579-ef8f314c41e4-kube-api-access-7shvr\") pod \"f146ff30-8207-40d9-a579-ef8f314c41e4\" (UID: \"f146ff30-8207-40d9-a579-ef8f314c41e4\") " Nov 25 16:01:05 crc kubenswrapper[4890]: I1125 16:01:05.903683 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f146ff30-8207-40d9-a579-ef8f314c41e4-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "f146ff30-8207-40d9-a579-ef8f314c41e4" (UID: "f146ff30-8207-40d9-a579-ef8f314c41e4"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:01:05 crc kubenswrapper[4890]: I1125 16:01:05.903706 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f146ff30-8207-40d9-a579-ef8f314c41e4-kube-api-access-7shvr" (OuterVolumeSpecName: "kube-api-access-7shvr") pod "f146ff30-8207-40d9-a579-ef8f314c41e4" (UID: "f146ff30-8207-40d9-a579-ef8f314c41e4"). InnerVolumeSpecName "kube-api-access-7shvr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:01:05 crc kubenswrapper[4890]: I1125 16:01:05.934859 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f146ff30-8207-40d9-a579-ef8f314c41e4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f146ff30-8207-40d9-a579-ef8f314c41e4" (UID: "f146ff30-8207-40d9-a579-ef8f314c41e4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:01:05 crc kubenswrapper[4890]: I1125 16:01:05.966582 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f146ff30-8207-40d9-a579-ef8f314c41e4-config-data" (OuterVolumeSpecName: "config-data") pod "f146ff30-8207-40d9-a579-ef8f314c41e4" (UID: "f146ff30-8207-40d9-a579-ef8f314c41e4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:01:05 crc kubenswrapper[4890]: I1125 16:01:05.999509 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7shvr\" (UniqueName: \"kubernetes.io/projected/f146ff30-8207-40d9-a579-ef8f314c41e4-kube-api-access-7shvr\") on node \"crc\" DevicePath \"\"" Nov 25 16:01:05 crc kubenswrapper[4890]: I1125 16:01:05.999559 4890 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f146ff30-8207-40d9-a579-ef8f314c41e4-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 16:01:05 crc kubenswrapper[4890]: I1125 16:01:05.999572 4890 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f146ff30-8207-40d9-a579-ef8f314c41e4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 16:01:05 crc kubenswrapper[4890]: I1125 16:01:05.999580 4890 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f146ff30-8207-40d9-a579-ef8f314c41e4-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 25 16:01:06 crc kubenswrapper[4890]: I1125 16:01:06.299442 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29401441-rfltq" event={"ID":"f146ff30-8207-40d9-a579-ef8f314c41e4","Type":"ContainerDied","Data":"245f07380c5c8ae6b18fff39d7f1dfc3cf1fc9dac5389dc942491860af5d120f"} Nov 25 16:01:06 crc kubenswrapper[4890]: I1125 16:01:06.300992 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="245f07380c5c8ae6b18fff39d7f1dfc3cf1fc9dac5389dc942491860af5d120f" Nov 25 16:01:06 crc kubenswrapper[4890]: I1125 16:01:06.301208 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29401441-rfltq" Nov 25 16:01:07 crc kubenswrapper[4890]: I1125 16:01:07.311571 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7gbs" event={"ID":"74d71c6f-0f1b-4de1-b608-91aa7b420464","Type":"ContainerStarted","Data":"923102e60c883750a0fd0768a5819b8749732c1a694d8d6d837fe3cc4e45b1fe"} Nov 25 16:01:22 crc kubenswrapper[4890]: I1125 16:01:22.466593 4890 generic.go:334] "Generic (PLEG): container finished" podID="74d71c6f-0f1b-4de1-b608-91aa7b420464" containerID="923102e60c883750a0fd0768a5819b8749732c1a694d8d6d837fe3cc4e45b1fe" exitCode=0 Nov 25 16:01:22 crc kubenswrapper[4890]: I1125 16:01:22.466687 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7gbs" event={"ID":"74d71c6f-0f1b-4de1-b608-91aa7b420464","Type":"ContainerDied","Data":"923102e60c883750a0fd0768a5819b8749732c1a694d8d6d837fe3cc4e45b1fe"} Nov 25 16:01:24 crc kubenswrapper[4890]: I1125 16:01:24.490403 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7gbs" event={"ID":"74d71c6f-0f1b-4de1-b608-91aa7b420464","Type":"ContainerStarted","Data":"169a576c8eaf5c3f4b8c263b45394fef55f286d612e2340e213bd6e0e05bacff"} Nov 25 16:01:25 crc kubenswrapper[4890]: I1125 16:01:25.528597 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-x7gbs" podStartSLOduration=3.590504417 podStartE2EDuration="22.528572049s" podCreationTimestamp="2025-11-25 16:01:03 +0000 UTC" firstStartedPulling="2025-11-25 16:01:05.285352559 +0000 UTC m=+3523.727815169" lastFinishedPulling="2025-11-25 16:01:24.223420191 +0000 UTC m=+3542.665882801" observedRunningTime="2025-11-25 16:01:25.524314064 +0000 UTC m=+3543.966776684" watchObservedRunningTime="2025-11-25 16:01:25.528572049 +0000 UTC m=+3543.971034659" Nov 25 16:01:33 crc kubenswrapper[4890]: I1125 16:01:33.663278 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-x7gbs" Nov 25 16:01:33 crc kubenswrapper[4890]: I1125 16:01:33.664278 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-x7gbs" Nov 25 16:01:33 crc kubenswrapper[4890]: I1125 16:01:33.723561 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-x7gbs" Nov 25 16:01:34 crc kubenswrapper[4890]: I1125 16:01:34.652068 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-x7gbs" Nov 25 16:01:34 crc kubenswrapper[4890]: I1125 16:01:34.710038 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x7gbs"] Nov 25 16:01:36 crc kubenswrapper[4890]: I1125 16:01:36.609403 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-x7gbs" podUID="74d71c6f-0f1b-4de1-b608-91aa7b420464" containerName="registry-server" containerID="cri-o://169a576c8eaf5c3f4b8c263b45394fef55f286d612e2340e213bd6e0e05bacff" gracePeriod=2 Nov 25 16:01:37 crc kubenswrapper[4890]: I1125 16:01:37.236266 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x7gbs" Nov 25 16:01:37 crc kubenswrapper[4890]: I1125 16:01:37.302987 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74d71c6f-0f1b-4de1-b608-91aa7b420464-catalog-content\") pod \"74d71c6f-0f1b-4de1-b608-91aa7b420464\" (UID: \"74d71c6f-0f1b-4de1-b608-91aa7b420464\") " Nov 25 16:01:37 crc kubenswrapper[4890]: I1125 16:01:37.303053 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ll2k\" (UniqueName: \"kubernetes.io/projected/74d71c6f-0f1b-4de1-b608-91aa7b420464-kube-api-access-6ll2k\") pod \"74d71c6f-0f1b-4de1-b608-91aa7b420464\" (UID: \"74d71c6f-0f1b-4de1-b608-91aa7b420464\") " Nov 25 16:01:37 crc kubenswrapper[4890]: I1125 16:01:37.303152 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74d71c6f-0f1b-4de1-b608-91aa7b420464-utilities\") pod \"74d71c6f-0f1b-4de1-b608-91aa7b420464\" (UID: \"74d71c6f-0f1b-4de1-b608-91aa7b420464\") " Nov 25 16:01:37 crc kubenswrapper[4890]: I1125 16:01:37.304259 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74d71c6f-0f1b-4de1-b608-91aa7b420464-utilities" (OuterVolumeSpecName: "utilities") pod "74d71c6f-0f1b-4de1-b608-91aa7b420464" (UID: "74d71c6f-0f1b-4de1-b608-91aa7b420464"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:01:37 crc kubenswrapper[4890]: I1125 16:01:37.311797 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74d71c6f-0f1b-4de1-b608-91aa7b420464-kube-api-access-6ll2k" (OuterVolumeSpecName: "kube-api-access-6ll2k") pod "74d71c6f-0f1b-4de1-b608-91aa7b420464" (UID: "74d71c6f-0f1b-4de1-b608-91aa7b420464"). InnerVolumeSpecName "kube-api-access-6ll2k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:01:37 crc kubenswrapper[4890]: I1125 16:01:37.406566 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ll2k\" (UniqueName: \"kubernetes.io/projected/74d71c6f-0f1b-4de1-b608-91aa7b420464-kube-api-access-6ll2k\") on node \"crc\" DevicePath \"\"" Nov 25 16:01:37 crc kubenswrapper[4890]: I1125 16:01:37.406603 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74d71c6f-0f1b-4de1-b608-91aa7b420464-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 16:01:37 crc kubenswrapper[4890]: I1125 16:01:37.408157 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74d71c6f-0f1b-4de1-b608-91aa7b420464-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "74d71c6f-0f1b-4de1-b608-91aa7b420464" (UID: "74d71c6f-0f1b-4de1-b608-91aa7b420464"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:01:37 crc kubenswrapper[4890]: I1125 16:01:37.508341 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74d71c6f-0f1b-4de1-b608-91aa7b420464-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 16:01:37 crc kubenswrapper[4890]: I1125 16:01:37.623400 4890 generic.go:334] "Generic (PLEG): container finished" podID="74d71c6f-0f1b-4de1-b608-91aa7b420464" containerID="169a576c8eaf5c3f4b8c263b45394fef55f286d612e2340e213bd6e0e05bacff" exitCode=0 Nov 25 16:01:37 crc kubenswrapper[4890]: I1125 16:01:37.623474 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7gbs" event={"ID":"74d71c6f-0f1b-4de1-b608-91aa7b420464","Type":"ContainerDied","Data":"169a576c8eaf5c3f4b8c263b45394fef55f286d612e2340e213bd6e0e05bacff"} Nov 25 16:01:37 crc kubenswrapper[4890]: I1125 16:01:37.623501 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x7gbs" Nov 25 16:01:37 crc kubenswrapper[4890]: I1125 16:01:37.623524 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7gbs" event={"ID":"74d71c6f-0f1b-4de1-b608-91aa7b420464","Type":"ContainerDied","Data":"c0f73d8a0c0233018bfa3b688ddafa6bc1139b07bb1b2cc53e827e5f55dfdcb7"} Nov 25 16:01:37 crc kubenswrapper[4890]: I1125 16:01:37.623549 4890 scope.go:117] "RemoveContainer" containerID="169a576c8eaf5c3f4b8c263b45394fef55f286d612e2340e213bd6e0e05bacff" Nov 25 16:01:37 crc kubenswrapper[4890]: I1125 16:01:37.651371 4890 scope.go:117] "RemoveContainer" containerID="923102e60c883750a0fd0768a5819b8749732c1a694d8d6d837fe3cc4e45b1fe" Nov 25 16:01:37 crc kubenswrapper[4890]: I1125 16:01:37.670065 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x7gbs"] Nov 25 16:01:37 crc kubenswrapper[4890]: I1125 16:01:37.683558 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-x7gbs"] Nov 25 16:01:37 crc kubenswrapper[4890]: I1125 16:01:37.690486 4890 scope.go:117] "RemoveContainer" containerID="607e132da773d8e4b6fd0e2f304ba09f13c86e617b9f6acf7ab88b13acdc6d64" Nov 25 16:01:37 crc kubenswrapper[4890]: I1125 16:01:37.737194 4890 scope.go:117] "RemoveContainer" containerID="169a576c8eaf5c3f4b8c263b45394fef55f286d612e2340e213bd6e0e05bacff" Nov 25 16:01:37 crc kubenswrapper[4890]: E1125 16:01:37.737821 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"169a576c8eaf5c3f4b8c263b45394fef55f286d612e2340e213bd6e0e05bacff\": container with ID starting with 169a576c8eaf5c3f4b8c263b45394fef55f286d612e2340e213bd6e0e05bacff not found: ID does not exist" containerID="169a576c8eaf5c3f4b8c263b45394fef55f286d612e2340e213bd6e0e05bacff" Nov 25 16:01:37 crc kubenswrapper[4890]: I1125 16:01:37.737880 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"169a576c8eaf5c3f4b8c263b45394fef55f286d612e2340e213bd6e0e05bacff"} err="failed to get container status \"169a576c8eaf5c3f4b8c263b45394fef55f286d612e2340e213bd6e0e05bacff\": rpc error: code = NotFound desc = could not find container \"169a576c8eaf5c3f4b8c263b45394fef55f286d612e2340e213bd6e0e05bacff\": container with ID starting with 169a576c8eaf5c3f4b8c263b45394fef55f286d612e2340e213bd6e0e05bacff not found: ID does not exist" Nov 25 16:01:37 crc kubenswrapper[4890]: I1125 16:01:37.737918 4890 scope.go:117] "RemoveContainer" containerID="923102e60c883750a0fd0768a5819b8749732c1a694d8d6d837fe3cc4e45b1fe" Nov 25 16:01:37 crc kubenswrapper[4890]: E1125 16:01:37.738406 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"923102e60c883750a0fd0768a5819b8749732c1a694d8d6d837fe3cc4e45b1fe\": container with ID starting with 923102e60c883750a0fd0768a5819b8749732c1a694d8d6d837fe3cc4e45b1fe not found: ID does not exist" containerID="923102e60c883750a0fd0768a5819b8749732c1a694d8d6d837fe3cc4e45b1fe" Nov 25 16:01:37 crc kubenswrapper[4890]: I1125 16:01:37.738446 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"923102e60c883750a0fd0768a5819b8749732c1a694d8d6d837fe3cc4e45b1fe"} err="failed to get container status \"923102e60c883750a0fd0768a5819b8749732c1a694d8d6d837fe3cc4e45b1fe\": rpc error: code = NotFound desc = could not find container \"923102e60c883750a0fd0768a5819b8749732c1a694d8d6d837fe3cc4e45b1fe\": container with ID starting with 923102e60c883750a0fd0768a5819b8749732c1a694d8d6d837fe3cc4e45b1fe not found: ID does not exist" Nov 25 16:01:37 crc kubenswrapper[4890]: I1125 16:01:37.738484 4890 scope.go:117] "RemoveContainer" containerID="607e132da773d8e4b6fd0e2f304ba09f13c86e617b9f6acf7ab88b13acdc6d64" Nov 25 16:01:37 crc kubenswrapper[4890]: E1125 16:01:37.738773 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"607e132da773d8e4b6fd0e2f304ba09f13c86e617b9f6acf7ab88b13acdc6d64\": container with ID starting with 607e132da773d8e4b6fd0e2f304ba09f13c86e617b9f6acf7ab88b13acdc6d64 not found: ID does not exist" containerID="607e132da773d8e4b6fd0e2f304ba09f13c86e617b9f6acf7ab88b13acdc6d64" Nov 25 16:01:37 crc kubenswrapper[4890]: I1125 16:01:37.738798 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"607e132da773d8e4b6fd0e2f304ba09f13c86e617b9f6acf7ab88b13acdc6d64"} err="failed to get container status \"607e132da773d8e4b6fd0e2f304ba09f13c86e617b9f6acf7ab88b13acdc6d64\": rpc error: code = NotFound desc = could not find container \"607e132da773d8e4b6fd0e2f304ba09f13c86e617b9f6acf7ab88b13acdc6d64\": container with ID starting with 607e132da773d8e4b6fd0e2f304ba09f13c86e617b9f6acf7ab88b13acdc6d64 not found: ID does not exist" Nov 25 16:01:38 crc kubenswrapper[4890]: I1125 16:01:38.186042 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74d71c6f-0f1b-4de1-b608-91aa7b420464" path="/var/lib/kubelet/pods/74d71c6f-0f1b-4de1-b608-91aa7b420464/volumes" Nov 25 16:01:56 crc kubenswrapper[4890]: I1125 16:01:56.447692 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 16:01:56 crc kubenswrapper[4890]: I1125 16:01:56.448607 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 16:02:26 crc kubenswrapper[4890]: I1125 16:02:26.447987 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 16:02:26 crc kubenswrapper[4890]: I1125 16:02:26.448783 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 16:02:56 crc kubenswrapper[4890]: I1125 16:02:56.447975 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 16:02:56 crc kubenswrapper[4890]: I1125 16:02:56.448489 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 16:02:56 crc kubenswrapper[4890]: I1125 16:02:56.448560 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" Nov 25 16:02:56 crc kubenswrapper[4890]: I1125 16:02:56.449426 4890 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a0bac43b712e038c240449718f3b53c0d88f1d20e0f5e92b7dfb0385a9de4f68"} pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 16:02:56 crc kubenswrapper[4890]: I1125 16:02:56.449482 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" containerID="cri-o://a0bac43b712e038c240449718f3b53c0d88f1d20e0f5e92b7dfb0385a9de4f68" gracePeriod=600 Nov 25 16:02:57 crc kubenswrapper[4890]: E1125 16:02:57.379039 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 16:02:57 crc kubenswrapper[4890]: I1125 16:02:57.484568 4890 generic.go:334] "Generic (PLEG): container finished" podID="4e4f849d-f239-4727-a73e-18327856929a" containerID="a0bac43b712e038c240449718f3b53c0d88f1d20e0f5e92b7dfb0385a9de4f68" exitCode=0 Nov 25 16:02:57 crc kubenswrapper[4890]: I1125 16:02:57.484625 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" event={"ID":"4e4f849d-f239-4727-a73e-18327856929a","Type":"ContainerDied","Data":"a0bac43b712e038c240449718f3b53c0d88f1d20e0f5e92b7dfb0385a9de4f68"} Nov 25 16:02:57 crc kubenswrapper[4890]: I1125 16:02:57.484671 4890 scope.go:117] "RemoveContainer" containerID="54ca4ab07f0e1f73cd740dc4474b2b043d2b98ae0c3d04e251edd62a255d4192" Nov 25 16:02:57 crc kubenswrapper[4890]: I1125 16:02:57.485671 4890 scope.go:117] "RemoveContainer" containerID="a0bac43b712e038c240449718f3b53c0d88f1d20e0f5e92b7dfb0385a9de4f68" Nov 25 16:02:57 crc kubenswrapper[4890]: E1125 16:02:57.486042 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 16:03:11 crc kubenswrapper[4890]: I1125 16:03:11.177104 4890 scope.go:117] "RemoveContainer" containerID="a0bac43b712e038c240449718f3b53c0d88f1d20e0f5e92b7dfb0385a9de4f68" Nov 25 16:03:11 crc kubenswrapper[4890]: E1125 16:03:11.178236 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 16:03:22 crc kubenswrapper[4890]: I1125 16:03:22.183215 4890 scope.go:117] "RemoveContainer" containerID="a0bac43b712e038c240449718f3b53c0d88f1d20e0f5e92b7dfb0385a9de4f68" Nov 25 16:03:22 crc kubenswrapper[4890]: E1125 16:03:22.184084 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 16:03:34 crc kubenswrapper[4890]: I1125 16:03:34.173660 4890 scope.go:117] "RemoveContainer" containerID="a0bac43b712e038c240449718f3b53c0d88f1d20e0f5e92b7dfb0385a9de4f68" Nov 25 16:03:34 crc kubenswrapper[4890]: E1125 16:03:34.174966 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 16:03:46 crc kubenswrapper[4890]: I1125 16:03:46.173447 4890 scope.go:117] "RemoveContainer" containerID="a0bac43b712e038c240449718f3b53c0d88f1d20e0f5e92b7dfb0385a9de4f68" Nov 25 16:03:46 crc kubenswrapper[4890]: E1125 16:03:46.174406 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 16:03:58 crc kubenswrapper[4890]: I1125 16:03:58.173903 4890 scope.go:117] "RemoveContainer" containerID="a0bac43b712e038c240449718f3b53c0d88f1d20e0f5e92b7dfb0385a9de4f68" Nov 25 16:03:58 crc kubenswrapper[4890]: E1125 16:03:58.175171 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 16:04:13 crc kubenswrapper[4890]: I1125 16:04:13.173885 4890 scope.go:117] "RemoveContainer" containerID="a0bac43b712e038c240449718f3b53c0d88f1d20e0f5e92b7dfb0385a9de4f68" Nov 25 16:04:13 crc kubenswrapper[4890]: E1125 16:04:13.174724 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 16:04:27 crc kubenswrapper[4890]: I1125 16:04:27.174342 4890 scope.go:117] "RemoveContainer" containerID="a0bac43b712e038c240449718f3b53c0d88f1d20e0f5e92b7dfb0385a9de4f68" Nov 25 16:04:27 crc kubenswrapper[4890]: E1125 16:04:27.175198 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 16:04:42 crc kubenswrapper[4890]: I1125 16:04:42.182288 4890 scope.go:117] "RemoveContainer" containerID="a0bac43b712e038c240449718f3b53c0d88f1d20e0f5e92b7dfb0385a9de4f68" Nov 25 16:04:42 crc kubenswrapper[4890]: E1125 16:04:42.183350 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 16:04:54 crc kubenswrapper[4890]: I1125 16:04:54.174029 4890 scope.go:117] "RemoveContainer" containerID="a0bac43b712e038c240449718f3b53c0d88f1d20e0f5e92b7dfb0385a9de4f68" Nov 25 16:04:54 crc kubenswrapper[4890]: E1125 16:04:54.174989 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 16:05:08 crc kubenswrapper[4890]: I1125 16:05:08.174066 4890 scope.go:117] "RemoveContainer" containerID="a0bac43b712e038c240449718f3b53c0d88f1d20e0f5e92b7dfb0385a9de4f68" Nov 25 16:05:08 crc kubenswrapper[4890]: E1125 16:05:08.175004 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 16:05:19 crc kubenswrapper[4890]: I1125 16:05:19.173229 4890 scope.go:117] "RemoveContainer" containerID="a0bac43b712e038c240449718f3b53c0d88f1d20e0f5e92b7dfb0385a9de4f68" Nov 25 16:05:19 crc kubenswrapper[4890]: E1125 16:05:19.173958 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 16:05:32 crc kubenswrapper[4890]: I1125 16:05:32.182017 4890 scope.go:117] "RemoveContainer" containerID="a0bac43b712e038c240449718f3b53c0d88f1d20e0f5e92b7dfb0385a9de4f68" Nov 25 16:05:32 crc kubenswrapper[4890]: E1125 16:05:32.182854 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 16:05:43 crc kubenswrapper[4890]: I1125 16:05:43.173671 4890 scope.go:117] "RemoveContainer" containerID="a0bac43b712e038c240449718f3b53c0d88f1d20e0f5e92b7dfb0385a9de4f68" Nov 25 16:05:43 crc kubenswrapper[4890]: E1125 16:05:43.174625 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 16:05:55 crc kubenswrapper[4890]: I1125 16:05:55.173786 4890 scope.go:117] "RemoveContainer" containerID="a0bac43b712e038c240449718f3b53c0d88f1d20e0f5e92b7dfb0385a9de4f68" Nov 25 16:05:55 crc kubenswrapper[4890]: E1125 16:05:55.174592 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 16:06:10 crc kubenswrapper[4890]: I1125 16:06:10.174474 4890 scope.go:117] "RemoveContainer" containerID="a0bac43b712e038c240449718f3b53c0d88f1d20e0f5e92b7dfb0385a9de4f68" Nov 25 16:06:10 crc kubenswrapper[4890]: E1125 16:06:10.175339 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 16:06:21 crc kubenswrapper[4890]: I1125 16:06:21.173952 4890 scope.go:117] "RemoveContainer" containerID="a0bac43b712e038c240449718f3b53c0d88f1d20e0f5e92b7dfb0385a9de4f68" Nov 25 16:06:21 crc kubenswrapper[4890]: E1125 16:06:21.174750 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 16:06:31 crc kubenswrapper[4890]: I1125 16:06:31.647230 4890 generic.go:334] "Generic (PLEG): container finished" podID="8b652448-2962-42f9-b0b5-e551b50edc1e" containerID="ba77aee782ce8f9ea592a0a8f000c7a382df11999dec90fd8e463c93097cbf2b" exitCode=0 Nov 25 16:06:31 crc kubenswrapper[4890]: I1125 16:06:31.647321 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"8b652448-2962-42f9-b0b5-e551b50edc1e","Type":"ContainerDied","Data":"ba77aee782ce8f9ea592a0a8f000c7a382df11999dec90fd8e463c93097cbf2b"} Nov 25 16:06:33 crc kubenswrapper[4890]: I1125 16:06:33.077325 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 25 16:06:33 crc kubenswrapper[4890]: I1125 16:06:33.242797 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8b652448-2962-42f9-b0b5-e551b50edc1e-openstack-config\") pod \"8b652448-2962-42f9-b0b5-e551b50edc1e\" (UID: \"8b652448-2962-42f9-b0b5-e551b50edc1e\") " Nov 25 16:06:33 crc kubenswrapper[4890]: I1125 16:06:33.242915 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8b652448-2962-42f9-b0b5-e551b50edc1e-ssh-key\") pod \"8b652448-2962-42f9-b0b5-e551b50edc1e\" (UID: \"8b652448-2962-42f9-b0b5-e551b50edc1e\") " Nov 25 16:06:33 crc kubenswrapper[4890]: I1125 16:06:33.242993 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/8b652448-2962-42f9-b0b5-e551b50edc1e-ca-certs\") pod \"8b652448-2962-42f9-b0b5-e551b50edc1e\" (UID: \"8b652448-2962-42f9-b0b5-e551b50edc1e\") " Nov 25 16:06:33 crc kubenswrapper[4890]: I1125 16:06:33.243093 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"8b652448-2962-42f9-b0b5-e551b50edc1e\" (UID: \"8b652448-2962-42f9-b0b5-e551b50edc1e\") " Nov 25 16:06:33 crc kubenswrapper[4890]: I1125 16:06:33.243384 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgmvt\" (UniqueName: \"kubernetes.io/projected/8b652448-2962-42f9-b0b5-e551b50edc1e-kube-api-access-fgmvt\") pod \"8b652448-2962-42f9-b0b5-e551b50edc1e\" (UID: \"8b652448-2962-42f9-b0b5-e551b50edc1e\") " Nov 25 16:06:33 crc kubenswrapper[4890]: I1125 16:06:33.243434 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/8b652448-2962-42f9-b0b5-e551b50edc1e-test-operator-ephemeral-workdir\") pod \"8b652448-2962-42f9-b0b5-e551b50edc1e\" (UID: \"8b652448-2962-42f9-b0b5-e551b50edc1e\") " Nov 25 16:06:33 crc kubenswrapper[4890]: I1125 16:06:33.243477 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8b652448-2962-42f9-b0b5-e551b50edc1e-config-data\") pod \"8b652448-2962-42f9-b0b5-e551b50edc1e\" (UID: \"8b652448-2962-42f9-b0b5-e551b50edc1e\") " Nov 25 16:06:33 crc kubenswrapper[4890]: I1125 16:06:33.243571 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8b652448-2962-42f9-b0b5-e551b50edc1e-openstack-config-secret\") pod \"8b652448-2962-42f9-b0b5-e551b50edc1e\" (UID: \"8b652448-2962-42f9-b0b5-e551b50edc1e\") " Nov 25 16:06:33 crc kubenswrapper[4890]: I1125 16:06:33.243627 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/8b652448-2962-42f9-b0b5-e551b50edc1e-test-operator-ephemeral-temporary\") pod \"8b652448-2962-42f9-b0b5-e551b50edc1e\" (UID: \"8b652448-2962-42f9-b0b5-e551b50edc1e\") " Nov 25 16:06:33 crc kubenswrapper[4890]: I1125 16:06:33.244420 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b652448-2962-42f9-b0b5-e551b50edc1e-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "8b652448-2962-42f9-b0b5-e551b50edc1e" (UID: "8b652448-2962-42f9-b0b5-e551b50edc1e"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:06:33 crc kubenswrapper[4890]: I1125 16:06:33.244734 4890 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/8b652448-2962-42f9-b0b5-e551b50edc1e-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Nov 25 16:06:33 crc kubenswrapper[4890]: I1125 16:06:33.244726 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b652448-2962-42f9-b0b5-e551b50edc1e-config-data" (OuterVolumeSpecName: "config-data") pod "8b652448-2962-42f9-b0b5-e551b50edc1e" (UID: "8b652448-2962-42f9-b0b5-e551b50edc1e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:06:33 crc kubenswrapper[4890]: I1125 16:06:33.256497 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b652448-2962-42f9-b0b5-e551b50edc1e-kube-api-access-fgmvt" (OuterVolumeSpecName: "kube-api-access-fgmvt") pod "8b652448-2962-42f9-b0b5-e551b50edc1e" (UID: "8b652448-2962-42f9-b0b5-e551b50edc1e"). InnerVolumeSpecName "kube-api-access-fgmvt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:06:33 crc kubenswrapper[4890]: I1125 16:06:33.257697 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b652448-2962-42f9-b0b5-e551b50edc1e-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "8b652448-2962-42f9-b0b5-e551b50edc1e" (UID: "8b652448-2962-42f9-b0b5-e551b50edc1e"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:06:33 crc kubenswrapper[4890]: I1125 16:06:33.262119 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "test-operator-logs") pod "8b652448-2962-42f9-b0b5-e551b50edc1e" (UID: "8b652448-2962-42f9-b0b5-e551b50edc1e"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 16:06:33 crc kubenswrapper[4890]: I1125 16:06:33.323298 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b652448-2962-42f9-b0b5-e551b50edc1e-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "8b652448-2962-42f9-b0b5-e551b50edc1e" (UID: "8b652448-2962-42f9-b0b5-e551b50edc1e"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:06:33 crc kubenswrapper[4890]: I1125 16:06:33.346879 4890 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Nov 25 16:06:33 crc kubenswrapper[4890]: I1125 16:06:33.346919 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fgmvt\" (UniqueName: \"kubernetes.io/projected/8b652448-2962-42f9-b0b5-e551b50edc1e-kube-api-access-fgmvt\") on node \"crc\" DevicePath \"\"" Nov 25 16:06:33 crc kubenswrapper[4890]: I1125 16:06:33.346930 4890 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/8b652448-2962-42f9-b0b5-e551b50edc1e-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Nov 25 16:06:33 crc kubenswrapper[4890]: I1125 16:06:33.346938 4890 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8b652448-2962-42f9-b0b5-e551b50edc1e-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 16:06:33 crc kubenswrapper[4890]: I1125 16:06:33.346948 4890 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8b652448-2962-42f9-b0b5-e551b50edc1e-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 25 16:06:33 crc kubenswrapper[4890]: I1125 16:06:33.426997 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b652448-2962-42f9-b0b5-e551b50edc1e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8b652448-2962-42f9-b0b5-e551b50edc1e" (UID: "8b652448-2962-42f9-b0b5-e551b50edc1e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:06:33 crc kubenswrapper[4890]: I1125 16:06:33.450754 4890 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8b652448-2962-42f9-b0b5-e551b50edc1e-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 16:06:33 crc kubenswrapper[4890]: I1125 16:06:33.454809 4890 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Nov 25 16:06:33 crc kubenswrapper[4890]: I1125 16:06:33.458705 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b652448-2962-42f9-b0b5-e551b50edc1e-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "8b652448-2962-42f9-b0b5-e551b50edc1e" (UID: "8b652448-2962-42f9-b0b5-e551b50edc1e"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:06:33 crc kubenswrapper[4890]: I1125 16:06:33.459217 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b652448-2962-42f9-b0b5-e551b50edc1e-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "8b652448-2962-42f9-b0b5-e551b50edc1e" (UID: "8b652448-2962-42f9-b0b5-e551b50edc1e"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:06:33 crc kubenswrapper[4890]: I1125 16:06:33.552347 4890 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8b652448-2962-42f9-b0b5-e551b50edc1e-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 25 16:06:33 crc kubenswrapper[4890]: I1125 16:06:33.552375 4890 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/8b652448-2962-42f9-b0b5-e551b50edc1e-ca-certs\") on node \"crc\" DevicePath \"\"" Nov 25 16:06:33 crc kubenswrapper[4890]: I1125 16:06:33.552385 4890 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Nov 25 16:06:33 crc kubenswrapper[4890]: I1125 16:06:33.673295 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"8b652448-2962-42f9-b0b5-e551b50edc1e","Type":"ContainerDied","Data":"10fafdacefa12996d94beffa374682d9bccca4cd4e5bc1dca33d1f626c9893e5"} Nov 25 16:06:33 crc kubenswrapper[4890]: I1125 16:06:33.673346 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="10fafdacefa12996d94beffa374682d9bccca4cd4e5bc1dca33d1f626c9893e5" Nov 25 16:06:33 crc kubenswrapper[4890]: I1125 16:06:33.673460 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 25 16:06:36 crc kubenswrapper[4890]: I1125 16:06:36.173316 4890 scope.go:117] "RemoveContainer" containerID="a0bac43b712e038c240449718f3b53c0d88f1d20e0f5e92b7dfb0385a9de4f68" Nov 25 16:06:36 crc kubenswrapper[4890]: E1125 16:06:36.174260 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 16:06:37 crc kubenswrapper[4890]: I1125 16:06:37.155298 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 25 16:06:37 crc kubenswrapper[4890]: E1125 16:06:37.156036 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74d71c6f-0f1b-4de1-b608-91aa7b420464" containerName="extract-utilities" Nov 25 16:06:37 crc kubenswrapper[4890]: I1125 16:06:37.156178 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="74d71c6f-0f1b-4de1-b608-91aa7b420464" containerName="extract-utilities" Nov 25 16:06:37 crc kubenswrapper[4890]: E1125 16:06:37.156277 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b652448-2962-42f9-b0b5-e551b50edc1e" containerName="tempest-tests-tempest-tests-runner" Nov 25 16:06:37 crc kubenswrapper[4890]: I1125 16:06:37.156337 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b652448-2962-42f9-b0b5-e551b50edc1e" containerName="tempest-tests-tempest-tests-runner" Nov 25 16:06:37 crc kubenswrapper[4890]: E1125 16:06:37.156439 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f146ff30-8207-40d9-a579-ef8f314c41e4" containerName="keystone-cron" Nov 25 16:06:37 crc kubenswrapper[4890]: I1125 16:06:37.156498 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="f146ff30-8207-40d9-a579-ef8f314c41e4" containerName="keystone-cron" Nov 25 16:06:37 crc kubenswrapper[4890]: E1125 16:06:37.156564 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74d71c6f-0f1b-4de1-b608-91aa7b420464" containerName="extract-content" Nov 25 16:06:37 crc kubenswrapper[4890]: I1125 16:06:37.156626 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="74d71c6f-0f1b-4de1-b608-91aa7b420464" containerName="extract-content" Nov 25 16:06:37 crc kubenswrapper[4890]: E1125 16:06:37.156743 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74d71c6f-0f1b-4de1-b608-91aa7b420464" containerName="registry-server" Nov 25 16:06:37 crc kubenswrapper[4890]: I1125 16:06:37.156839 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="74d71c6f-0f1b-4de1-b608-91aa7b420464" containerName="registry-server" Nov 25 16:06:37 crc kubenswrapper[4890]: I1125 16:06:37.157120 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="74d71c6f-0f1b-4de1-b608-91aa7b420464" containerName="registry-server" Nov 25 16:06:37 crc kubenswrapper[4890]: I1125 16:06:37.157269 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b652448-2962-42f9-b0b5-e551b50edc1e" containerName="tempest-tests-tempest-tests-runner" Nov 25 16:06:37 crc kubenswrapper[4890]: I1125 16:06:37.157359 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="f146ff30-8207-40d9-a579-ef8f314c41e4" containerName="keystone-cron" Nov 25 16:06:37 crc kubenswrapper[4890]: I1125 16:06:37.158388 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 25 16:06:37 crc kubenswrapper[4890]: I1125 16:06:37.163868 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-wpdrd" Nov 25 16:06:37 crc kubenswrapper[4890]: I1125 16:06:37.167096 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 25 16:06:37 crc kubenswrapper[4890]: I1125 16:06:37.236504 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"a76905fc-a86c-4ed6-982d-0a50bada561f\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 25 16:06:37 crc kubenswrapper[4890]: I1125 16:06:37.236664 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlf64\" (UniqueName: \"kubernetes.io/projected/a76905fc-a86c-4ed6-982d-0a50bada561f-kube-api-access-zlf64\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"a76905fc-a86c-4ed6-982d-0a50bada561f\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 25 16:06:37 crc kubenswrapper[4890]: I1125 16:06:37.339070 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"a76905fc-a86c-4ed6-982d-0a50bada561f\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 25 16:06:37 crc kubenswrapper[4890]: I1125 16:06:37.339315 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlf64\" (UniqueName: \"kubernetes.io/projected/a76905fc-a86c-4ed6-982d-0a50bada561f-kube-api-access-zlf64\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"a76905fc-a86c-4ed6-982d-0a50bada561f\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 25 16:06:37 crc kubenswrapper[4890]: I1125 16:06:37.339872 4890 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"a76905fc-a86c-4ed6-982d-0a50bada561f\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 25 16:06:37 crc kubenswrapper[4890]: I1125 16:06:37.362185 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlf64\" (UniqueName: \"kubernetes.io/projected/a76905fc-a86c-4ed6-982d-0a50bada561f-kube-api-access-zlf64\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"a76905fc-a86c-4ed6-982d-0a50bada561f\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 25 16:06:37 crc kubenswrapper[4890]: I1125 16:06:37.368010 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"a76905fc-a86c-4ed6-982d-0a50bada561f\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 25 16:06:37 crc kubenswrapper[4890]: I1125 16:06:37.487213 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 25 16:06:37 crc kubenswrapper[4890]: I1125 16:06:37.945468 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 25 16:06:37 crc kubenswrapper[4890]: I1125 16:06:37.950320 4890 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 16:06:38 crc kubenswrapper[4890]: I1125 16:06:38.740745 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"a76905fc-a86c-4ed6-982d-0a50bada561f","Type":"ContainerStarted","Data":"984ae7270cd2580ba7a164408703639dad4b4fbee0f8bf499dc37f6a34ebb8d3"} Nov 25 16:06:41 crc kubenswrapper[4890]: I1125 16:06:41.791951 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"a76905fc-a86c-4ed6-982d-0a50bada561f","Type":"ContainerStarted","Data":"4d2b34d34660116dcbad05ea04c616fb05d420945cadcabd7021c067b0356eaf"} Nov 25 16:06:41 crc kubenswrapper[4890]: I1125 16:06:41.811322 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=1.725266678 podStartE2EDuration="4.811302261s" podCreationTimestamp="2025-11-25 16:06:37 +0000 UTC" firstStartedPulling="2025-11-25 16:06:37.950026883 +0000 UTC m=+3856.392489493" lastFinishedPulling="2025-11-25 16:06:41.036062466 +0000 UTC m=+3859.478525076" observedRunningTime="2025-11-25 16:06:41.807588919 +0000 UTC m=+3860.250051549" watchObservedRunningTime="2025-11-25 16:06:41.811302261 +0000 UTC m=+3860.253764871" Nov 25 16:06:48 crc kubenswrapper[4890]: I1125 16:06:48.173448 4890 scope.go:117] "RemoveContainer" containerID="a0bac43b712e038c240449718f3b53c0d88f1d20e0f5e92b7dfb0385a9de4f68" Nov 25 16:06:48 crc kubenswrapper[4890]: E1125 16:06:48.174325 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 16:07:00 crc kubenswrapper[4890]: I1125 16:07:00.104404 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xptmz"] Nov 25 16:07:00 crc kubenswrapper[4890]: I1125 16:07:00.107345 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xptmz" Nov 25 16:07:00 crc kubenswrapper[4890]: I1125 16:07:00.114614 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xptmz"] Nov 25 16:07:00 crc kubenswrapper[4890]: I1125 16:07:00.248018 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f38d160-8131-4f1f-8625-e992e6900787-utilities\") pod \"redhat-marketplace-xptmz\" (UID: \"5f38d160-8131-4f1f-8625-e992e6900787\") " pod="openshift-marketplace/redhat-marketplace-xptmz" Nov 25 16:07:00 crc kubenswrapper[4890]: I1125 16:07:00.248125 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f38d160-8131-4f1f-8625-e992e6900787-catalog-content\") pod \"redhat-marketplace-xptmz\" (UID: \"5f38d160-8131-4f1f-8625-e992e6900787\") " pod="openshift-marketplace/redhat-marketplace-xptmz" Nov 25 16:07:00 crc kubenswrapper[4890]: I1125 16:07:00.248149 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwrfs\" (UniqueName: \"kubernetes.io/projected/5f38d160-8131-4f1f-8625-e992e6900787-kube-api-access-dwrfs\") pod \"redhat-marketplace-xptmz\" (UID: \"5f38d160-8131-4f1f-8625-e992e6900787\") " pod="openshift-marketplace/redhat-marketplace-xptmz" Nov 25 16:07:00 crc kubenswrapper[4890]: I1125 16:07:00.350918 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f38d160-8131-4f1f-8625-e992e6900787-utilities\") pod \"redhat-marketplace-xptmz\" (UID: \"5f38d160-8131-4f1f-8625-e992e6900787\") " pod="openshift-marketplace/redhat-marketplace-xptmz" Nov 25 16:07:00 crc kubenswrapper[4890]: I1125 16:07:00.351048 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f38d160-8131-4f1f-8625-e992e6900787-catalog-content\") pod \"redhat-marketplace-xptmz\" (UID: \"5f38d160-8131-4f1f-8625-e992e6900787\") " pod="openshift-marketplace/redhat-marketplace-xptmz" Nov 25 16:07:00 crc kubenswrapper[4890]: I1125 16:07:00.351080 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwrfs\" (UniqueName: \"kubernetes.io/projected/5f38d160-8131-4f1f-8625-e992e6900787-kube-api-access-dwrfs\") pod \"redhat-marketplace-xptmz\" (UID: \"5f38d160-8131-4f1f-8625-e992e6900787\") " pod="openshift-marketplace/redhat-marketplace-xptmz" Nov 25 16:07:00 crc kubenswrapper[4890]: I1125 16:07:00.351520 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f38d160-8131-4f1f-8625-e992e6900787-utilities\") pod \"redhat-marketplace-xptmz\" (UID: \"5f38d160-8131-4f1f-8625-e992e6900787\") " pod="openshift-marketplace/redhat-marketplace-xptmz" Nov 25 16:07:00 crc kubenswrapper[4890]: I1125 16:07:00.351564 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f38d160-8131-4f1f-8625-e992e6900787-catalog-content\") pod \"redhat-marketplace-xptmz\" (UID: \"5f38d160-8131-4f1f-8625-e992e6900787\") " pod="openshift-marketplace/redhat-marketplace-xptmz" Nov 25 16:07:00 crc kubenswrapper[4890]: I1125 16:07:00.371838 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwrfs\" (UniqueName: \"kubernetes.io/projected/5f38d160-8131-4f1f-8625-e992e6900787-kube-api-access-dwrfs\") pod \"redhat-marketplace-xptmz\" (UID: \"5f38d160-8131-4f1f-8625-e992e6900787\") " pod="openshift-marketplace/redhat-marketplace-xptmz" Nov 25 16:07:00 crc kubenswrapper[4890]: I1125 16:07:00.431259 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xptmz" Nov 25 16:07:00 crc kubenswrapper[4890]: I1125 16:07:00.946675 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xptmz"] Nov 25 16:07:00 crc kubenswrapper[4890]: I1125 16:07:00.998096 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xptmz" event={"ID":"5f38d160-8131-4f1f-8625-e992e6900787","Type":"ContainerStarted","Data":"34ceeadf364d8dc1bb275853924c60892b0817b75ae0eb4613c71d5457e54bb0"} Nov 25 16:07:02 crc kubenswrapper[4890]: I1125 16:07:02.011849 4890 generic.go:334] "Generic (PLEG): container finished" podID="5f38d160-8131-4f1f-8625-e992e6900787" containerID="bea981f4822f437813754061c6710cf56da6bf149352dc27a207d07a7e822127" exitCode=0 Nov 25 16:07:02 crc kubenswrapper[4890]: I1125 16:07:02.011955 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xptmz" event={"ID":"5f38d160-8131-4f1f-8625-e992e6900787","Type":"ContainerDied","Data":"bea981f4822f437813754061c6710cf56da6bf149352dc27a207d07a7e822127"} Nov 25 16:07:03 crc kubenswrapper[4890]: I1125 16:07:03.174189 4890 scope.go:117] "RemoveContainer" containerID="a0bac43b712e038c240449718f3b53c0d88f1d20e0f5e92b7dfb0385a9de4f68" Nov 25 16:07:03 crc kubenswrapper[4890]: E1125 16:07:03.174820 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 16:07:04 crc kubenswrapper[4890]: I1125 16:07:04.869378 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-sv4x5/must-gather-t6wvd"] Nov 25 16:07:04 crc kubenswrapper[4890]: I1125 16:07:04.871596 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sv4x5/must-gather-t6wvd" Nov 25 16:07:04 crc kubenswrapper[4890]: I1125 16:07:04.874286 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdvhk\" (UniqueName: \"kubernetes.io/projected/745efd4c-6be9-4f7e-84d0-81154833dd25-kube-api-access-cdvhk\") pod \"must-gather-t6wvd\" (UID: \"745efd4c-6be9-4f7e-84d0-81154833dd25\") " pod="openshift-must-gather-sv4x5/must-gather-t6wvd" Nov 25 16:07:04 crc kubenswrapper[4890]: I1125 16:07:04.874325 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/745efd4c-6be9-4f7e-84d0-81154833dd25-must-gather-output\") pod \"must-gather-t6wvd\" (UID: \"745efd4c-6be9-4f7e-84d0-81154833dd25\") " pod="openshift-must-gather-sv4x5/must-gather-t6wvd" Nov 25 16:07:04 crc kubenswrapper[4890]: I1125 16:07:04.874609 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-sv4x5"/"kube-root-ca.crt" Nov 25 16:07:04 crc kubenswrapper[4890]: I1125 16:07:04.874783 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-sv4x5"/"default-dockercfg-q94hs" Nov 25 16:07:04 crc kubenswrapper[4890]: I1125 16:07:04.874884 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-sv4x5"/"openshift-service-ca.crt" Nov 25 16:07:04 crc kubenswrapper[4890]: I1125 16:07:04.885927 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-sv4x5/must-gather-t6wvd"] Nov 25 16:07:04 crc kubenswrapper[4890]: I1125 16:07:04.975695 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/745efd4c-6be9-4f7e-84d0-81154833dd25-must-gather-output\") pod \"must-gather-t6wvd\" (UID: \"745efd4c-6be9-4f7e-84d0-81154833dd25\") " pod="openshift-must-gather-sv4x5/must-gather-t6wvd" Nov 25 16:07:04 crc kubenswrapper[4890]: I1125 16:07:04.976251 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdvhk\" (UniqueName: \"kubernetes.io/projected/745efd4c-6be9-4f7e-84d0-81154833dd25-kube-api-access-cdvhk\") pod \"must-gather-t6wvd\" (UID: \"745efd4c-6be9-4f7e-84d0-81154833dd25\") " pod="openshift-must-gather-sv4x5/must-gather-t6wvd" Nov 25 16:07:04 crc kubenswrapper[4890]: I1125 16:07:04.976822 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/745efd4c-6be9-4f7e-84d0-81154833dd25-must-gather-output\") pod \"must-gather-t6wvd\" (UID: \"745efd4c-6be9-4f7e-84d0-81154833dd25\") " pod="openshift-must-gather-sv4x5/must-gather-t6wvd" Nov 25 16:07:05 crc kubenswrapper[4890]: I1125 16:07:05.006082 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdvhk\" (UniqueName: \"kubernetes.io/projected/745efd4c-6be9-4f7e-84d0-81154833dd25-kube-api-access-cdvhk\") pod \"must-gather-t6wvd\" (UID: \"745efd4c-6be9-4f7e-84d0-81154833dd25\") " pod="openshift-must-gather-sv4x5/must-gather-t6wvd" Nov 25 16:07:05 crc kubenswrapper[4890]: I1125 16:07:05.193989 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sv4x5/must-gather-t6wvd" Nov 25 16:07:05 crc kubenswrapper[4890]: I1125 16:07:05.703597 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-sv4x5/must-gather-t6wvd"] Nov 25 16:07:06 crc kubenswrapper[4890]: I1125 16:07:06.059479 4890 generic.go:334] "Generic (PLEG): container finished" podID="5f38d160-8131-4f1f-8625-e992e6900787" containerID="976a6c8e131069d9e2df64cfffbd0f3b30f6702ef6a3aa49e70f2a01adda6eba" exitCode=0 Nov 25 16:07:06 crc kubenswrapper[4890]: I1125 16:07:06.059616 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xptmz" event={"ID":"5f38d160-8131-4f1f-8625-e992e6900787","Type":"ContainerDied","Data":"976a6c8e131069d9e2df64cfffbd0f3b30f6702ef6a3aa49e70f2a01adda6eba"} Nov 25 16:07:06 crc kubenswrapper[4890]: I1125 16:07:06.060944 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-sv4x5/must-gather-t6wvd" event={"ID":"745efd4c-6be9-4f7e-84d0-81154833dd25","Type":"ContainerStarted","Data":"b3e0066ae7e27d4bcdabb3ccdabfd99046922d3980a7c11bcf341e123c298d61"} Nov 25 16:07:09 crc kubenswrapper[4890]: I1125 16:07:09.115123 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xptmz" event={"ID":"5f38d160-8131-4f1f-8625-e992e6900787","Type":"ContainerStarted","Data":"2a6608fc15686baf07faf3c9f7068cb85c19174f9876360267600a6e9ee3e937"} Nov 25 16:07:09 crc kubenswrapper[4890]: I1125 16:07:09.146045 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xptmz" podStartSLOduration=3.096884402 podStartE2EDuration="9.14602521s" podCreationTimestamp="2025-11-25 16:07:00 +0000 UTC" firstStartedPulling="2025-11-25 16:07:02.014095359 +0000 UTC m=+3880.456557969" lastFinishedPulling="2025-11-25 16:07:08.063236167 +0000 UTC m=+3886.505698777" observedRunningTime="2025-11-25 16:07:09.13997479 +0000 UTC m=+3887.582437410" watchObservedRunningTime="2025-11-25 16:07:09.14602521 +0000 UTC m=+3887.588487820" Nov 25 16:07:10 crc kubenswrapper[4890]: I1125 16:07:10.431672 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xptmz" Nov 25 16:07:10 crc kubenswrapper[4890]: I1125 16:07:10.432429 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xptmz" Nov 25 16:07:10 crc kubenswrapper[4890]: I1125 16:07:10.491637 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xptmz" Nov 25 16:07:14 crc kubenswrapper[4890]: I1125 16:07:14.173865 4890 scope.go:117] "RemoveContainer" containerID="a0bac43b712e038c240449718f3b53c0d88f1d20e0f5e92b7dfb0385a9de4f68" Nov 25 16:07:14 crc kubenswrapper[4890]: E1125 16:07:14.175991 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 16:07:15 crc kubenswrapper[4890]: I1125 16:07:15.175288 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-sv4x5/must-gather-t6wvd" event={"ID":"745efd4c-6be9-4f7e-84d0-81154833dd25","Type":"ContainerStarted","Data":"bd5b95d54d1173336682277661b67f4c6ebde09df93ffb10b569c5c18ec42c4e"} Nov 25 16:07:16 crc kubenswrapper[4890]: I1125 16:07:16.192439 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-sv4x5/must-gather-t6wvd" event={"ID":"745efd4c-6be9-4f7e-84d0-81154833dd25","Type":"ContainerStarted","Data":"d5dfa60ad3022889688752f0425048acc7fc507e094c6c6f9fcd6c67ea81c8d6"} Nov 25 16:07:16 crc kubenswrapper[4890]: I1125 16:07:16.209611 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-sv4x5/must-gather-t6wvd" podStartSLOduration=3.617615396 podStartE2EDuration="12.209588865s" podCreationTimestamp="2025-11-25 16:07:04 +0000 UTC" firstStartedPulling="2025-11-25 16:07:05.731255994 +0000 UTC m=+3884.173718604" lastFinishedPulling="2025-11-25 16:07:14.323229463 +0000 UTC m=+3892.765692073" observedRunningTime="2025-11-25 16:07:16.206081358 +0000 UTC m=+3894.648543968" watchObservedRunningTime="2025-11-25 16:07:16.209588865 +0000 UTC m=+3894.652051475" Nov 25 16:07:18 crc kubenswrapper[4890]: I1125 16:07:18.588062 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-sv4x5/crc-debug-8nscj"] Nov 25 16:07:18 crc kubenswrapper[4890]: I1125 16:07:18.589931 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sv4x5/crc-debug-8nscj" Nov 25 16:07:18 crc kubenswrapper[4890]: I1125 16:07:18.679874 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4348f329-d6ac-4f5c-872c-454ab305c34a-host\") pod \"crc-debug-8nscj\" (UID: \"4348f329-d6ac-4f5c-872c-454ab305c34a\") " pod="openshift-must-gather-sv4x5/crc-debug-8nscj" Nov 25 16:07:18 crc kubenswrapper[4890]: I1125 16:07:18.679950 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brnvp\" (UniqueName: \"kubernetes.io/projected/4348f329-d6ac-4f5c-872c-454ab305c34a-kube-api-access-brnvp\") pod \"crc-debug-8nscj\" (UID: \"4348f329-d6ac-4f5c-872c-454ab305c34a\") " pod="openshift-must-gather-sv4x5/crc-debug-8nscj" Nov 25 16:07:18 crc kubenswrapper[4890]: I1125 16:07:18.782141 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4348f329-d6ac-4f5c-872c-454ab305c34a-host\") pod \"crc-debug-8nscj\" (UID: \"4348f329-d6ac-4f5c-872c-454ab305c34a\") " pod="openshift-must-gather-sv4x5/crc-debug-8nscj" Nov 25 16:07:18 crc kubenswrapper[4890]: I1125 16:07:18.782224 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brnvp\" (UniqueName: \"kubernetes.io/projected/4348f329-d6ac-4f5c-872c-454ab305c34a-kube-api-access-brnvp\") pod \"crc-debug-8nscj\" (UID: \"4348f329-d6ac-4f5c-872c-454ab305c34a\") " pod="openshift-must-gather-sv4x5/crc-debug-8nscj" Nov 25 16:07:18 crc kubenswrapper[4890]: I1125 16:07:18.782301 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4348f329-d6ac-4f5c-872c-454ab305c34a-host\") pod \"crc-debug-8nscj\" (UID: \"4348f329-d6ac-4f5c-872c-454ab305c34a\") " pod="openshift-must-gather-sv4x5/crc-debug-8nscj" Nov 25 16:07:18 crc kubenswrapper[4890]: I1125 16:07:18.855258 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brnvp\" (UniqueName: \"kubernetes.io/projected/4348f329-d6ac-4f5c-872c-454ab305c34a-kube-api-access-brnvp\") pod \"crc-debug-8nscj\" (UID: \"4348f329-d6ac-4f5c-872c-454ab305c34a\") " pod="openshift-must-gather-sv4x5/crc-debug-8nscj" Nov 25 16:07:18 crc kubenswrapper[4890]: I1125 16:07:18.911041 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sv4x5/crc-debug-8nscj" Nov 25 16:07:20 crc kubenswrapper[4890]: I1125 16:07:20.232650 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-sv4x5/crc-debug-8nscj" event={"ID":"4348f329-d6ac-4f5c-872c-454ab305c34a","Type":"ContainerStarted","Data":"5c2b69d4865c21b7a06b088c31191bcaec3841045d1f8be84ef1431017d953d6"} Nov 25 16:07:20 crc kubenswrapper[4890]: I1125 16:07:20.494335 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xptmz" Nov 25 16:07:20 crc kubenswrapper[4890]: I1125 16:07:20.561152 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xptmz"] Nov 25 16:07:21 crc kubenswrapper[4890]: I1125 16:07:21.243877 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xptmz" podUID="5f38d160-8131-4f1f-8625-e992e6900787" containerName="registry-server" containerID="cri-o://2a6608fc15686baf07faf3c9f7068cb85c19174f9876360267600a6e9ee3e937" gracePeriod=2 Nov 25 16:07:21 crc kubenswrapper[4890]: I1125 16:07:21.936199 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xptmz" Nov 25 16:07:22 crc kubenswrapper[4890]: I1125 16:07:22.056350 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwrfs\" (UniqueName: \"kubernetes.io/projected/5f38d160-8131-4f1f-8625-e992e6900787-kube-api-access-dwrfs\") pod \"5f38d160-8131-4f1f-8625-e992e6900787\" (UID: \"5f38d160-8131-4f1f-8625-e992e6900787\") " Nov 25 16:07:22 crc kubenswrapper[4890]: I1125 16:07:22.057140 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f38d160-8131-4f1f-8625-e992e6900787-utilities\") pod \"5f38d160-8131-4f1f-8625-e992e6900787\" (UID: \"5f38d160-8131-4f1f-8625-e992e6900787\") " Nov 25 16:07:22 crc kubenswrapper[4890]: I1125 16:07:22.057301 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f38d160-8131-4f1f-8625-e992e6900787-catalog-content\") pod \"5f38d160-8131-4f1f-8625-e992e6900787\" (UID: \"5f38d160-8131-4f1f-8625-e992e6900787\") " Nov 25 16:07:22 crc kubenswrapper[4890]: I1125 16:07:22.057817 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f38d160-8131-4f1f-8625-e992e6900787-utilities" (OuterVolumeSpecName: "utilities") pod "5f38d160-8131-4f1f-8625-e992e6900787" (UID: "5f38d160-8131-4f1f-8625-e992e6900787"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:07:22 crc kubenswrapper[4890]: I1125 16:07:22.058302 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f38d160-8131-4f1f-8625-e992e6900787-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 16:07:22 crc kubenswrapper[4890]: I1125 16:07:22.079605 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f38d160-8131-4f1f-8625-e992e6900787-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5f38d160-8131-4f1f-8625-e992e6900787" (UID: "5f38d160-8131-4f1f-8625-e992e6900787"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:07:22 crc kubenswrapper[4890]: I1125 16:07:22.083481 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f38d160-8131-4f1f-8625-e992e6900787-kube-api-access-dwrfs" (OuterVolumeSpecName: "kube-api-access-dwrfs") pod "5f38d160-8131-4f1f-8625-e992e6900787" (UID: "5f38d160-8131-4f1f-8625-e992e6900787"). InnerVolumeSpecName "kube-api-access-dwrfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:07:22 crc kubenswrapper[4890]: I1125 16:07:22.160752 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwrfs\" (UniqueName: \"kubernetes.io/projected/5f38d160-8131-4f1f-8625-e992e6900787-kube-api-access-dwrfs\") on node \"crc\" DevicePath \"\"" Nov 25 16:07:22 crc kubenswrapper[4890]: I1125 16:07:22.160807 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f38d160-8131-4f1f-8625-e992e6900787-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 16:07:22 crc kubenswrapper[4890]: I1125 16:07:22.263137 4890 generic.go:334] "Generic (PLEG): container finished" podID="5f38d160-8131-4f1f-8625-e992e6900787" containerID="2a6608fc15686baf07faf3c9f7068cb85c19174f9876360267600a6e9ee3e937" exitCode=0 Nov 25 16:07:22 crc kubenswrapper[4890]: I1125 16:07:22.263209 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xptmz" event={"ID":"5f38d160-8131-4f1f-8625-e992e6900787","Type":"ContainerDied","Data":"2a6608fc15686baf07faf3c9f7068cb85c19174f9876360267600a6e9ee3e937"} Nov 25 16:07:22 crc kubenswrapper[4890]: I1125 16:07:22.263252 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xptmz" event={"ID":"5f38d160-8131-4f1f-8625-e992e6900787","Type":"ContainerDied","Data":"34ceeadf364d8dc1bb275853924c60892b0817b75ae0eb4613c71d5457e54bb0"} Nov 25 16:07:22 crc kubenswrapper[4890]: I1125 16:07:22.263275 4890 scope.go:117] "RemoveContainer" containerID="2a6608fc15686baf07faf3c9f7068cb85c19174f9876360267600a6e9ee3e937" Nov 25 16:07:22 crc kubenswrapper[4890]: I1125 16:07:22.263326 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xptmz" Nov 25 16:07:22 crc kubenswrapper[4890]: I1125 16:07:22.297941 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xptmz"] Nov 25 16:07:22 crc kubenswrapper[4890]: I1125 16:07:22.302712 4890 scope.go:117] "RemoveContainer" containerID="976a6c8e131069d9e2df64cfffbd0f3b30f6702ef6a3aa49e70f2a01adda6eba" Nov 25 16:07:22 crc kubenswrapper[4890]: I1125 16:07:22.310793 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xptmz"] Nov 25 16:07:22 crc kubenswrapper[4890]: I1125 16:07:22.361399 4890 scope.go:117] "RemoveContainer" containerID="bea981f4822f437813754061c6710cf56da6bf149352dc27a207d07a7e822127" Nov 25 16:07:22 crc kubenswrapper[4890]: I1125 16:07:22.404427 4890 scope.go:117] "RemoveContainer" containerID="2a6608fc15686baf07faf3c9f7068cb85c19174f9876360267600a6e9ee3e937" Nov 25 16:07:22 crc kubenswrapper[4890]: E1125 16:07:22.405013 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a6608fc15686baf07faf3c9f7068cb85c19174f9876360267600a6e9ee3e937\": container with ID starting with 2a6608fc15686baf07faf3c9f7068cb85c19174f9876360267600a6e9ee3e937 not found: ID does not exist" containerID="2a6608fc15686baf07faf3c9f7068cb85c19174f9876360267600a6e9ee3e937" Nov 25 16:07:22 crc kubenswrapper[4890]: I1125 16:07:22.405083 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a6608fc15686baf07faf3c9f7068cb85c19174f9876360267600a6e9ee3e937"} err="failed to get container status \"2a6608fc15686baf07faf3c9f7068cb85c19174f9876360267600a6e9ee3e937\": rpc error: code = NotFound desc = could not find container \"2a6608fc15686baf07faf3c9f7068cb85c19174f9876360267600a6e9ee3e937\": container with ID starting with 2a6608fc15686baf07faf3c9f7068cb85c19174f9876360267600a6e9ee3e937 not found: ID does not exist" Nov 25 16:07:22 crc kubenswrapper[4890]: I1125 16:07:22.405124 4890 scope.go:117] "RemoveContainer" containerID="976a6c8e131069d9e2df64cfffbd0f3b30f6702ef6a3aa49e70f2a01adda6eba" Nov 25 16:07:22 crc kubenswrapper[4890]: E1125 16:07:22.405797 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"976a6c8e131069d9e2df64cfffbd0f3b30f6702ef6a3aa49e70f2a01adda6eba\": container with ID starting with 976a6c8e131069d9e2df64cfffbd0f3b30f6702ef6a3aa49e70f2a01adda6eba not found: ID does not exist" containerID="976a6c8e131069d9e2df64cfffbd0f3b30f6702ef6a3aa49e70f2a01adda6eba" Nov 25 16:07:22 crc kubenswrapper[4890]: I1125 16:07:22.405852 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"976a6c8e131069d9e2df64cfffbd0f3b30f6702ef6a3aa49e70f2a01adda6eba"} err="failed to get container status \"976a6c8e131069d9e2df64cfffbd0f3b30f6702ef6a3aa49e70f2a01adda6eba\": rpc error: code = NotFound desc = could not find container \"976a6c8e131069d9e2df64cfffbd0f3b30f6702ef6a3aa49e70f2a01adda6eba\": container with ID starting with 976a6c8e131069d9e2df64cfffbd0f3b30f6702ef6a3aa49e70f2a01adda6eba not found: ID does not exist" Nov 25 16:07:22 crc kubenswrapper[4890]: I1125 16:07:22.405890 4890 scope.go:117] "RemoveContainer" containerID="bea981f4822f437813754061c6710cf56da6bf149352dc27a207d07a7e822127" Nov 25 16:07:22 crc kubenswrapper[4890]: E1125 16:07:22.406349 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bea981f4822f437813754061c6710cf56da6bf149352dc27a207d07a7e822127\": container with ID starting with bea981f4822f437813754061c6710cf56da6bf149352dc27a207d07a7e822127 not found: ID does not exist" containerID="bea981f4822f437813754061c6710cf56da6bf149352dc27a207d07a7e822127" Nov 25 16:07:22 crc kubenswrapper[4890]: I1125 16:07:22.406398 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bea981f4822f437813754061c6710cf56da6bf149352dc27a207d07a7e822127"} err="failed to get container status \"bea981f4822f437813754061c6710cf56da6bf149352dc27a207d07a7e822127\": rpc error: code = NotFound desc = could not find container \"bea981f4822f437813754061c6710cf56da6bf149352dc27a207d07a7e822127\": container with ID starting with bea981f4822f437813754061c6710cf56da6bf149352dc27a207d07a7e822127 not found: ID does not exist" Nov 25 16:07:23 crc kubenswrapper[4890]: E1125 16:07:23.166054 4890 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f38d160_8131_4f1f_8625_e992e6900787.slice/crio-34ceeadf364d8dc1bb275853924c60892b0817b75ae0eb4613c71d5457e54bb0\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f38d160_8131_4f1f_8625_e992e6900787.slice\": RecentStats: unable to find data in memory cache]" Nov 25 16:07:24 crc kubenswrapper[4890]: I1125 16:07:24.235760 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f38d160-8131-4f1f-8625-e992e6900787" path="/var/lib/kubelet/pods/5f38d160-8131-4f1f-8625-e992e6900787/volumes" Nov 25 16:07:27 crc kubenswrapper[4890]: I1125 16:07:27.173019 4890 scope.go:117] "RemoveContainer" containerID="a0bac43b712e038c240449718f3b53c0d88f1d20e0f5e92b7dfb0385a9de4f68" Nov 25 16:07:27 crc kubenswrapper[4890]: E1125 16:07:27.173643 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 16:07:33 crc kubenswrapper[4890]: E1125 16:07:33.470223 4890 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f38d160_8131_4f1f_8625_e992e6900787.slice/crio-34ceeadf364d8dc1bb275853924c60892b0817b75ae0eb4613c71d5457e54bb0\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f38d160_8131_4f1f_8625_e992e6900787.slice\": RecentStats: unable to find data in memory cache]" Nov 25 16:07:36 crc kubenswrapper[4890]: E1125 16:07:36.398086 4890 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296" Nov 25 16:07:36 crc kubenswrapper[4890]: E1125 16:07:36.398759 4890 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:container-00,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296,Command:[chroot /host bash -c echo 'TOOLBOX_NAME=toolbox-osp' > /root/.toolboxrc ; rm -rf \"/var/tmp/sos-osp\" && mkdir -p \"/var/tmp/sos-osp\" && sudo podman rm --force toolbox-osp; sudo --preserve-env podman pull --authfile /var/lib/kubelet/config.json registry.redhat.io/rhel9/support-tools && toolbox sos report --batch --all-logs --only-plugins block,cifs,crio,devicemapper,devices,firewall_tables,firewalld,iscsi,lvm2,memory,multipath,nfs,nis,nvme,podman,process,processor,selinux,scsi,udev,logs,crypto --tmp-dir=\"/var/tmp/sos-osp\" && if [[ \"$(ls /var/log/pods/*/{*.log.*,*/*.log.*} 2>/dev/null)\" != '' ]]; then tar --ignore-failed-read --warning=no-file-changed -cJf \"/var/tmp/sos-osp/podlogs.tar.xz\" --transform 's,^,podlogs/,' /var/log/pods/*/{*.log.*,*/*.log.*} || true; fi],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:TMOUT,Value:900,ValueFrom:nil,},EnvVar{Name:HOST,Value:/host,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host,ReadOnly:false,MountPath:/host,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-brnvp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod crc-debug-8nscj_openshift-must-gather-sv4x5(4348f329-d6ac-4f5c-872c-454ab305c34a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 16:07:36 crc kubenswrapper[4890]: E1125 16:07:36.401022 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"container-00\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openshift-must-gather-sv4x5/crc-debug-8nscj" podUID="4348f329-d6ac-4f5c-872c-454ab305c34a" Nov 25 16:07:37 crc kubenswrapper[4890]: E1125 16:07:37.415064 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"container-00\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296\\\"\"" pod="openshift-must-gather-sv4x5/crc-debug-8nscj" podUID="4348f329-d6ac-4f5c-872c-454ab305c34a" Nov 25 16:07:39 crc kubenswrapper[4890]: I1125 16:07:39.174357 4890 scope.go:117] "RemoveContainer" containerID="a0bac43b712e038c240449718f3b53c0d88f1d20e0f5e92b7dfb0385a9de4f68" Nov 25 16:07:39 crc kubenswrapper[4890]: E1125 16:07:39.175030 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 16:07:42 crc kubenswrapper[4890]: I1125 16:07:42.010282 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qpsjg"] Nov 25 16:07:42 crc kubenswrapper[4890]: E1125 16:07:42.011268 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f38d160-8131-4f1f-8625-e992e6900787" containerName="extract-content" Nov 25 16:07:42 crc kubenswrapper[4890]: I1125 16:07:42.011285 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f38d160-8131-4f1f-8625-e992e6900787" containerName="extract-content" Nov 25 16:07:42 crc kubenswrapper[4890]: E1125 16:07:42.011321 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f38d160-8131-4f1f-8625-e992e6900787" containerName="registry-server" Nov 25 16:07:42 crc kubenswrapper[4890]: I1125 16:07:42.011329 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f38d160-8131-4f1f-8625-e992e6900787" containerName="registry-server" Nov 25 16:07:42 crc kubenswrapper[4890]: E1125 16:07:42.011364 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f38d160-8131-4f1f-8625-e992e6900787" containerName="extract-utilities" Nov 25 16:07:42 crc kubenswrapper[4890]: I1125 16:07:42.011374 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f38d160-8131-4f1f-8625-e992e6900787" containerName="extract-utilities" Nov 25 16:07:42 crc kubenswrapper[4890]: I1125 16:07:42.011776 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f38d160-8131-4f1f-8625-e992e6900787" containerName="registry-server" Nov 25 16:07:42 crc kubenswrapper[4890]: I1125 16:07:42.013655 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qpsjg" Nov 25 16:07:42 crc kubenswrapper[4890]: I1125 16:07:42.024545 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qpsjg"] Nov 25 16:07:42 crc kubenswrapper[4890]: I1125 16:07:42.096750 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5570e14-7c3e-44cf-a400-220a62fb7502-utilities\") pod \"certified-operators-qpsjg\" (UID: \"f5570e14-7c3e-44cf-a400-220a62fb7502\") " pod="openshift-marketplace/certified-operators-qpsjg" Nov 25 16:07:42 crc kubenswrapper[4890]: I1125 16:07:42.096878 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dwc4\" (UniqueName: \"kubernetes.io/projected/f5570e14-7c3e-44cf-a400-220a62fb7502-kube-api-access-7dwc4\") pod \"certified-operators-qpsjg\" (UID: \"f5570e14-7c3e-44cf-a400-220a62fb7502\") " pod="openshift-marketplace/certified-operators-qpsjg" Nov 25 16:07:42 crc kubenswrapper[4890]: I1125 16:07:42.097052 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5570e14-7c3e-44cf-a400-220a62fb7502-catalog-content\") pod \"certified-operators-qpsjg\" (UID: \"f5570e14-7c3e-44cf-a400-220a62fb7502\") " pod="openshift-marketplace/certified-operators-qpsjg" Nov 25 16:07:42 crc kubenswrapper[4890]: I1125 16:07:42.198844 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5570e14-7c3e-44cf-a400-220a62fb7502-catalog-content\") pod \"certified-operators-qpsjg\" (UID: \"f5570e14-7c3e-44cf-a400-220a62fb7502\") " pod="openshift-marketplace/certified-operators-qpsjg" Nov 25 16:07:42 crc kubenswrapper[4890]: I1125 16:07:42.199232 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5570e14-7c3e-44cf-a400-220a62fb7502-utilities\") pod \"certified-operators-qpsjg\" (UID: \"f5570e14-7c3e-44cf-a400-220a62fb7502\") " pod="openshift-marketplace/certified-operators-qpsjg" Nov 25 16:07:42 crc kubenswrapper[4890]: I1125 16:07:42.199383 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dwc4\" (UniqueName: \"kubernetes.io/projected/f5570e14-7c3e-44cf-a400-220a62fb7502-kube-api-access-7dwc4\") pod \"certified-operators-qpsjg\" (UID: \"f5570e14-7c3e-44cf-a400-220a62fb7502\") " pod="openshift-marketplace/certified-operators-qpsjg" Nov 25 16:07:42 crc kubenswrapper[4890]: I1125 16:07:42.199646 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5570e14-7c3e-44cf-a400-220a62fb7502-catalog-content\") pod \"certified-operators-qpsjg\" (UID: \"f5570e14-7c3e-44cf-a400-220a62fb7502\") " pod="openshift-marketplace/certified-operators-qpsjg" Nov 25 16:07:42 crc kubenswrapper[4890]: I1125 16:07:42.199762 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5570e14-7c3e-44cf-a400-220a62fb7502-utilities\") pod \"certified-operators-qpsjg\" (UID: \"f5570e14-7c3e-44cf-a400-220a62fb7502\") " pod="openshift-marketplace/certified-operators-qpsjg" Nov 25 16:07:42 crc kubenswrapper[4890]: I1125 16:07:42.217873 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dwc4\" (UniqueName: \"kubernetes.io/projected/f5570e14-7c3e-44cf-a400-220a62fb7502-kube-api-access-7dwc4\") pod \"certified-operators-qpsjg\" (UID: \"f5570e14-7c3e-44cf-a400-220a62fb7502\") " pod="openshift-marketplace/certified-operators-qpsjg" Nov 25 16:07:42 crc kubenswrapper[4890]: I1125 16:07:42.338268 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qpsjg" Nov 25 16:07:42 crc kubenswrapper[4890]: I1125 16:07:42.887855 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qpsjg"] Nov 25 16:07:43 crc kubenswrapper[4890]: I1125 16:07:43.508697 4890 generic.go:334] "Generic (PLEG): container finished" podID="f5570e14-7c3e-44cf-a400-220a62fb7502" containerID="256a806d0841e24e1b7aa4e5267e165b136aef351d4cc7074743c27db37680af" exitCode=0 Nov 25 16:07:43 crc kubenswrapper[4890]: I1125 16:07:43.508783 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qpsjg" event={"ID":"f5570e14-7c3e-44cf-a400-220a62fb7502","Type":"ContainerDied","Data":"256a806d0841e24e1b7aa4e5267e165b136aef351d4cc7074743c27db37680af"} Nov 25 16:07:43 crc kubenswrapper[4890]: I1125 16:07:43.509026 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qpsjg" event={"ID":"f5570e14-7c3e-44cf-a400-220a62fb7502","Type":"ContainerStarted","Data":"3aa6930101b292b65ff2ed9804a9c522c693ece229d59b8e3623edc01cac627e"} Nov 25 16:07:43 crc kubenswrapper[4890]: E1125 16:07:43.729556 4890 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f38d160_8131_4f1f_8625_e992e6900787.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f38d160_8131_4f1f_8625_e992e6900787.slice/crio-34ceeadf364d8dc1bb275853924c60892b0817b75ae0eb4613c71d5457e54bb0\": RecentStats: unable to find data in memory cache]" Nov 25 16:07:45 crc kubenswrapper[4890]: I1125 16:07:45.534620 4890 generic.go:334] "Generic (PLEG): container finished" podID="f5570e14-7c3e-44cf-a400-220a62fb7502" containerID="a26b20dd889462559b9f2ffa1828aeb9e5940d565b1b1663b42a58eba54a5ad9" exitCode=0 Nov 25 16:07:45 crc kubenswrapper[4890]: I1125 16:07:45.534845 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qpsjg" event={"ID":"f5570e14-7c3e-44cf-a400-220a62fb7502","Type":"ContainerDied","Data":"a26b20dd889462559b9f2ffa1828aeb9e5940d565b1b1663b42a58eba54a5ad9"} Nov 25 16:07:50 crc kubenswrapper[4890]: I1125 16:07:50.174400 4890 scope.go:117] "RemoveContainer" containerID="a0bac43b712e038c240449718f3b53c0d88f1d20e0f5e92b7dfb0385a9de4f68" Nov 25 16:07:50 crc kubenswrapper[4890]: E1125 16:07:50.175430 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 16:07:53 crc kubenswrapper[4890]: I1125 16:07:53.639335 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qpsjg" event={"ID":"f5570e14-7c3e-44cf-a400-220a62fb7502","Type":"ContainerStarted","Data":"80d0ac3dd10b50752d2b25927d4c9b9437781619f285817b1228a4ddcc32b5b3"} Nov 25 16:07:53 crc kubenswrapper[4890]: I1125 16:07:53.668582 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qpsjg" podStartSLOduration=3.201333936 podStartE2EDuration="12.668540572s" podCreationTimestamp="2025-11-25 16:07:41 +0000 UTC" firstStartedPulling="2025-11-25 16:07:43.51475115 +0000 UTC m=+3921.957213760" lastFinishedPulling="2025-11-25 16:07:52.981957786 +0000 UTC m=+3931.424420396" observedRunningTime="2025-11-25 16:07:53.656869542 +0000 UTC m=+3932.099332152" watchObservedRunningTime="2025-11-25 16:07:53.668540572 +0000 UTC m=+3932.111003182" Nov 25 16:07:54 crc kubenswrapper[4890]: E1125 16:07:54.013483 4890 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f38d160_8131_4f1f_8625_e992e6900787.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f38d160_8131_4f1f_8625_e992e6900787.slice/crio-34ceeadf364d8dc1bb275853924c60892b0817b75ae0eb4613c71d5457e54bb0\": RecentStats: unable to find data in memory cache]" Nov 25 16:07:55 crc kubenswrapper[4890]: I1125 16:07:55.663204 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-sv4x5/crc-debug-8nscj" event={"ID":"4348f329-d6ac-4f5c-872c-454ab305c34a","Type":"ContainerStarted","Data":"cdd466e25f42e244d833b2e793ad80fb89bfbfbbbafe96c20bb8b1961b955f21"} Nov 25 16:08:02 crc kubenswrapper[4890]: I1125 16:08:02.339357 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qpsjg" Nov 25 16:08:02 crc kubenswrapper[4890]: I1125 16:08:02.339916 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qpsjg" Nov 25 16:08:02 crc kubenswrapper[4890]: I1125 16:08:02.396855 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qpsjg" Nov 25 16:08:02 crc kubenswrapper[4890]: I1125 16:08:02.431238 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-sv4x5/crc-debug-8nscj" podStartSLOduration=8.524051493 podStartE2EDuration="44.431219656s" podCreationTimestamp="2025-11-25 16:07:18 +0000 UTC" firstStartedPulling="2025-11-25 16:07:19.254106469 +0000 UTC m=+3897.696569079" lastFinishedPulling="2025-11-25 16:07:55.161274632 +0000 UTC m=+3933.603737242" observedRunningTime="2025-11-25 16:07:55.679215827 +0000 UTC m=+3934.121678437" watchObservedRunningTime="2025-11-25 16:08:02.431219656 +0000 UTC m=+3940.873682256" Nov 25 16:08:02 crc kubenswrapper[4890]: I1125 16:08:02.787075 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qpsjg" Nov 25 16:08:02 crc kubenswrapper[4890]: I1125 16:08:02.846121 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qpsjg"] Nov 25 16:08:04 crc kubenswrapper[4890]: E1125 16:08:04.303879 4890 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f38d160_8131_4f1f_8625_e992e6900787.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f38d160_8131_4f1f_8625_e992e6900787.slice/crio-34ceeadf364d8dc1bb275853924c60892b0817b75ae0eb4613c71d5457e54bb0\": RecentStats: unable to find data in memory cache]" Nov 25 16:08:05 crc kubenswrapper[4890]: I1125 16:08:05.013946 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qpsjg" podUID="f5570e14-7c3e-44cf-a400-220a62fb7502" containerName="registry-server" containerID="cri-o://80d0ac3dd10b50752d2b25927d4c9b9437781619f285817b1228a4ddcc32b5b3" gracePeriod=2 Nov 25 16:08:05 crc kubenswrapper[4890]: I1125 16:08:05.178308 4890 scope.go:117] "RemoveContainer" containerID="a0bac43b712e038c240449718f3b53c0d88f1d20e0f5e92b7dfb0385a9de4f68" Nov 25 16:08:06 crc kubenswrapper[4890]: I1125 16:08:06.027648 4890 generic.go:334] "Generic (PLEG): container finished" podID="f5570e14-7c3e-44cf-a400-220a62fb7502" containerID="80d0ac3dd10b50752d2b25927d4c9b9437781619f285817b1228a4ddcc32b5b3" exitCode=0 Nov 25 16:08:06 crc kubenswrapper[4890]: I1125 16:08:06.027722 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qpsjg" event={"ID":"f5570e14-7c3e-44cf-a400-220a62fb7502","Type":"ContainerDied","Data":"80d0ac3dd10b50752d2b25927d4c9b9437781619f285817b1228a4ddcc32b5b3"} Nov 25 16:08:08 crc kubenswrapper[4890]: I1125 16:08:08.529469 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qpsjg" Nov 25 16:08:08 crc kubenswrapper[4890]: I1125 16:08:08.672071 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7dwc4\" (UniqueName: \"kubernetes.io/projected/f5570e14-7c3e-44cf-a400-220a62fb7502-kube-api-access-7dwc4\") pod \"f5570e14-7c3e-44cf-a400-220a62fb7502\" (UID: \"f5570e14-7c3e-44cf-a400-220a62fb7502\") " Nov 25 16:08:08 crc kubenswrapper[4890]: I1125 16:08:08.672276 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5570e14-7c3e-44cf-a400-220a62fb7502-utilities\") pod \"f5570e14-7c3e-44cf-a400-220a62fb7502\" (UID: \"f5570e14-7c3e-44cf-a400-220a62fb7502\") " Nov 25 16:08:08 crc kubenswrapper[4890]: I1125 16:08:08.672372 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5570e14-7c3e-44cf-a400-220a62fb7502-catalog-content\") pod \"f5570e14-7c3e-44cf-a400-220a62fb7502\" (UID: \"f5570e14-7c3e-44cf-a400-220a62fb7502\") " Nov 25 16:08:08 crc kubenswrapper[4890]: I1125 16:08:08.672929 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5570e14-7c3e-44cf-a400-220a62fb7502-utilities" (OuterVolumeSpecName: "utilities") pod "f5570e14-7c3e-44cf-a400-220a62fb7502" (UID: "f5570e14-7c3e-44cf-a400-220a62fb7502"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:08:08 crc kubenswrapper[4890]: I1125 16:08:08.678646 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5570e14-7c3e-44cf-a400-220a62fb7502-kube-api-access-7dwc4" (OuterVolumeSpecName: "kube-api-access-7dwc4") pod "f5570e14-7c3e-44cf-a400-220a62fb7502" (UID: "f5570e14-7c3e-44cf-a400-220a62fb7502"). InnerVolumeSpecName "kube-api-access-7dwc4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:08:08 crc kubenswrapper[4890]: I1125 16:08:08.727237 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5570e14-7c3e-44cf-a400-220a62fb7502-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f5570e14-7c3e-44cf-a400-220a62fb7502" (UID: "f5570e14-7c3e-44cf-a400-220a62fb7502"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:08:08 crc kubenswrapper[4890]: I1125 16:08:08.775425 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7dwc4\" (UniqueName: \"kubernetes.io/projected/f5570e14-7c3e-44cf-a400-220a62fb7502-kube-api-access-7dwc4\") on node \"crc\" DevicePath \"\"" Nov 25 16:08:08 crc kubenswrapper[4890]: I1125 16:08:08.775467 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5570e14-7c3e-44cf-a400-220a62fb7502-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 16:08:08 crc kubenswrapper[4890]: I1125 16:08:08.775477 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5570e14-7c3e-44cf-a400-220a62fb7502-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 16:08:09 crc kubenswrapper[4890]: I1125 16:08:09.063458 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" event={"ID":"4e4f849d-f239-4727-a73e-18327856929a","Type":"ContainerStarted","Data":"0ee5f289dd255f61c3f8c4e86c84d7044d423518d51541add4da1a581b273eb1"} Nov 25 16:08:09 crc kubenswrapper[4890]: I1125 16:08:09.068521 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qpsjg" event={"ID":"f5570e14-7c3e-44cf-a400-220a62fb7502","Type":"ContainerDied","Data":"3aa6930101b292b65ff2ed9804a9c522c693ece229d59b8e3623edc01cac627e"} Nov 25 16:08:09 crc kubenswrapper[4890]: I1125 16:08:09.068597 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qpsjg" Nov 25 16:08:09 crc kubenswrapper[4890]: I1125 16:08:09.068656 4890 scope.go:117] "RemoveContainer" containerID="80d0ac3dd10b50752d2b25927d4c9b9437781619f285817b1228a4ddcc32b5b3" Nov 25 16:08:09 crc kubenswrapper[4890]: I1125 16:08:09.115636 4890 scope.go:117] "RemoveContainer" containerID="a26b20dd889462559b9f2ffa1828aeb9e5940d565b1b1663b42a58eba54a5ad9" Nov 25 16:08:09 crc kubenswrapper[4890]: I1125 16:08:09.146129 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qpsjg"] Nov 25 16:08:09 crc kubenswrapper[4890]: I1125 16:08:09.157440 4890 scope.go:117] "RemoveContainer" containerID="256a806d0841e24e1b7aa4e5267e165b136aef351d4cc7074743c27db37680af" Nov 25 16:08:09 crc kubenswrapper[4890]: I1125 16:08:09.157656 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qpsjg"] Nov 25 16:08:10 crc kubenswrapper[4890]: I1125 16:08:10.185793 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5570e14-7c3e-44cf-a400-220a62fb7502" path="/var/lib/kubelet/pods/f5570e14-7c3e-44cf-a400-220a62fb7502/volumes" Nov 25 16:08:14 crc kubenswrapper[4890]: E1125 16:08:14.600206 4890 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f38d160_8131_4f1f_8625_e992e6900787.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f38d160_8131_4f1f_8625_e992e6900787.slice/crio-34ceeadf364d8dc1bb275853924c60892b0817b75ae0eb4613c71d5457e54bb0\": RecentStats: unable to find data in memory cache]" Nov 25 16:08:50 crc kubenswrapper[4890]: I1125 16:08:50.520670 4890 generic.go:334] "Generic (PLEG): container finished" podID="4348f329-d6ac-4f5c-872c-454ab305c34a" containerID="cdd466e25f42e244d833b2e793ad80fb89bfbfbbbafe96c20bb8b1961b955f21" exitCode=0 Nov 25 16:08:50 crc kubenswrapper[4890]: I1125 16:08:50.520718 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-sv4x5/crc-debug-8nscj" event={"ID":"4348f329-d6ac-4f5c-872c-454ab305c34a","Type":"ContainerDied","Data":"cdd466e25f42e244d833b2e793ad80fb89bfbfbbbafe96c20bb8b1961b955f21"} Nov 25 16:08:51 crc kubenswrapper[4890]: I1125 16:08:51.634100 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sv4x5/crc-debug-8nscj" Nov 25 16:08:51 crc kubenswrapper[4890]: I1125 16:08:51.731223 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4348f329-d6ac-4f5c-872c-454ab305c34a-host\") pod \"4348f329-d6ac-4f5c-872c-454ab305c34a\" (UID: \"4348f329-d6ac-4f5c-872c-454ab305c34a\") " Nov 25 16:08:51 crc kubenswrapper[4890]: I1125 16:08:51.731315 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-brnvp\" (UniqueName: \"kubernetes.io/projected/4348f329-d6ac-4f5c-872c-454ab305c34a-kube-api-access-brnvp\") pod \"4348f329-d6ac-4f5c-872c-454ab305c34a\" (UID: \"4348f329-d6ac-4f5c-872c-454ab305c34a\") " Nov 25 16:08:51 crc kubenswrapper[4890]: I1125 16:08:51.731325 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4348f329-d6ac-4f5c-872c-454ab305c34a-host" (OuterVolumeSpecName: "host") pod "4348f329-d6ac-4f5c-872c-454ab305c34a" (UID: "4348f329-d6ac-4f5c-872c-454ab305c34a"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 16:08:51 crc kubenswrapper[4890]: I1125 16:08:51.731874 4890 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4348f329-d6ac-4f5c-872c-454ab305c34a-host\") on node \"crc\" DevicePath \"\"" Nov 25 16:08:51 crc kubenswrapper[4890]: I1125 16:08:51.740366 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-sv4x5/crc-debug-8nscj"] Nov 25 16:08:51 crc kubenswrapper[4890]: I1125 16:08:51.746416 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4348f329-d6ac-4f5c-872c-454ab305c34a-kube-api-access-brnvp" (OuterVolumeSpecName: "kube-api-access-brnvp") pod "4348f329-d6ac-4f5c-872c-454ab305c34a" (UID: "4348f329-d6ac-4f5c-872c-454ab305c34a"). InnerVolumeSpecName "kube-api-access-brnvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:08:51 crc kubenswrapper[4890]: I1125 16:08:51.753793 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-sv4x5/crc-debug-8nscj"] Nov 25 16:08:51 crc kubenswrapper[4890]: I1125 16:08:51.834489 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-brnvp\" (UniqueName: \"kubernetes.io/projected/4348f329-d6ac-4f5c-872c-454ab305c34a-kube-api-access-brnvp\") on node \"crc\" DevicePath \"\"" Nov 25 16:08:52 crc kubenswrapper[4890]: I1125 16:08:52.183480 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4348f329-d6ac-4f5c-872c-454ab305c34a" path="/var/lib/kubelet/pods/4348f329-d6ac-4f5c-872c-454ab305c34a/volumes" Nov 25 16:08:52 crc kubenswrapper[4890]: I1125 16:08:52.542646 4890 scope.go:117] "RemoveContainer" containerID="cdd466e25f42e244d833b2e793ad80fb89bfbfbbbafe96c20bb8b1961b955f21" Nov 25 16:08:52 crc kubenswrapper[4890]: I1125 16:08:52.542659 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sv4x5/crc-debug-8nscj" Nov 25 16:08:52 crc kubenswrapper[4890]: I1125 16:08:52.919865 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-sv4x5/crc-debug-4fv8n"] Nov 25 16:08:52 crc kubenswrapper[4890]: E1125 16:08:52.920761 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4348f329-d6ac-4f5c-872c-454ab305c34a" containerName="container-00" Nov 25 16:08:52 crc kubenswrapper[4890]: I1125 16:08:52.920775 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="4348f329-d6ac-4f5c-872c-454ab305c34a" containerName="container-00" Nov 25 16:08:52 crc kubenswrapper[4890]: E1125 16:08:52.920791 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5570e14-7c3e-44cf-a400-220a62fb7502" containerName="extract-content" Nov 25 16:08:52 crc kubenswrapper[4890]: I1125 16:08:52.920797 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5570e14-7c3e-44cf-a400-220a62fb7502" containerName="extract-content" Nov 25 16:08:52 crc kubenswrapper[4890]: E1125 16:08:52.920826 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5570e14-7c3e-44cf-a400-220a62fb7502" containerName="registry-server" Nov 25 16:08:52 crc kubenswrapper[4890]: I1125 16:08:52.920836 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5570e14-7c3e-44cf-a400-220a62fb7502" containerName="registry-server" Nov 25 16:08:52 crc kubenswrapper[4890]: E1125 16:08:52.920851 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5570e14-7c3e-44cf-a400-220a62fb7502" containerName="extract-utilities" Nov 25 16:08:52 crc kubenswrapper[4890]: I1125 16:08:52.920859 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5570e14-7c3e-44cf-a400-220a62fb7502" containerName="extract-utilities" Nov 25 16:08:52 crc kubenswrapper[4890]: I1125 16:08:52.921061 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5570e14-7c3e-44cf-a400-220a62fb7502" containerName="registry-server" Nov 25 16:08:52 crc kubenswrapper[4890]: I1125 16:08:52.921078 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="4348f329-d6ac-4f5c-872c-454ab305c34a" containerName="container-00" Nov 25 16:08:52 crc kubenswrapper[4890]: I1125 16:08:52.921825 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sv4x5/crc-debug-4fv8n" Nov 25 16:08:52 crc kubenswrapper[4890]: I1125 16:08:52.962148 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/682c875a-ddfa-4b64-b669-617bf9e17f2f-host\") pod \"crc-debug-4fv8n\" (UID: \"682c875a-ddfa-4b64-b669-617bf9e17f2f\") " pod="openshift-must-gather-sv4x5/crc-debug-4fv8n" Nov 25 16:08:52 crc kubenswrapper[4890]: I1125 16:08:52.962280 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlkqn\" (UniqueName: \"kubernetes.io/projected/682c875a-ddfa-4b64-b669-617bf9e17f2f-kube-api-access-jlkqn\") pod \"crc-debug-4fv8n\" (UID: \"682c875a-ddfa-4b64-b669-617bf9e17f2f\") " pod="openshift-must-gather-sv4x5/crc-debug-4fv8n" Nov 25 16:08:53 crc kubenswrapper[4890]: I1125 16:08:53.064202 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlkqn\" (UniqueName: \"kubernetes.io/projected/682c875a-ddfa-4b64-b669-617bf9e17f2f-kube-api-access-jlkqn\") pod \"crc-debug-4fv8n\" (UID: \"682c875a-ddfa-4b64-b669-617bf9e17f2f\") " pod="openshift-must-gather-sv4x5/crc-debug-4fv8n" Nov 25 16:08:53 crc kubenswrapper[4890]: I1125 16:08:53.064381 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/682c875a-ddfa-4b64-b669-617bf9e17f2f-host\") pod \"crc-debug-4fv8n\" (UID: \"682c875a-ddfa-4b64-b669-617bf9e17f2f\") " pod="openshift-must-gather-sv4x5/crc-debug-4fv8n" Nov 25 16:08:53 crc kubenswrapper[4890]: I1125 16:08:53.064524 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/682c875a-ddfa-4b64-b669-617bf9e17f2f-host\") pod \"crc-debug-4fv8n\" (UID: \"682c875a-ddfa-4b64-b669-617bf9e17f2f\") " pod="openshift-must-gather-sv4x5/crc-debug-4fv8n" Nov 25 16:08:53 crc kubenswrapper[4890]: I1125 16:08:53.086871 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlkqn\" (UniqueName: \"kubernetes.io/projected/682c875a-ddfa-4b64-b669-617bf9e17f2f-kube-api-access-jlkqn\") pod \"crc-debug-4fv8n\" (UID: \"682c875a-ddfa-4b64-b669-617bf9e17f2f\") " pod="openshift-must-gather-sv4x5/crc-debug-4fv8n" Nov 25 16:08:53 crc kubenswrapper[4890]: I1125 16:08:53.242029 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sv4x5/crc-debug-4fv8n" Nov 25 16:08:53 crc kubenswrapper[4890]: I1125 16:08:53.556933 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-sv4x5/crc-debug-4fv8n" event={"ID":"682c875a-ddfa-4b64-b669-617bf9e17f2f","Type":"ContainerStarted","Data":"71e25fe5fa672420688c2c248d1dc3442f1d52fde0faae968b620082f34571bc"} Nov 25 16:08:53 crc kubenswrapper[4890]: I1125 16:08:53.557421 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-sv4x5/crc-debug-4fv8n" event={"ID":"682c875a-ddfa-4b64-b669-617bf9e17f2f","Type":"ContainerStarted","Data":"4357976d84b27069df69f183a96e671076c5e69286ebe2e1532f81a091284e11"} Nov 25 16:08:53 crc kubenswrapper[4890]: I1125 16:08:53.578846 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-sv4x5/crc-debug-4fv8n" podStartSLOduration=1.578812329 podStartE2EDuration="1.578812329s" podCreationTimestamp="2025-11-25 16:08:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:08:53.572038881 +0000 UTC m=+3992.014501491" watchObservedRunningTime="2025-11-25 16:08:53.578812329 +0000 UTC m=+3992.021274939" Nov 25 16:08:54 crc kubenswrapper[4890]: I1125 16:08:54.570004 4890 generic.go:334] "Generic (PLEG): container finished" podID="682c875a-ddfa-4b64-b669-617bf9e17f2f" containerID="71e25fe5fa672420688c2c248d1dc3442f1d52fde0faae968b620082f34571bc" exitCode=0 Nov 25 16:08:54 crc kubenswrapper[4890]: I1125 16:08:54.570080 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-sv4x5/crc-debug-4fv8n" event={"ID":"682c875a-ddfa-4b64-b669-617bf9e17f2f","Type":"ContainerDied","Data":"71e25fe5fa672420688c2c248d1dc3442f1d52fde0faae968b620082f34571bc"} Nov 25 16:08:55 crc kubenswrapper[4890]: I1125 16:08:55.698005 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sv4x5/crc-debug-4fv8n" Nov 25 16:08:55 crc kubenswrapper[4890]: I1125 16:08:55.721350 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/682c875a-ddfa-4b64-b669-617bf9e17f2f-host\") pod \"682c875a-ddfa-4b64-b669-617bf9e17f2f\" (UID: \"682c875a-ddfa-4b64-b669-617bf9e17f2f\") " Nov 25 16:08:55 crc kubenswrapper[4890]: I1125 16:08:55.721429 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jlkqn\" (UniqueName: \"kubernetes.io/projected/682c875a-ddfa-4b64-b669-617bf9e17f2f-kube-api-access-jlkqn\") pod \"682c875a-ddfa-4b64-b669-617bf9e17f2f\" (UID: \"682c875a-ddfa-4b64-b669-617bf9e17f2f\") " Nov 25 16:08:55 crc kubenswrapper[4890]: I1125 16:08:55.721507 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/682c875a-ddfa-4b64-b669-617bf9e17f2f-host" (OuterVolumeSpecName: "host") pod "682c875a-ddfa-4b64-b669-617bf9e17f2f" (UID: "682c875a-ddfa-4b64-b669-617bf9e17f2f"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 16:08:55 crc kubenswrapper[4890]: I1125 16:08:55.722093 4890 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/682c875a-ddfa-4b64-b669-617bf9e17f2f-host\") on node \"crc\" DevicePath \"\"" Nov 25 16:08:55 crc kubenswrapper[4890]: I1125 16:08:55.737598 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/682c875a-ddfa-4b64-b669-617bf9e17f2f-kube-api-access-jlkqn" (OuterVolumeSpecName: "kube-api-access-jlkqn") pod "682c875a-ddfa-4b64-b669-617bf9e17f2f" (UID: "682c875a-ddfa-4b64-b669-617bf9e17f2f"). InnerVolumeSpecName "kube-api-access-jlkqn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:08:55 crc kubenswrapper[4890]: I1125 16:08:55.739751 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-sv4x5/crc-debug-4fv8n"] Nov 25 16:08:55 crc kubenswrapper[4890]: I1125 16:08:55.749219 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-sv4x5/crc-debug-4fv8n"] Nov 25 16:08:55 crc kubenswrapper[4890]: I1125 16:08:55.823997 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jlkqn\" (UniqueName: \"kubernetes.io/projected/682c875a-ddfa-4b64-b669-617bf9e17f2f-kube-api-access-jlkqn\") on node \"crc\" DevicePath \"\"" Nov 25 16:08:56 crc kubenswrapper[4890]: I1125 16:08:56.212605 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="682c875a-ddfa-4b64-b669-617bf9e17f2f" path="/var/lib/kubelet/pods/682c875a-ddfa-4b64-b669-617bf9e17f2f/volumes" Nov 25 16:08:56 crc kubenswrapper[4890]: I1125 16:08:56.591033 4890 scope.go:117] "RemoveContainer" containerID="71e25fe5fa672420688c2c248d1dc3442f1d52fde0faae968b620082f34571bc" Nov 25 16:08:56 crc kubenswrapper[4890]: I1125 16:08:56.591099 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sv4x5/crc-debug-4fv8n" Nov 25 16:08:56 crc kubenswrapper[4890]: I1125 16:08:56.947786 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-sv4x5/crc-debug-msdvk"] Nov 25 16:08:56 crc kubenswrapper[4890]: E1125 16:08:56.948249 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="682c875a-ddfa-4b64-b669-617bf9e17f2f" containerName="container-00" Nov 25 16:08:56 crc kubenswrapper[4890]: I1125 16:08:56.948263 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="682c875a-ddfa-4b64-b669-617bf9e17f2f" containerName="container-00" Nov 25 16:08:56 crc kubenswrapper[4890]: I1125 16:08:56.948485 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="682c875a-ddfa-4b64-b669-617bf9e17f2f" containerName="container-00" Nov 25 16:08:56 crc kubenswrapper[4890]: I1125 16:08:56.949342 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sv4x5/crc-debug-msdvk" Nov 25 16:08:57 crc kubenswrapper[4890]: I1125 16:08:57.048204 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wz66\" (UniqueName: \"kubernetes.io/projected/e995f803-3996-491e-8e7a-a6db48cda9ee-kube-api-access-4wz66\") pod \"crc-debug-msdvk\" (UID: \"e995f803-3996-491e-8e7a-a6db48cda9ee\") " pod="openshift-must-gather-sv4x5/crc-debug-msdvk" Nov 25 16:08:57 crc kubenswrapper[4890]: I1125 16:08:57.048760 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e995f803-3996-491e-8e7a-a6db48cda9ee-host\") pod \"crc-debug-msdvk\" (UID: \"e995f803-3996-491e-8e7a-a6db48cda9ee\") " pod="openshift-must-gather-sv4x5/crc-debug-msdvk" Nov 25 16:08:57 crc kubenswrapper[4890]: I1125 16:08:57.151564 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e995f803-3996-491e-8e7a-a6db48cda9ee-host\") pod \"crc-debug-msdvk\" (UID: \"e995f803-3996-491e-8e7a-a6db48cda9ee\") " pod="openshift-must-gather-sv4x5/crc-debug-msdvk" Nov 25 16:08:57 crc kubenswrapper[4890]: I1125 16:08:57.151668 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wz66\" (UniqueName: \"kubernetes.io/projected/e995f803-3996-491e-8e7a-a6db48cda9ee-kube-api-access-4wz66\") pod \"crc-debug-msdvk\" (UID: \"e995f803-3996-491e-8e7a-a6db48cda9ee\") " pod="openshift-must-gather-sv4x5/crc-debug-msdvk" Nov 25 16:08:57 crc kubenswrapper[4890]: I1125 16:08:57.151728 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e995f803-3996-491e-8e7a-a6db48cda9ee-host\") pod \"crc-debug-msdvk\" (UID: \"e995f803-3996-491e-8e7a-a6db48cda9ee\") " pod="openshift-must-gather-sv4x5/crc-debug-msdvk" Nov 25 16:08:57 crc kubenswrapper[4890]: I1125 16:08:57.172412 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wz66\" (UniqueName: \"kubernetes.io/projected/e995f803-3996-491e-8e7a-a6db48cda9ee-kube-api-access-4wz66\") pod \"crc-debug-msdvk\" (UID: \"e995f803-3996-491e-8e7a-a6db48cda9ee\") " pod="openshift-must-gather-sv4x5/crc-debug-msdvk" Nov 25 16:08:57 crc kubenswrapper[4890]: I1125 16:08:57.268615 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sv4x5/crc-debug-msdvk" Nov 25 16:08:57 crc kubenswrapper[4890]: W1125 16:08:57.298724 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode995f803_3996_491e_8e7a_a6db48cda9ee.slice/crio-96732b5b8042d79e575cad699725dfe1d607d9bd8722695546bba83a4acfd553 WatchSource:0}: Error finding container 96732b5b8042d79e575cad699725dfe1d607d9bd8722695546bba83a4acfd553: Status 404 returned error can't find the container with id 96732b5b8042d79e575cad699725dfe1d607d9bd8722695546bba83a4acfd553 Nov 25 16:08:57 crc kubenswrapper[4890]: I1125 16:08:57.602931 4890 generic.go:334] "Generic (PLEG): container finished" podID="e995f803-3996-491e-8e7a-a6db48cda9ee" containerID="5f32c3faa9c266cf2254e69c91dfa94c7c3798329f7a3618df17493a5adb1b6e" exitCode=0 Nov 25 16:08:57 crc kubenswrapper[4890]: I1125 16:08:57.603019 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-sv4x5/crc-debug-msdvk" event={"ID":"e995f803-3996-491e-8e7a-a6db48cda9ee","Type":"ContainerDied","Data":"5f32c3faa9c266cf2254e69c91dfa94c7c3798329f7a3618df17493a5adb1b6e"} Nov 25 16:08:57 crc kubenswrapper[4890]: I1125 16:08:57.603462 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-sv4x5/crc-debug-msdvk" event={"ID":"e995f803-3996-491e-8e7a-a6db48cda9ee","Type":"ContainerStarted","Data":"96732b5b8042d79e575cad699725dfe1d607d9bd8722695546bba83a4acfd553"} Nov 25 16:08:57 crc kubenswrapper[4890]: I1125 16:08:57.644146 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-sv4x5/crc-debug-msdvk"] Nov 25 16:08:57 crc kubenswrapper[4890]: I1125 16:08:57.653924 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-sv4x5/crc-debug-msdvk"] Nov 25 16:08:59 crc kubenswrapper[4890]: I1125 16:08:59.321146 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sv4x5/crc-debug-msdvk" Nov 25 16:08:59 crc kubenswrapper[4890]: I1125 16:08:59.398789 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e995f803-3996-491e-8e7a-a6db48cda9ee-host\") pod \"e995f803-3996-491e-8e7a-a6db48cda9ee\" (UID: \"e995f803-3996-491e-8e7a-a6db48cda9ee\") " Nov 25 16:08:59 crc kubenswrapper[4890]: I1125 16:08:59.398949 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e995f803-3996-491e-8e7a-a6db48cda9ee-host" (OuterVolumeSpecName: "host") pod "e995f803-3996-491e-8e7a-a6db48cda9ee" (UID: "e995f803-3996-491e-8e7a-a6db48cda9ee"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 16:08:59 crc kubenswrapper[4890]: I1125 16:08:59.399303 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4wz66\" (UniqueName: \"kubernetes.io/projected/e995f803-3996-491e-8e7a-a6db48cda9ee-kube-api-access-4wz66\") pod \"e995f803-3996-491e-8e7a-a6db48cda9ee\" (UID: \"e995f803-3996-491e-8e7a-a6db48cda9ee\") " Nov 25 16:08:59 crc kubenswrapper[4890]: I1125 16:08:59.400292 4890 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e995f803-3996-491e-8e7a-a6db48cda9ee-host\") on node \"crc\" DevicePath \"\"" Nov 25 16:08:59 crc kubenswrapper[4890]: I1125 16:08:59.406693 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e995f803-3996-491e-8e7a-a6db48cda9ee-kube-api-access-4wz66" (OuterVolumeSpecName: "kube-api-access-4wz66") pod "e995f803-3996-491e-8e7a-a6db48cda9ee" (UID: "e995f803-3996-491e-8e7a-a6db48cda9ee"). InnerVolumeSpecName "kube-api-access-4wz66". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:08:59 crc kubenswrapper[4890]: I1125 16:08:59.502511 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4wz66\" (UniqueName: \"kubernetes.io/projected/e995f803-3996-491e-8e7a-a6db48cda9ee-kube-api-access-4wz66\") on node \"crc\" DevicePath \"\"" Nov 25 16:08:59 crc kubenswrapper[4890]: I1125 16:08:59.626514 4890 scope.go:117] "RemoveContainer" containerID="5f32c3faa9c266cf2254e69c91dfa94c7c3798329f7a3618df17493a5adb1b6e" Nov 25 16:08:59 crc kubenswrapper[4890]: I1125 16:08:59.626550 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sv4x5/crc-debug-msdvk" Nov 25 16:09:00 crc kubenswrapper[4890]: I1125 16:09:00.192291 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e995f803-3996-491e-8e7a-a6db48cda9ee" path="/var/lib/kubelet/pods/e995f803-3996-491e-8e7a-a6db48cda9ee/volumes" Nov 25 16:09:15 crc kubenswrapper[4890]: I1125 16:09:15.082285 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7d4dcdcc94-g9drz_117b4c5d-110a-472c-adc1-80846e0d37cb/barbican-api/0.log" Nov 25 16:09:15 crc kubenswrapper[4890]: I1125 16:09:15.201629 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7d4dcdcc94-g9drz_117b4c5d-110a-472c-adc1-80846e0d37cb/barbican-api-log/0.log" Nov 25 16:09:15 crc kubenswrapper[4890]: I1125 16:09:15.299410 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-54dc864df6-phkw8_0b033caa-b23d-43c7-b7d4-a5465d13588b/barbican-keystone-listener/0.log" Nov 25 16:09:15 crc kubenswrapper[4890]: I1125 16:09:15.355976 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-54dc864df6-phkw8_0b033caa-b23d-43c7-b7d4-a5465d13588b/barbican-keystone-listener-log/0.log" Nov 25 16:09:15 crc kubenswrapper[4890]: I1125 16:09:15.532581 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-77c9dbb88c-b285m_56f2ab61-294c-4099-99d8-e5177128adec/barbican-worker/0.log" Nov 25 16:09:15 crc kubenswrapper[4890]: I1125 16:09:15.540969 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-77c9dbb88c-b285m_56f2ab61-294c-4099-99d8-e5177128adec/barbican-worker-log/0.log" Nov 25 16:09:15 crc kubenswrapper[4890]: I1125 16:09:15.798456 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-ddcgt_a2e80fb3-23be-4b01-b663-554207d5d538/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 16:09:15 crc kubenswrapper[4890]: I1125 16:09:15.809848 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f27f06e7-28d2-403a-92c8-b1d858d90619/ceilometer-central-agent/0.log" Nov 25 16:09:15 crc kubenswrapper[4890]: I1125 16:09:15.903236 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f27f06e7-28d2-403a-92c8-b1d858d90619/ceilometer-notification-agent/0.log" Nov 25 16:09:16 crc kubenswrapper[4890]: I1125 16:09:16.040838 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f27f06e7-28d2-403a-92c8-b1d858d90619/proxy-httpd/0.log" Nov 25 16:09:16 crc kubenswrapper[4890]: I1125 16:09:16.162937 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f27f06e7-28d2-403a-92c8-b1d858d90619/sg-core/0.log" Nov 25 16:09:16 crc kubenswrapper[4890]: I1125 16:09:16.249746 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_309052f6-3a50-436c-89e0-b97d5454b935/cinder-api/0.log" Nov 25 16:09:16 crc kubenswrapper[4890]: I1125 16:09:16.335637 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_309052f6-3a50-436c-89e0-b97d5454b935/cinder-api-log/0.log" Nov 25 16:09:16 crc kubenswrapper[4890]: I1125 16:09:16.532267 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_bb86ab7c-e51d-46ad-a1d6-6a7f5ebabc41/probe/0.log" Nov 25 16:09:16 crc kubenswrapper[4890]: I1125 16:09:16.545668 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_bb86ab7c-e51d-46ad-a1d6-6a7f5ebabc41/cinder-scheduler/0.log" Nov 25 16:09:16 crc kubenswrapper[4890]: I1125 16:09:16.715688 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-9h2zh_35c7409b-9251-4ebb-9a04-8de78ea0fde0/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 16:09:16 crc kubenswrapper[4890]: I1125 16:09:16.893008 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-wdv5b_2875b0db-cd24-42e1-91b7-3308bec9d7d1/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 16:09:17 crc kubenswrapper[4890]: I1125 16:09:17.188515 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-j8b46_2d2eef50-1148-4cd4-8b76-67dc29bf8846/init/0.log" Nov 25 16:09:17 crc kubenswrapper[4890]: I1125 16:09:17.403200 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-sgklf_ef539e05-0227-4553-b6a8-0a70f5eadf8c/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 16:09:17 crc kubenswrapper[4890]: I1125 16:09:17.414891 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-j8b46_2d2eef50-1148-4cd4-8b76-67dc29bf8846/init/0.log" Nov 25 16:09:17 crc kubenswrapper[4890]: I1125 16:09:17.459920 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-j8b46_2d2eef50-1148-4cd4-8b76-67dc29bf8846/dnsmasq-dns/0.log" Nov 25 16:09:17 crc kubenswrapper[4890]: I1125 16:09:17.663198 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_9c617481-223e-4a84-8ecc-087c98dbd459/glance-httpd/0.log" Nov 25 16:09:17 crc kubenswrapper[4890]: I1125 16:09:17.740209 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_9c617481-223e-4a84-8ecc-087c98dbd459/glance-log/0.log" Nov 25 16:09:17 crc kubenswrapper[4890]: I1125 16:09:17.936550 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_31454d1f-69db-4da1-a219-8c8ded147188/glance-httpd/0.log" Nov 25 16:09:17 crc kubenswrapper[4890]: I1125 16:09:17.981639 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_31454d1f-69db-4da1-a219-8c8ded147188/glance-log/0.log" Nov 25 16:09:18 crc kubenswrapper[4890]: I1125 16:09:18.227290 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7d75bbdb6d-5dvcn_2bf94c7c-4031-4868-bb7a-39c84abc96ea/horizon/0.log" Nov 25 16:09:18 crc kubenswrapper[4890]: I1125 16:09:18.308090 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-j96lg_2a24b42b-12c8-43bc-9dc7-f07991224497/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 16:09:18 crc kubenswrapper[4890]: I1125 16:09:18.607174 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7d75bbdb6d-5dvcn_2bf94c7c-4031-4868-bb7a-39c84abc96ea/horizon-log/0.log" Nov 25 16:09:18 crc kubenswrapper[4890]: I1125 16:09:18.638401 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-nn65s_6679d2f8-52a9-4539-bb89-f4c9c986bc6b/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 16:09:18 crc kubenswrapper[4890]: I1125 16:09:18.878870 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29401441-rfltq_f146ff30-8207-40d9-a579-ef8f314c41e4/keystone-cron/0.log" Nov 25 16:09:19 crc kubenswrapper[4890]: I1125 16:09:19.023397 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-64fdb55f6-22c8w_e127320b-04a5-48f0-b434-ba6bfe1c2696/keystone-api/0.log" Nov 25 16:09:19 crc kubenswrapper[4890]: I1125 16:09:19.358712 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_a47203e9-a88a-42f7-86e1-164c6ad532b5/kube-state-metrics/3.log" Nov 25 16:09:19 crc kubenswrapper[4890]: I1125 16:09:19.366705 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_a47203e9-a88a-42f7-86e1-164c6ad532b5/kube-state-metrics/2.log" Nov 25 16:09:19 crc kubenswrapper[4890]: I1125 16:09:19.470213 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-4fkzh_ca9e616d-0ce2-4784-945c-9e99ba4f437b/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 16:09:19 crc kubenswrapper[4890]: I1125 16:09:19.879296 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-59d57fcc85-nkqp6_5baff989-9642-4963-816c-83904cab7811/neutron-api/0.log" Nov 25 16:09:19 crc kubenswrapper[4890]: I1125 16:09:19.940925 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-59d57fcc85-nkqp6_5baff989-9642-4963-816c-83904cab7811/neutron-httpd/0.log" Nov 25 16:09:20 crc kubenswrapper[4890]: I1125 16:09:20.012737 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-dvxc6_cb4c455d-9d31-430d-b225-9bdb0b9690e0/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 16:09:21 crc kubenswrapper[4890]: I1125 16:09:21.449809 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_3c6b60ed-19f6-472a-92d6-2a1e06c92ea6/nova-api-log/0.log" Nov 25 16:09:21 crc kubenswrapper[4890]: I1125 16:09:21.454398 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_f7929bde-463b-4dff-b685-db4bf8368909/nova-cell0-conductor-conductor/0.log" Nov 25 16:09:21 crc kubenswrapper[4890]: I1125 16:09:21.866404 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_59c66e1f-c917-441e-99d4-10980ac6891e/nova-cell1-conductor-conductor/0.log" Nov 25 16:09:22 crc kubenswrapper[4890]: I1125 16:09:22.009154 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_85887546-bd08-4f9d-826a-ec30d6dd3dee/nova-cell1-novncproxy-novncproxy/0.log" Nov 25 16:09:22 crc kubenswrapper[4890]: I1125 16:09:22.030333 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_3c6b60ed-19f6-472a-92d6-2a1e06c92ea6/nova-api-api/0.log" Nov 25 16:09:22 crc kubenswrapper[4890]: I1125 16:09:22.164435 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-49nlj_e890bd11-32e9-4d73-9dd0-ae5d69a1fe58/nova-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 16:09:22 crc kubenswrapper[4890]: I1125 16:09:22.462143 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_ab6971fb-880d-4e68-96f6-bb23533105cd/nova-metadata-log/0.log" Nov 25 16:09:23 crc kubenswrapper[4890]: I1125 16:09:23.014918 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab/mysql-bootstrap/0.log" Nov 25 16:09:23 crc kubenswrapper[4890]: I1125 16:09:23.187751 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_8df24e08-7452-4c77-9360-b555fdecf971/nova-scheduler-scheduler/0.log" Nov 25 16:09:23 crc kubenswrapper[4890]: I1125 16:09:23.264595 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab/mysql-bootstrap/0.log" Nov 25 16:09:23 crc kubenswrapper[4890]: I1125 16:09:23.353081 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab/galera/0.log" Nov 25 16:09:23 crc kubenswrapper[4890]: I1125 16:09:23.591924 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_90881a25-bc18-4b43-a226-a44e3165cb3a/mysql-bootstrap/0.log" Nov 25 16:09:23 crc kubenswrapper[4890]: I1125 16:09:23.799573 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_90881a25-bc18-4b43-a226-a44e3165cb3a/mysql-bootstrap/0.log" Nov 25 16:09:23 crc kubenswrapper[4890]: I1125 16:09:23.813188 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_90881a25-bc18-4b43-a226-a44e3165cb3a/galera/0.log" Nov 25 16:09:24 crc kubenswrapper[4890]: I1125 16:09:24.017050 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_ab6971fb-880d-4e68-96f6-bb23533105cd/nova-metadata-metadata/0.log" Nov 25 16:09:24 crc kubenswrapper[4890]: I1125 16:09:24.242383 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_1d4df90d-315a-4483-a6e5-c6e5b322db54/openstackclient/0.log" Nov 25 16:09:24 crc kubenswrapper[4890]: I1125 16:09:24.352304 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-v59z7_f282d61d-d8c2-4933-99c1-d95a9d4f40e2/openstack-network-exporter/0.log" Nov 25 16:09:24 crc kubenswrapper[4890]: I1125 16:09:24.517932 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-nrbtc_524ff107-da15-4e2f-a0c8-acca0e91b27a/ovn-controller/0.log" Nov 25 16:09:24 crc kubenswrapper[4890]: I1125 16:09:24.737150 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-5tg6b_289df7bb-0b8e-4411-ab2f-935629ed4019/ovsdb-server-init/0.log" Nov 25 16:09:24 crc kubenswrapper[4890]: I1125 16:09:24.960796 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-5tg6b_289df7bb-0b8e-4411-ab2f-935629ed4019/ovsdb-server-init/0.log" Nov 25 16:09:24 crc kubenswrapper[4890]: I1125 16:09:24.967558 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-5tg6b_289df7bb-0b8e-4411-ab2f-935629ed4019/ovsdb-server/0.log" Nov 25 16:09:25 crc kubenswrapper[4890]: I1125 16:09:25.068278 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-5tg6b_289df7bb-0b8e-4411-ab2f-935629ed4019/ovs-vswitchd/0.log" Nov 25 16:09:25 crc kubenswrapper[4890]: I1125 16:09:25.318995 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-llz86_995132c2-f199-4d15-b3ef-3c71fed454f5/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 16:09:25 crc kubenswrapper[4890]: I1125 16:09:25.347228 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_effbba39-45db-4d3b-a883-03f38d0b49f8/openstack-network-exporter/0.log" Nov 25 16:09:25 crc kubenswrapper[4890]: I1125 16:09:25.384588 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_effbba39-45db-4d3b-a883-03f38d0b49f8/ovn-northd/0.log" Nov 25 16:09:25 crc kubenswrapper[4890]: I1125 16:09:25.528618 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_49e51424-feba-4038-88a6-1cd56637984c/openstack-network-exporter/0.log" Nov 25 16:09:25 crc kubenswrapper[4890]: I1125 16:09:25.606653 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_49e51424-feba-4038-88a6-1cd56637984c/ovsdbserver-nb/0.log" Nov 25 16:09:25 crc kubenswrapper[4890]: I1125 16:09:25.818005 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_c9667a27-2ed9-4b9a-a75a-4901a93cab5e/openstack-network-exporter/0.log" Nov 25 16:09:25 crc kubenswrapper[4890]: I1125 16:09:25.868120 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_c9667a27-2ed9-4b9a-a75a-4901a93cab5e/ovsdbserver-sb/0.log" Nov 25 16:09:26 crc kubenswrapper[4890]: I1125 16:09:26.121448 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-74c777b4cb-m7k57_46d00836-60a2-459f-b4c9-ed2d5e3b1fb1/placement-api/0.log" Nov 25 16:09:26 crc kubenswrapper[4890]: I1125 16:09:26.292097 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_d08ea973-c3f2-4192-8bfc-b025f0c5391b/setup-container/0.log" Nov 25 16:09:26 crc kubenswrapper[4890]: I1125 16:09:26.303081 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-74c777b4cb-m7k57_46d00836-60a2-459f-b4c9-ed2d5e3b1fb1/placement-log/0.log" Nov 25 16:09:26 crc kubenswrapper[4890]: I1125 16:09:26.794762 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_d08ea973-c3f2-4192-8bfc-b025f0c5391b/rabbitmq/0.log" Nov 25 16:09:26 crc kubenswrapper[4890]: I1125 16:09:26.822536 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_9bab37b3-681b-4cd5-9fee-503afa0d6d22/setup-container/0.log" Nov 25 16:09:26 crc kubenswrapper[4890]: I1125 16:09:26.832260 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_d08ea973-c3f2-4192-8bfc-b025f0c5391b/setup-container/0.log" Nov 25 16:09:27 crc kubenswrapper[4890]: I1125 16:09:27.160019 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_9bab37b3-681b-4cd5-9fee-503afa0d6d22/setup-container/0.log" Nov 25 16:09:27 crc kubenswrapper[4890]: I1125 16:09:27.161796 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_9bab37b3-681b-4cd5-9fee-503afa0d6d22/rabbitmq/0.log" Nov 25 16:09:27 crc kubenswrapper[4890]: I1125 16:09:27.225240 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-7bjzl_2ba10718-5ae6-4670-bbc6-b02e9c5562e0/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 16:09:27 crc kubenswrapper[4890]: I1125 16:09:27.406784 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-w4tcq_4f0ccf80-fad3-4920-83e3-1c68111ed1ac/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 16:09:27 crc kubenswrapper[4890]: I1125 16:09:27.475695 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-cr58b_6d15a3f2-bbd4-43ae-b620-93251f74369e/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 16:09:27 crc kubenswrapper[4890]: I1125 16:09:27.759362 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-59sbs_788b5bd1-adf2-4416-a8f0-975895de101c/ssh-known-hosts-edpm-deployment/0.log" Nov 25 16:09:27 crc kubenswrapper[4890]: I1125 16:09:27.794747 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-l94p8_90b03f38-bd28-4c62-ae05-5cc72f4e6ee7/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 16:09:28 crc kubenswrapper[4890]: I1125 16:09:28.104685 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5f45cd8b8f-8wpds_66cfb07e-ac1a-4aee-8fbd-7871e2e30d88/proxy-server/0.log" Nov 25 16:09:28 crc kubenswrapper[4890]: I1125 16:09:28.184913 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5f45cd8b8f-8wpds_66cfb07e-ac1a-4aee-8fbd-7871e2e30d88/proxy-httpd/0.log" Nov 25 16:09:28 crc kubenswrapper[4890]: I1125 16:09:28.417809 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6517688d-0723-4809-ba51-bcaf48524157/account-auditor/0.log" Nov 25 16:09:28 crc kubenswrapper[4890]: I1125 16:09:28.419145 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-vs9dw_cf34ad11-7dd6-4572-b3d2-4c14127a1d76/swift-ring-rebalance/0.log" Nov 25 16:09:28 crc kubenswrapper[4890]: I1125 16:09:28.476356 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6517688d-0723-4809-ba51-bcaf48524157/account-reaper/0.log" Nov 25 16:09:28 crc kubenswrapper[4890]: I1125 16:09:28.727251 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6517688d-0723-4809-ba51-bcaf48524157/container-auditor/0.log" Nov 25 16:09:28 crc kubenswrapper[4890]: I1125 16:09:28.742375 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6517688d-0723-4809-ba51-bcaf48524157/account-replicator/0.log" Nov 25 16:09:28 crc kubenswrapper[4890]: I1125 16:09:28.759310 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6517688d-0723-4809-ba51-bcaf48524157/container-replicator/0.log" Nov 25 16:09:28 crc kubenswrapper[4890]: I1125 16:09:28.823530 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6517688d-0723-4809-ba51-bcaf48524157/account-server/0.log" Nov 25 16:09:29 crc kubenswrapper[4890]: I1125 16:09:29.021571 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6517688d-0723-4809-ba51-bcaf48524157/object-auditor/0.log" Nov 25 16:09:29 crc kubenswrapper[4890]: I1125 16:09:29.269238 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6517688d-0723-4809-ba51-bcaf48524157/object-expirer/0.log" Nov 25 16:09:29 crc kubenswrapper[4890]: I1125 16:09:29.273958 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6517688d-0723-4809-ba51-bcaf48524157/container-server/0.log" Nov 25 16:09:29 crc kubenswrapper[4890]: I1125 16:09:29.282849 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6517688d-0723-4809-ba51-bcaf48524157/container-updater/0.log" Nov 25 16:09:29 crc kubenswrapper[4890]: I1125 16:09:29.519563 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6517688d-0723-4809-ba51-bcaf48524157/object-updater/0.log" Nov 25 16:09:29 crc kubenswrapper[4890]: I1125 16:09:29.533695 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6517688d-0723-4809-ba51-bcaf48524157/object-server/0.log" Nov 25 16:09:29 crc kubenswrapper[4890]: I1125 16:09:29.541076 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6517688d-0723-4809-ba51-bcaf48524157/object-replicator/0.log" Nov 25 16:09:29 crc kubenswrapper[4890]: I1125 16:09:29.560274 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6517688d-0723-4809-ba51-bcaf48524157/rsync/0.log" Nov 25 16:09:29 crc kubenswrapper[4890]: I1125 16:09:29.813351 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6517688d-0723-4809-ba51-bcaf48524157/swift-recon-cron/0.log" Nov 25 16:09:29 crc kubenswrapper[4890]: I1125 16:09:29.866324 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-jm99p_588df9b2-9cfb-449a-9dca-17912f88a4c6/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 16:09:30 crc kubenswrapper[4890]: I1125 16:09:30.108693 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_8b652448-2962-42f9-b0b5-e551b50edc1e/tempest-tests-tempest-tests-runner/0.log" Nov 25 16:09:30 crc kubenswrapper[4890]: I1125 16:09:30.113849 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_a76905fc-a86c-4ed6-982d-0a50bada561f/test-operator-logs-container/0.log" Nov 25 16:09:30 crc kubenswrapper[4890]: I1125 16:09:30.353034 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-6vbxl_8394f488-cfd0-4fdc-aa22-96fe01aec73e/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 16:09:44 crc kubenswrapper[4890]: I1125 16:09:44.196816 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_c5630332-c335-4b9a-8486-e845c6baaa9c/memcached/0.log" Nov 25 16:09:58 crc kubenswrapper[4890]: I1125 16:09:58.452605 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9b547be2eac4c3e8f15e2800f340fb44715065f09b75b108844ab9dee67zk4x_4050c336-b097-49f8-bc65-17bb8d191fa4/util/0.log" Nov 25 16:09:59 crc kubenswrapper[4890]: I1125 16:09:59.087979 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9b547be2eac4c3e8f15e2800f340fb44715065f09b75b108844ab9dee67zk4x_4050c336-b097-49f8-bc65-17bb8d191fa4/util/0.log" Nov 25 16:09:59 crc kubenswrapper[4890]: I1125 16:09:59.128380 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9b547be2eac4c3e8f15e2800f340fb44715065f09b75b108844ab9dee67zk4x_4050c336-b097-49f8-bc65-17bb8d191fa4/pull/0.log" Nov 25 16:09:59 crc kubenswrapper[4890]: I1125 16:09:59.136274 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9b547be2eac4c3e8f15e2800f340fb44715065f09b75b108844ab9dee67zk4x_4050c336-b097-49f8-bc65-17bb8d191fa4/pull/0.log" Nov 25 16:09:59 crc kubenswrapper[4890]: I1125 16:09:59.294828 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9b547be2eac4c3e8f15e2800f340fb44715065f09b75b108844ab9dee67zk4x_4050c336-b097-49f8-bc65-17bb8d191fa4/util/0.log" Nov 25 16:09:59 crc kubenswrapper[4890]: I1125 16:09:59.375189 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9b547be2eac4c3e8f15e2800f340fb44715065f09b75b108844ab9dee67zk4x_4050c336-b097-49f8-bc65-17bb8d191fa4/extract/0.log" Nov 25 16:09:59 crc kubenswrapper[4890]: I1125 16:09:59.385065 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9b547be2eac4c3e8f15e2800f340fb44715065f09b75b108844ab9dee67zk4x_4050c336-b097-49f8-bc65-17bb8d191fa4/pull/0.log" Nov 25 16:09:59 crc kubenswrapper[4890]: I1125 16:09:59.527882 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-86dc4d89c8-pmmf2_d175a35b-8b9b-49c0-8fc2-02ef62f131d9/kube-rbac-proxy/0.log" Nov 25 16:09:59 crc kubenswrapper[4890]: I1125 16:09:59.574637 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-86dc4d89c8-pmmf2_d175a35b-8b9b-49c0-8fc2-02ef62f131d9/manager/2.log" Nov 25 16:09:59 crc kubenswrapper[4890]: I1125 16:09:59.601009 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-86dc4d89c8-pmmf2_d175a35b-8b9b-49c0-8fc2-02ef62f131d9/manager/3.log" Nov 25 16:09:59 crc kubenswrapper[4890]: I1125 16:09:59.719289 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-79856dc55c-jcw7x_0bdbada3-e08c-4f8a-bd55-bac955d4370d/kube-rbac-proxy/0.log" Nov 25 16:09:59 crc kubenswrapper[4890]: I1125 16:09:59.780469 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-79856dc55c-jcw7x_0bdbada3-e08c-4f8a-bd55-bac955d4370d/manager/2.log" Nov 25 16:09:59 crc kubenswrapper[4890]: I1125 16:09:59.785376 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-79856dc55c-jcw7x_0bdbada3-e08c-4f8a-bd55-bac955d4370d/manager/3.log" Nov 25 16:09:59 crc kubenswrapper[4890]: I1125 16:09:59.911902 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-7d695c9b56-stwjm_77a5e8f9-92ab-48b5-be19-29a7a8e5df49/kube-rbac-proxy/0.log" Nov 25 16:09:59 crc kubenswrapper[4890]: I1125 16:09:59.956614 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-7d695c9b56-stwjm_77a5e8f9-92ab-48b5-be19-29a7a8e5df49/manager/3.log" Nov 25 16:09:59 crc kubenswrapper[4890]: I1125 16:09:59.998994 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-7d695c9b56-stwjm_77a5e8f9-92ab-48b5-be19-29a7a8e5df49/manager/2.log" Nov 25 16:10:00 crc kubenswrapper[4890]: I1125 16:10:00.175629 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-68b95954c9-lfnwl_076c2a0e-7fdb-42e1-ab73-e296bba2b7a9/kube-rbac-proxy/0.log" Nov 25 16:10:00 crc kubenswrapper[4890]: I1125 16:10:00.201839 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-68b95954c9-lfnwl_076c2a0e-7fdb-42e1-ab73-e296bba2b7a9/manager/3.log" Nov 25 16:10:00 crc kubenswrapper[4890]: I1125 16:10:00.218508 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-68b95954c9-lfnwl_076c2a0e-7fdb-42e1-ab73-e296bba2b7a9/manager/2.log" Nov 25 16:10:00 crc kubenswrapper[4890]: I1125 16:10:00.418998 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-774b86978c-xvfgd_72c26c5b-3370-4994-a5f0-82128c62c2df/kube-rbac-proxy/0.log" Nov 25 16:10:00 crc kubenswrapper[4890]: I1125 16:10:00.436430 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-774b86978c-xvfgd_72c26c5b-3370-4994-a5f0-82128c62c2df/manager/3.log" Nov 25 16:10:00 crc kubenswrapper[4890]: I1125 16:10:00.494922 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-774b86978c-xvfgd_72c26c5b-3370-4994-a5f0-82128c62c2df/manager/2.log" Nov 25 16:10:00 crc kubenswrapper[4890]: I1125 16:10:00.650863 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c9694994-j6pqs_c528348f-0907-423a-a0c2-f15cf18ff99a/manager/3.log" Nov 25 16:10:00 crc kubenswrapper[4890]: I1125 16:10:00.653321 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c9694994-j6pqs_c528348f-0907-423a-a0c2-f15cf18ff99a/kube-rbac-proxy/0.log" Nov 25 16:10:00 crc kubenswrapper[4890]: I1125 16:10:00.719119 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c9694994-j6pqs_c528348f-0907-423a-a0c2-f15cf18ff99a/manager/2.log" Nov 25 16:10:00 crc kubenswrapper[4890]: I1125 16:10:00.849022 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-d5cc86f4b-stflb_7e576602-19f5-4656-84dd-5ab80abca895/kube-rbac-proxy/0.log" Nov 25 16:10:00 crc kubenswrapper[4890]: I1125 16:10:00.863451 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-d5cc86f4b-stflb_7e576602-19f5-4656-84dd-5ab80abca895/manager/3.log" Nov 25 16:10:00 crc kubenswrapper[4890]: I1125 16:10:00.913219 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-d5cc86f4b-stflb_7e576602-19f5-4656-84dd-5ab80abca895/manager/2.log" Nov 25 16:10:01 crc kubenswrapper[4890]: I1125 16:10:01.075967 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5bfcdc958c-hxzf4_679b3bf3-1ebc-4d19-9ad2-588bd41565dd/manager/3.log" Nov 25 16:10:01 crc kubenswrapper[4890]: I1125 16:10:01.093421 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5bfcdc958c-hxzf4_679b3bf3-1ebc-4d19-9ad2-588bd41565dd/kube-rbac-proxy/0.log" Nov 25 16:10:01 crc kubenswrapper[4890]: I1125 16:10:01.126563 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5bfcdc958c-hxzf4_679b3bf3-1ebc-4d19-9ad2-588bd41565dd/manager/2.log" Nov 25 16:10:01 crc kubenswrapper[4890]: I1125 16:10:01.303580 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-748dc6576f-rs6hx_a5d8a8ba-50c0-4fb6-823f-f28cf0d0f125/kube-rbac-proxy/0.log" Nov 25 16:10:01 crc kubenswrapper[4890]: I1125 16:10:01.349625 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-748dc6576f-rs6hx_a5d8a8ba-50c0-4fb6-823f-f28cf0d0f125/manager/2.log" Nov 25 16:10:01 crc kubenswrapper[4890]: I1125 16:10:01.371378 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-748dc6576f-rs6hx_a5d8a8ba-50c0-4fb6-823f-f28cf0d0f125/manager/3.log" Nov 25 16:10:01 crc kubenswrapper[4890]: I1125 16:10:01.492730 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58bb8d67cc-slpt4_f272410f-230d-43f2-946e-e34e4d9c05ea/kube-rbac-proxy/0.log" Nov 25 16:10:01 crc kubenswrapper[4890]: I1125 16:10:01.529519 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58bb8d67cc-slpt4_f272410f-230d-43f2-946e-e34e4d9c05ea/manager/3.log" Nov 25 16:10:01 crc kubenswrapper[4890]: I1125 16:10:01.567710 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58bb8d67cc-slpt4_f272410f-230d-43f2-946e-e34e4d9c05ea/manager/2.log" Nov 25 16:10:01 crc kubenswrapper[4890]: I1125 16:10:01.743335 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-cb6c4fdb7-wvjw7_03353c8b-91f4-41ae-afc1-c522b1c73bb4/kube-rbac-proxy/0.log" Nov 25 16:10:01 crc kubenswrapper[4890]: I1125 16:10:01.756200 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-cb6c4fdb7-wvjw7_03353c8b-91f4-41ae-afc1-c522b1c73bb4/manager/3.log" Nov 25 16:10:01 crc kubenswrapper[4890]: I1125 16:10:01.792268 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-cb6c4fdb7-wvjw7_03353c8b-91f4-41ae-afc1-c522b1c73bb4/manager/2.log" Nov 25 16:10:01 crc kubenswrapper[4890]: I1125 16:10:01.971374 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-7c57c8bbc4-829nk_4364eec5-444a-41f6-af1c-fe16b24bd044/kube-rbac-proxy/0.log" Nov 25 16:10:02 crc kubenswrapper[4890]: I1125 16:10:02.008879 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-7c57c8bbc4-829nk_4364eec5-444a-41f6-af1c-fe16b24bd044/manager/3.log" Nov 25 16:10:02 crc kubenswrapper[4890]: I1125 16:10:02.107103 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-7c57c8bbc4-829nk_4364eec5-444a-41f6-af1c-fe16b24bd044/manager/2.log" Nov 25 16:10:02 crc kubenswrapper[4890]: I1125 16:10:02.299691 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-6jq7q_0769cde4-b49f-41c5-b720-e0dbc667eb4a/kube-rbac-proxy/0.log" Nov 25 16:10:02 crc kubenswrapper[4890]: I1125 16:10:02.308118 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-6jq7q_0769cde4-b49f-41c5-b720-e0dbc667eb4a/manager/2.log" Nov 25 16:10:02 crc kubenswrapper[4890]: I1125 16:10:02.365030 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-6jq7q_0769cde4-b49f-41c5-b720-e0dbc667eb4a/manager/3.log" Nov 25 16:10:02 crc kubenswrapper[4890]: I1125 16:10:02.444825 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-fd75fd47d-56bcr_a9efeabb-bf1c-4de2-8322-961b284e95cc/kube-rbac-proxy/0.log" Nov 25 16:10:02 crc kubenswrapper[4890]: I1125 16:10:02.515882 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-fd75fd47d-56bcr_a9efeabb-bf1c-4de2-8322-961b284e95cc/manager/3.log" Nov 25 16:10:02 crc kubenswrapper[4890]: I1125 16:10:02.540208 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-fd75fd47d-56bcr_a9efeabb-bf1c-4de2-8322-961b284e95cc/manager/2.log" Nov 25 16:10:02 crc kubenswrapper[4890]: I1125 16:10:02.630671 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-544b9bb9-5lr7r_10534635-efe1-4e20-8c9b-9cbea73fbe87/kube-rbac-proxy/0.log" Nov 25 16:10:02 crc kubenswrapper[4890]: I1125 16:10:02.692297 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-544b9bb9-5lr7r_10534635-efe1-4e20-8c9b-9cbea73fbe87/manager/1.log" Nov 25 16:10:02 crc kubenswrapper[4890]: I1125 16:10:02.745907 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-544b9bb9-5lr7r_10534635-efe1-4e20-8c9b-9cbea73fbe87/manager/0.log" Nov 25 16:10:02 crc kubenswrapper[4890]: I1125 16:10:02.815984 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-68587559f4-9b4cf_3e496c3d-9485-44bc-b3b1-dc1682a88434/manager/2.log" Nov 25 16:10:02 crc kubenswrapper[4890]: I1125 16:10:02.960671 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6d88ccc4fc-9smnn_bd60efda-e354-44d0-af56-c110651bdbaf/operator/1.log" Nov 25 16:10:03 crc kubenswrapper[4890]: I1125 16:10:03.144361 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-vtg6p_bd47d0df-89c3-465a-8f11-8f0e7e1117eb/registry-server/0.log" Nov 25 16:10:03 crc kubenswrapper[4890]: I1125 16:10:03.282582 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-68587559f4-9b4cf_3e496c3d-9485-44bc-b3b1-dc1682a88434/manager/3.log" Nov 25 16:10:03 crc kubenswrapper[4890]: I1125 16:10:03.284639 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6d88ccc4fc-9smnn_bd60efda-e354-44d0-af56-c110651bdbaf/operator/0.log" Nov 25 16:10:03 crc kubenswrapper[4890]: I1125 16:10:03.353408 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-66cf5c67ff-vx75t_29f250db-2a96-41ad-bb8f-5b503a9288d8/kube-rbac-proxy/0.log" Nov 25 16:10:03 crc kubenswrapper[4890]: I1125 16:10:03.420445 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-66cf5c67ff-vx75t_29f250db-2a96-41ad-bb8f-5b503a9288d8/manager/3.log" Nov 25 16:10:03 crc kubenswrapper[4890]: I1125 16:10:03.470543 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-66cf5c67ff-vx75t_29f250db-2a96-41ad-bb8f-5b503a9288d8/manager/2.log" Nov 25 16:10:03 crc kubenswrapper[4890]: I1125 16:10:03.555239 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5db546f9d9-nshw7_9f0e3cf9-dcc0-405f-a2de-9148844ec3ae/kube-rbac-proxy/0.log" Nov 25 16:10:03 crc kubenswrapper[4890]: I1125 16:10:03.556385 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5db546f9d9-nshw7_9f0e3cf9-dcc0-405f-a2de-9148844ec3ae/manager/3.log" Nov 25 16:10:03 crc kubenswrapper[4890]: I1125 16:10:03.683574 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5db546f9d9-nshw7_9f0e3cf9-dcc0-405f-a2de-9148844ec3ae/manager/2.log" Nov 25 16:10:03 crc kubenswrapper[4890]: I1125 16:10:03.732238 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-dmlx4_e26594ec-4bc2-4051-8a49-fc490bd3b6c9/operator/3.log" Nov 25 16:10:03 crc kubenswrapper[4890]: I1125 16:10:03.777848 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-dmlx4_e26594ec-4bc2-4051-8a49-fc490bd3b6c9/operator/2.log" Nov 25 16:10:03 crc kubenswrapper[4890]: I1125 16:10:03.893583 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6fdc4fcf86-dzkgt_861317e1-222d-4f14-b931-e9e7d100ebd6/kube-rbac-proxy/0.log" Nov 25 16:10:03 crc kubenswrapper[4890]: I1125 16:10:03.897849 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6fdc4fcf86-dzkgt_861317e1-222d-4f14-b931-e9e7d100ebd6/manager/3.log" Nov 25 16:10:03 crc kubenswrapper[4890]: I1125 16:10:03.947890 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6fdc4fcf86-dzkgt_861317e1-222d-4f14-b931-e9e7d100ebd6/manager/2.log" Nov 25 16:10:04 crc kubenswrapper[4890]: I1125 16:10:04.051394 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-567f98c9d-6zr4x_d50aec9d-1673-4b81-bb57-04cae418843e/kube-rbac-proxy/0.log" Nov 25 16:10:04 crc kubenswrapper[4890]: I1125 16:10:04.128400 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-567f98c9d-6zr4x_d50aec9d-1673-4b81-bb57-04cae418843e/manager/2.log" Nov 25 16:10:04 crc kubenswrapper[4890]: I1125 16:10:04.146809 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-567f98c9d-6zr4x_d50aec9d-1673-4b81-bb57-04cae418843e/manager/3.log" Nov 25 16:10:04 crc kubenswrapper[4890]: I1125 16:10:04.228912 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cb74df96-6mqt2_9637fded-765d-4037-be8f-82e59ffdfc14/kube-rbac-proxy/0.log" Nov 25 16:10:04 crc kubenswrapper[4890]: I1125 16:10:04.291912 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cb74df96-6mqt2_9637fded-765d-4037-be8f-82e59ffdfc14/manager/1.log" Nov 25 16:10:04 crc kubenswrapper[4890]: I1125 16:10:04.376013 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cb74df96-6mqt2_9637fded-765d-4037-be8f-82e59ffdfc14/manager/0.log" Nov 25 16:10:04 crc kubenswrapper[4890]: I1125 16:10:04.400790 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-864885998-62ddz_a9a69828-3f0d-4d55-a840-10c2c791444c/kube-rbac-proxy/0.log" Nov 25 16:10:04 crc kubenswrapper[4890]: I1125 16:10:04.453891 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-864885998-62ddz_a9a69828-3f0d-4d55-a840-10c2c791444c/manager/2.log" Nov 25 16:10:04 crc kubenswrapper[4890]: I1125 16:10:04.541051 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-864885998-62ddz_a9a69828-3f0d-4d55-a840-10c2c791444c/manager/1.log" Nov 25 16:10:23 crc kubenswrapper[4890]: I1125 16:10:23.412911 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-zs6w7_a3f62807-e422-4d74-82c5-81756d963b77/control-plane-machine-set-operator/0.log" Nov 25 16:10:23 crc kubenswrapper[4890]: I1125 16:10:23.621424 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-6dmh8_8b5c8177-1153-453a-9110-6023d01e3fb7/kube-rbac-proxy/0.log" Nov 25 16:10:23 crc kubenswrapper[4890]: I1125 16:10:23.633018 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-6dmh8_8b5c8177-1153-453a-9110-6023d01e3fb7/machine-api-operator/0.log" Nov 25 16:10:26 crc kubenswrapper[4890]: I1125 16:10:26.448387 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 16:10:26 crc kubenswrapper[4890]: I1125 16:10:26.448933 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 16:10:36 crc kubenswrapper[4890]: I1125 16:10:36.980919 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-9c9wk_e14d1071-6d9a-4543-b593-8e999b7dcad2/cert-manager-controller/0.log" Nov 25 16:10:36 crc kubenswrapper[4890]: I1125 16:10:36.984704 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-9c9wk_e14d1071-6d9a-4543-b593-8e999b7dcad2/cert-manager-controller/1.log" Nov 25 16:10:37 crc kubenswrapper[4890]: I1125 16:10:37.225206 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-x7dwr_d0f0970d-430a-4ea0-b18c-6af23eb40c71/cert-manager-cainjector/1.log" Nov 25 16:10:37 crc kubenswrapper[4890]: I1125 16:10:37.252199 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-x7dwr_d0f0970d-430a-4ea0-b18c-6af23eb40c71/cert-manager-cainjector/0.log" Nov 25 16:10:37 crc kubenswrapper[4890]: I1125 16:10:37.375132 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-r5kd7_3e0ccee5-4080-4973-99df-e4d6e929e471/cert-manager-webhook/0.log" Nov 25 16:10:38 crc kubenswrapper[4890]: I1125 16:10:38.792401 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nm58z"] Nov 25 16:10:38 crc kubenswrapper[4890]: E1125 16:10:38.793568 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e995f803-3996-491e-8e7a-a6db48cda9ee" containerName="container-00" Nov 25 16:10:38 crc kubenswrapper[4890]: I1125 16:10:38.793587 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="e995f803-3996-491e-8e7a-a6db48cda9ee" containerName="container-00" Nov 25 16:10:38 crc kubenswrapper[4890]: I1125 16:10:38.793868 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="e995f803-3996-491e-8e7a-a6db48cda9ee" containerName="container-00" Nov 25 16:10:38 crc kubenswrapper[4890]: I1125 16:10:38.795816 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nm58z" Nov 25 16:10:38 crc kubenswrapper[4890]: I1125 16:10:38.805937 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nm58z"] Nov 25 16:10:38 crc kubenswrapper[4890]: I1125 16:10:38.951095 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a17e0cb0-68c0-4327-b7d0-94d146abcb7e-catalog-content\") pod \"community-operators-nm58z\" (UID: \"a17e0cb0-68c0-4327-b7d0-94d146abcb7e\") " pod="openshift-marketplace/community-operators-nm58z" Nov 25 16:10:38 crc kubenswrapper[4890]: I1125 16:10:38.951241 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fwvq\" (UniqueName: \"kubernetes.io/projected/a17e0cb0-68c0-4327-b7d0-94d146abcb7e-kube-api-access-8fwvq\") pod \"community-operators-nm58z\" (UID: \"a17e0cb0-68c0-4327-b7d0-94d146abcb7e\") " pod="openshift-marketplace/community-operators-nm58z" Nov 25 16:10:38 crc kubenswrapper[4890]: I1125 16:10:38.952747 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a17e0cb0-68c0-4327-b7d0-94d146abcb7e-utilities\") pod \"community-operators-nm58z\" (UID: \"a17e0cb0-68c0-4327-b7d0-94d146abcb7e\") " pod="openshift-marketplace/community-operators-nm58z" Nov 25 16:10:39 crc kubenswrapper[4890]: I1125 16:10:39.055142 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a17e0cb0-68c0-4327-b7d0-94d146abcb7e-utilities\") pod \"community-operators-nm58z\" (UID: \"a17e0cb0-68c0-4327-b7d0-94d146abcb7e\") " pod="openshift-marketplace/community-operators-nm58z" Nov 25 16:10:39 crc kubenswrapper[4890]: I1125 16:10:39.055246 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a17e0cb0-68c0-4327-b7d0-94d146abcb7e-catalog-content\") pod \"community-operators-nm58z\" (UID: \"a17e0cb0-68c0-4327-b7d0-94d146abcb7e\") " pod="openshift-marketplace/community-operators-nm58z" Nov 25 16:10:39 crc kubenswrapper[4890]: I1125 16:10:39.055287 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fwvq\" (UniqueName: \"kubernetes.io/projected/a17e0cb0-68c0-4327-b7d0-94d146abcb7e-kube-api-access-8fwvq\") pod \"community-operators-nm58z\" (UID: \"a17e0cb0-68c0-4327-b7d0-94d146abcb7e\") " pod="openshift-marketplace/community-operators-nm58z" Nov 25 16:10:39 crc kubenswrapper[4890]: I1125 16:10:39.056209 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a17e0cb0-68c0-4327-b7d0-94d146abcb7e-utilities\") pod \"community-operators-nm58z\" (UID: \"a17e0cb0-68c0-4327-b7d0-94d146abcb7e\") " pod="openshift-marketplace/community-operators-nm58z" Nov 25 16:10:39 crc kubenswrapper[4890]: I1125 16:10:39.056430 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a17e0cb0-68c0-4327-b7d0-94d146abcb7e-catalog-content\") pod \"community-operators-nm58z\" (UID: \"a17e0cb0-68c0-4327-b7d0-94d146abcb7e\") " pod="openshift-marketplace/community-operators-nm58z" Nov 25 16:10:39 crc kubenswrapper[4890]: I1125 16:10:39.082603 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fwvq\" (UniqueName: \"kubernetes.io/projected/a17e0cb0-68c0-4327-b7d0-94d146abcb7e-kube-api-access-8fwvq\") pod \"community-operators-nm58z\" (UID: \"a17e0cb0-68c0-4327-b7d0-94d146abcb7e\") " pod="openshift-marketplace/community-operators-nm58z" Nov 25 16:10:39 crc kubenswrapper[4890]: I1125 16:10:39.128041 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nm58z" Nov 25 16:10:39 crc kubenswrapper[4890]: I1125 16:10:39.751771 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nm58z"] Nov 25 16:10:39 crc kubenswrapper[4890]: I1125 16:10:39.786257 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nm58z" event={"ID":"a17e0cb0-68c0-4327-b7d0-94d146abcb7e","Type":"ContainerStarted","Data":"f3b9dd2ab5b7cba357b9fcbe2b5c33926d856bc5e1920bd7809d994b7adcd061"} Nov 25 16:10:40 crc kubenswrapper[4890]: I1125 16:10:40.798716 4890 generic.go:334] "Generic (PLEG): container finished" podID="a17e0cb0-68c0-4327-b7d0-94d146abcb7e" containerID="803b18f295753dbaca654c84cfd05616cce379be16a3587b2f20c531368ada81" exitCode=0 Nov 25 16:10:40 crc kubenswrapper[4890]: I1125 16:10:40.798816 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nm58z" event={"ID":"a17e0cb0-68c0-4327-b7d0-94d146abcb7e","Type":"ContainerDied","Data":"803b18f295753dbaca654c84cfd05616cce379be16a3587b2f20c531368ada81"} Nov 25 16:10:43 crc kubenswrapper[4890]: I1125 16:10:43.832104 4890 generic.go:334] "Generic (PLEG): container finished" podID="a17e0cb0-68c0-4327-b7d0-94d146abcb7e" containerID="e7d09b0568a74792250a27d04030cc57d5e5bcd897339facdac174ee0e39df63" exitCode=0 Nov 25 16:10:43 crc kubenswrapper[4890]: I1125 16:10:43.832213 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nm58z" event={"ID":"a17e0cb0-68c0-4327-b7d0-94d146abcb7e","Type":"ContainerDied","Data":"e7d09b0568a74792250a27d04030cc57d5e5bcd897339facdac174ee0e39df63"} Nov 25 16:10:44 crc kubenswrapper[4890]: I1125 16:10:44.848247 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nm58z" event={"ID":"a17e0cb0-68c0-4327-b7d0-94d146abcb7e","Type":"ContainerStarted","Data":"a64f294aa64d8b49d90a4e3a9d93b29d2a6455721ac18032c93addaeea49c32f"} Nov 25 16:10:45 crc kubenswrapper[4890]: I1125 16:10:45.883776 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nm58z" podStartSLOduration=4.2427652 podStartE2EDuration="7.883756174s" podCreationTimestamp="2025-11-25 16:10:38 +0000 UTC" firstStartedPulling="2025-11-25 16:10:40.800535692 +0000 UTC m=+4099.242998302" lastFinishedPulling="2025-11-25 16:10:44.441526666 +0000 UTC m=+4102.883989276" observedRunningTime="2025-11-25 16:10:45.877856858 +0000 UTC m=+4104.320319478" watchObservedRunningTime="2025-11-25 16:10:45.883756174 +0000 UTC m=+4104.326218784" Nov 25 16:10:49 crc kubenswrapper[4890]: I1125 16:10:49.128504 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nm58z" Nov 25 16:10:49 crc kubenswrapper[4890]: I1125 16:10:49.129234 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nm58z" Nov 25 16:10:49 crc kubenswrapper[4890]: I1125 16:10:49.192314 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nm58z" Nov 25 16:10:49 crc kubenswrapper[4890]: I1125 16:10:49.964613 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nm58z" Nov 25 16:10:50 crc kubenswrapper[4890]: I1125 16:10:50.044438 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nm58z"] Nov 25 16:10:51 crc kubenswrapper[4890]: I1125 16:10:51.663782 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5874bd7bc5-mrnt9_85206fda-8548-4b94-82ce-df958e8d4045/nmstate-console-plugin/0.log" Nov 25 16:10:51 crc kubenswrapper[4890]: I1125 16:10:51.881450 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-sk8fz_1e35f1ea-5729-463f-b875-6b897b4d5944/nmstate-handler/0.log" Nov 25 16:10:51 crc kubenswrapper[4890]: I1125 16:10:51.928808 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nm58z" podUID="a17e0cb0-68c0-4327-b7d0-94d146abcb7e" containerName="registry-server" containerID="cri-o://a64f294aa64d8b49d90a4e3a9d93b29d2a6455721ac18032c93addaeea49c32f" gracePeriod=2 Nov 25 16:10:51 crc kubenswrapper[4890]: I1125 16:10:51.934516 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-8mxxw_9d98268c-c10a-4157-9737-e2df6cbdcfe5/kube-rbac-proxy/0.log" Nov 25 16:10:52 crc kubenswrapper[4890]: I1125 16:10:52.016501 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-8mxxw_9d98268c-c10a-4157-9737-e2df6cbdcfe5/nmstate-metrics/0.log" Nov 25 16:10:52 crc kubenswrapper[4890]: I1125 16:10:52.239221 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-557fdffb88-v4blp_213afbd7-4c85-41e6-94c9-dac6db9da0d0/nmstate-operator/0.log" Nov 25 16:10:52 crc kubenswrapper[4890]: I1125 16:10:52.271847 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6b89b748d8-nhh4b_f6edfb37-872a-4515-ada7-07d552977de0/nmstate-webhook/0.log" Nov 25 16:10:52 crc kubenswrapper[4890]: I1125 16:10:52.942528 4890 generic.go:334] "Generic (PLEG): container finished" podID="a17e0cb0-68c0-4327-b7d0-94d146abcb7e" containerID="a64f294aa64d8b49d90a4e3a9d93b29d2a6455721ac18032c93addaeea49c32f" exitCode=0 Nov 25 16:10:52 crc kubenswrapper[4890]: I1125 16:10:52.942629 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nm58z" event={"ID":"a17e0cb0-68c0-4327-b7d0-94d146abcb7e","Type":"ContainerDied","Data":"a64f294aa64d8b49d90a4e3a9d93b29d2a6455721ac18032c93addaeea49c32f"} Nov 25 16:10:52 crc kubenswrapper[4890]: I1125 16:10:52.942888 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nm58z" event={"ID":"a17e0cb0-68c0-4327-b7d0-94d146abcb7e","Type":"ContainerDied","Data":"f3b9dd2ab5b7cba357b9fcbe2b5c33926d856bc5e1920bd7809d994b7adcd061"} Nov 25 16:10:52 crc kubenswrapper[4890]: I1125 16:10:52.942905 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f3b9dd2ab5b7cba357b9fcbe2b5c33926d856bc5e1920bd7809d994b7adcd061" Nov 25 16:10:52 crc kubenswrapper[4890]: I1125 16:10:52.999497 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nm58z" Nov 25 16:10:53 crc kubenswrapper[4890]: I1125 16:10:53.098894 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a17e0cb0-68c0-4327-b7d0-94d146abcb7e-utilities\") pod \"a17e0cb0-68c0-4327-b7d0-94d146abcb7e\" (UID: \"a17e0cb0-68c0-4327-b7d0-94d146abcb7e\") " Nov 25 16:10:53 crc kubenswrapper[4890]: I1125 16:10:53.099270 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8fwvq\" (UniqueName: \"kubernetes.io/projected/a17e0cb0-68c0-4327-b7d0-94d146abcb7e-kube-api-access-8fwvq\") pod \"a17e0cb0-68c0-4327-b7d0-94d146abcb7e\" (UID: \"a17e0cb0-68c0-4327-b7d0-94d146abcb7e\") " Nov 25 16:10:53 crc kubenswrapper[4890]: I1125 16:10:53.100051 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a17e0cb0-68c0-4327-b7d0-94d146abcb7e-utilities" (OuterVolumeSpecName: "utilities") pod "a17e0cb0-68c0-4327-b7d0-94d146abcb7e" (UID: "a17e0cb0-68c0-4327-b7d0-94d146abcb7e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:10:53 crc kubenswrapper[4890]: I1125 16:10:53.100228 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a17e0cb0-68c0-4327-b7d0-94d146abcb7e-catalog-content\") pod \"a17e0cb0-68c0-4327-b7d0-94d146abcb7e\" (UID: \"a17e0cb0-68c0-4327-b7d0-94d146abcb7e\") " Nov 25 16:10:53 crc kubenswrapper[4890]: I1125 16:10:53.100816 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a17e0cb0-68c0-4327-b7d0-94d146abcb7e-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 16:10:53 crc kubenswrapper[4890]: I1125 16:10:53.111405 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a17e0cb0-68c0-4327-b7d0-94d146abcb7e-kube-api-access-8fwvq" (OuterVolumeSpecName: "kube-api-access-8fwvq") pod "a17e0cb0-68c0-4327-b7d0-94d146abcb7e" (UID: "a17e0cb0-68c0-4327-b7d0-94d146abcb7e"). InnerVolumeSpecName "kube-api-access-8fwvq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:10:53 crc kubenswrapper[4890]: I1125 16:10:53.203277 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8fwvq\" (UniqueName: \"kubernetes.io/projected/a17e0cb0-68c0-4327-b7d0-94d146abcb7e-kube-api-access-8fwvq\") on node \"crc\" DevicePath \"\"" Nov 25 16:10:53 crc kubenswrapper[4890]: I1125 16:10:53.952385 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nm58z" Nov 25 16:10:54 crc kubenswrapper[4890]: I1125 16:10:54.599771 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a17e0cb0-68c0-4327-b7d0-94d146abcb7e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a17e0cb0-68c0-4327-b7d0-94d146abcb7e" (UID: "a17e0cb0-68c0-4327-b7d0-94d146abcb7e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:10:54 crc kubenswrapper[4890]: I1125 16:10:54.660054 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a17e0cb0-68c0-4327-b7d0-94d146abcb7e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 16:10:54 crc kubenswrapper[4890]: I1125 16:10:54.896778 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nm58z"] Nov 25 16:10:54 crc kubenswrapper[4890]: I1125 16:10:54.907452 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nm58z"] Nov 25 16:10:56 crc kubenswrapper[4890]: I1125 16:10:56.189825 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a17e0cb0-68c0-4327-b7d0-94d146abcb7e" path="/var/lib/kubelet/pods/a17e0cb0-68c0-4327-b7d0-94d146abcb7e/volumes" Nov 25 16:10:56 crc kubenswrapper[4890]: I1125 16:10:56.447895 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 16:10:56 crc kubenswrapper[4890]: I1125 16:10:56.448072 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 16:11:08 crc kubenswrapper[4890]: I1125 16:11:08.830905 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-hknzk_d97f8477-4597-4fb0-b2fc-e3faf124ba4e/kube-rbac-proxy/0.log" Nov 25 16:11:09 crc kubenswrapper[4890]: I1125 16:11:09.038728 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhqzp_60b160c7-c63e-4ffc-b031-9c9fa5c248e2/cp-frr-files/0.log" Nov 25 16:11:09 crc kubenswrapper[4890]: I1125 16:11:09.050316 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-hknzk_d97f8477-4597-4fb0-b2fc-e3faf124ba4e/controller/0.log" Nov 25 16:11:09 crc kubenswrapper[4890]: I1125 16:11:09.273550 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-sz4z6"] Nov 25 16:11:09 crc kubenswrapper[4890]: E1125 16:11:09.274388 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a17e0cb0-68c0-4327-b7d0-94d146abcb7e" containerName="registry-server" Nov 25 16:11:09 crc kubenswrapper[4890]: I1125 16:11:09.274407 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="a17e0cb0-68c0-4327-b7d0-94d146abcb7e" containerName="registry-server" Nov 25 16:11:09 crc kubenswrapper[4890]: E1125 16:11:09.274443 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a17e0cb0-68c0-4327-b7d0-94d146abcb7e" containerName="extract-content" Nov 25 16:11:09 crc kubenswrapper[4890]: I1125 16:11:09.274450 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="a17e0cb0-68c0-4327-b7d0-94d146abcb7e" containerName="extract-content" Nov 25 16:11:09 crc kubenswrapper[4890]: E1125 16:11:09.274477 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a17e0cb0-68c0-4327-b7d0-94d146abcb7e" containerName="extract-utilities" Nov 25 16:11:09 crc kubenswrapper[4890]: I1125 16:11:09.274484 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="a17e0cb0-68c0-4327-b7d0-94d146abcb7e" containerName="extract-utilities" Nov 25 16:11:09 crc kubenswrapper[4890]: I1125 16:11:09.274654 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="a17e0cb0-68c0-4327-b7d0-94d146abcb7e" containerName="registry-server" Nov 25 16:11:09 crc kubenswrapper[4890]: I1125 16:11:09.277322 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sz4z6" Nov 25 16:11:09 crc kubenswrapper[4890]: I1125 16:11:09.283754 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhqzp_60b160c7-c63e-4ffc-b031-9c9fa5c248e2/cp-reloader/0.log" Nov 25 16:11:09 crc kubenswrapper[4890]: I1125 16:11:09.303070 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sz4z6"] Nov 25 16:11:09 crc kubenswrapper[4890]: I1125 16:11:09.345881 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhqzp_60b160c7-c63e-4ffc-b031-9c9fa5c248e2/cp-frr-files/0.log" Nov 25 16:11:09 crc kubenswrapper[4890]: I1125 16:11:09.373895 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1deef159-1a05-493e-aad2-c3467c5bd03b-utilities\") pod \"redhat-operators-sz4z6\" (UID: \"1deef159-1a05-493e-aad2-c3467c5bd03b\") " pod="openshift-marketplace/redhat-operators-sz4z6" Nov 25 16:11:09 crc kubenswrapper[4890]: I1125 16:11:09.374017 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvxc2\" (UniqueName: \"kubernetes.io/projected/1deef159-1a05-493e-aad2-c3467c5bd03b-kube-api-access-fvxc2\") pod \"redhat-operators-sz4z6\" (UID: \"1deef159-1a05-493e-aad2-c3467c5bd03b\") " pod="openshift-marketplace/redhat-operators-sz4z6" Nov 25 16:11:09 crc kubenswrapper[4890]: I1125 16:11:09.374100 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1deef159-1a05-493e-aad2-c3467c5bd03b-catalog-content\") pod \"redhat-operators-sz4z6\" (UID: \"1deef159-1a05-493e-aad2-c3467c5bd03b\") " pod="openshift-marketplace/redhat-operators-sz4z6" Nov 25 16:11:09 crc kubenswrapper[4890]: I1125 16:11:09.390111 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhqzp_60b160c7-c63e-4ffc-b031-9c9fa5c248e2/cp-reloader/0.log" Nov 25 16:11:09 crc kubenswrapper[4890]: I1125 16:11:09.396808 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhqzp_60b160c7-c63e-4ffc-b031-9c9fa5c248e2/cp-metrics/0.log" Nov 25 16:11:09 crc kubenswrapper[4890]: I1125 16:11:09.475845 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1deef159-1a05-493e-aad2-c3467c5bd03b-utilities\") pod \"redhat-operators-sz4z6\" (UID: \"1deef159-1a05-493e-aad2-c3467c5bd03b\") " pod="openshift-marketplace/redhat-operators-sz4z6" Nov 25 16:11:09 crc kubenswrapper[4890]: I1125 16:11:09.475973 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvxc2\" (UniqueName: \"kubernetes.io/projected/1deef159-1a05-493e-aad2-c3467c5bd03b-kube-api-access-fvxc2\") pod \"redhat-operators-sz4z6\" (UID: \"1deef159-1a05-493e-aad2-c3467c5bd03b\") " pod="openshift-marketplace/redhat-operators-sz4z6" Nov 25 16:11:09 crc kubenswrapper[4890]: I1125 16:11:09.476381 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1deef159-1a05-493e-aad2-c3467c5bd03b-catalog-content\") pod \"redhat-operators-sz4z6\" (UID: \"1deef159-1a05-493e-aad2-c3467c5bd03b\") " pod="openshift-marketplace/redhat-operators-sz4z6" Nov 25 16:11:09 crc kubenswrapper[4890]: I1125 16:11:09.476556 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1deef159-1a05-493e-aad2-c3467c5bd03b-utilities\") pod \"redhat-operators-sz4z6\" (UID: \"1deef159-1a05-493e-aad2-c3467c5bd03b\") " pod="openshift-marketplace/redhat-operators-sz4z6" Nov 25 16:11:09 crc kubenswrapper[4890]: I1125 16:11:09.476782 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1deef159-1a05-493e-aad2-c3467c5bd03b-catalog-content\") pod \"redhat-operators-sz4z6\" (UID: \"1deef159-1a05-493e-aad2-c3467c5bd03b\") " pod="openshift-marketplace/redhat-operators-sz4z6" Nov 25 16:11:09 crc kubenswrapper[4890]: I1125 16:11:09.500780 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvxc2\" (UniqueName: \"kubernetes.io/projected/1deef159-1a05-493e-aad2-c3467c5bd03b-kube-api-access-fvxc2\") pod \"redhat-operators-sz4z6\" (UID: \"1deef159-1a05-493e-aad2-c3467c5bd03b\") " pod="openshift-marketplace/redhat-operators-sz4z6" Nov 25 16:11:09 crc kubenswrapper[4890]: I1125 16:11:09.557674 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhqzp_60b160c7-c63e-4ffc-b031-9c9fa5c248e2/cp-frr-files/0.log" Nov 25 16:11:09 crc kubenswrapper[4890]: I1125 16:11:09.618355 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhqzp_60b160c7-c63e-4ffc-b031-9c9fa5c248e2/cp-reloader/0.log" Nov 25 16:11:09 crc kubenswrapper[4890]: I1125 16:11:09.621120 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sz4z6" Nov 25 16:11:09 crc kubenswrapper[4890]: I1125 16:11:09.646039 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhqzp_60b160c7-c63e-4ffc-b031-9c9fa5c248e2/cp-metrics/0.log" Nov 25 16:11:09 crc kubenswrapper[4890]: I1125 16:11:09.722753 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhqzp_60b160c7-c63e-4ffc-b031-9c9fa5c248e2/cp-metrics/0.log" Nov 25 16:11:09 crc kubenswrapper[4890]: I1125 16:11:09.961874 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhqzp_60b160c7-c63e-4ffc-b031-9c9fa5c248e2/cp-reloader/0.log" Nov 25 16:11:09 crc kubenswrapper[4890]: I1125 16:11:09.962134 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhqzp_60b160c7-c63e-4ffc-b031-9c9fa5c248e2/cp-frr-files/0.log" Nov 25 16:11:10 crc kubenswrapper[4890]: I1125 16:11:10.062633 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhqzp_60b160c7-c63e-4ffc-b031-9c9fa5c248e2/cp-metrics/0.log" Nov 25 16:11:10 crc kubenswrapper[4890]: I1125 16:11:10.105945 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhqzp_60b160c7-c63e-4ffc-b031-9c9fa5c248e2/controller/0.log" Nov 25 16:11:10 crc kubenswrapper[4890]: I1125 16:11:10.152754 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sz4z6"] Nov 25 16:11:10 crc kubenswrapper[4890]: I1125 16:11:10.233093 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhqzp_60b160c7-c63e-4ffc-b031-9c9fa5c248e2/frr-metrics/0.log" Nov 25 16:11:10 crc kubenswrapper[4890]: I1125 16:11:10.302538 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhqzp_60b160c7-c63e-4ffc-b031-9c9fa5c248e2/kube-rbac-proxy/0.log" Nov 25 16:11:11 crc kubenswrapper[4890]: I1125 16:11:11.011884 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhqzp_60b160c7-c63e-4ffc-b031-9c9fa5c248e2/kube-rbac-proxy-frr/0.log" Nov 25 16:11:11 crc kubenswrapper[4890]: I1125 16:11:11.024209 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhqzp_60b160c7-c63e-4ffc-b031-9c9fa5c248e2/reloader/0.log" Nov 25 16:11:11 crc kubenswrapper[4890]: I1125 16:11:11.127076 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sz4z6" event={"ID":"1deef159-1a05-493e-aad2-c3467c5bd03b","Type":"ContainerStarted","Data":"9b12495830bbdcf36a4bcba983fa10108d80ca8d040c5476019a8caf9a425ac5"} Nov 25 16:11:11 crc kubenswrapper[4890]: I1125 16:11:11.216245 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-6998585d5-9ncz9_742ce93d-d49e-4b12-a1cf-5413ec8475ac/frr-k8s-webhook-server/0.log" Nov 25 16:11:11 crc kubenswrapper[4890]: I1125 16:11:11.339390 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-66fc8c7fbf-lnjxm_dad943fe-2209-40f9-a9bb-0be7ba80cca2/manager/3.log" Nov 25 16:11:11 crc kubenswrapper[4890]: I1125 16:11:11.493372 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-66fc8c7fbf-lnjxm_dad943fe-2209-40f9-a9bb-0be7ba80cca2/manager/2.log" Nov 25 16:11:11 crc kubenswrapper[4890]: I1125 16:11:11.657192 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-6b856c5bd9-slq4g_fb5c6944-bf04-42b5-98f6-550b9b2811d5/webhook-server/0.log" Nov 25 16:11:11 crc kubenswrapper[4890]: I1125 16:11:11.769344 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-ndl5n_9c563abc-f1fc-4c15-9a03-901371600338/kube-rbac-proxy/0.log" Nov 25 16:11:11 crc kubenswrapper[4890]: I1125 16:11:11.843458 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhqzp_60b160c7-c63e-4ffc-b031-9c9fa5c248e2/frr/0.log" Nov 25 16:11:12 crc kubenswrapper[4890]: I1125 16:11:12.138690 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sz4z6" event={"ID":"1deef159-1a05-493e-aad2-c3467c5bd03b","Type":"ContainerDied","Data":"aa80cf31a849e912874a4937d448b34ee93bfeab1d6e97a94100845d839d3bcb"} Nov 25 16:11:12 crc kubenswrapper[4890]: I1125 16:11:12.139583 4890 generic.go:334] "Generic (PLEG): container finished" podID="1deef159-1a05-493e-aad2-c3467c5bd03b" containerID="aa80cf31a849e912874a4937d448b34ee93bfeab1d6e97a94100845d839d3bcb" exitCode=0 Nov 25 16:11:12 crc kubenswrapper[4890]: I1125 16:11:12.571643 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-ndl5n_9c563abc-f1fc-4c15-9a03-901371600338/speaker/0.log" Nov 25 16:11:19 crc kubenswrapper[4890]: I1125 16:11:19.216154 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sz4z6" event={"ID":"1deef159-1a05-493e-aad2-c3467c5bd03b","Type":"ContainerStarted","Data":"a5770458e17dc975fd913d5757b0dbe1183c0c4dd2effb731d5700e65ac3ff06"} Nov 25 16:11:24 crc kubenswrapper[4890]: I1125 16:11:24.191086 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772er4qx8_d89e1065-7553-4d5b-aef6-38108a5cb8ae/util/0.log" Nov 25 16:11:24 crc kubenswrapper[4890]: I1125 16:11:24.313555 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772er4qx8_d89e1065-7553-4d5b-aef6-38108a5cb8ae/util/0.log" Nov 25 16:11:24 crc kubenswrapper[4890]: I1125 16:11:24.348860 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772er4qx8_d89e1065-7553-4d5b-aef6-38108a5cb8ae/pull/0.log" Nov 25 16:11:24 crc kubenswrapper[4890]: I1125 16:11:24.352579 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772er4qx8_d89e1065-7553-4d5b-aef6-38108a5cb8ae/pull/0.log" Nov 25 16:11:24 crc kubenswrapper[4890]: I1125 16:11:24.532475 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772er4qx8_d89e1065-7553-4d5b-aef6-38108a5cb8ae/extract/0.log" Nov 25 16:11:24 crc kubenswrapper[4890]: I1125 16:11:24.675134 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-fksn8_6afbc261-80f8-4f4a-814f-5df9ab43502e/extract-utilities/0.log" Nov 25 16:11:24 crc kubenswrapper[4890]: I1125 16:11:24.759746 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772er4qx8_d89e1065-7553-4d5b-aef6-38108a5cb8ae/util/0.log" Nov 25 16:11:24 crc kubenswrapper[4890]: I1125 16:11:24.819467 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772er4qx8_d89e1065-7553-4d5b-aef6-38108a5cb8ae/pull/0.log" Nov 25 16:11:24 crc kubenswrapper[4890]: I1125 16:11:24.903277 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-fksn8_6afbc261-80f8-4f4a-814f-5df9ab43502e/extract-utilities/0.log" Nov 25 16:11:24 crc kubenswrapper[4890]: I1125 16:11:24.942077 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-fksn8_6afbc261-80f8-4f4a-814f-5df9ab43502e/extract-content/0.log" Nov 25 16:11:24 crc kubenswrapper[4890]: I1125 16:11:24.996780 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-fksn8_6afbc261-80f8-4f4a-814f-5df9ab43502e/extract-content/0.log" Nov 25 16:11:25 crc kubenswrapper[4890]: I1125 16:11:25.220395 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-fksn8_6afbc261-80f8-4f4a-814f-5df9ab43502e/extract-utilities/0.log" Nov 25 16:11:25 crc kubenswrapper[4890]: I1125 16:11:25.241636 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-fksn8_6afbc261-80f8-4f4a-814f-5df9ab43502e/extract-content/0.log" Nov 25 16:11:25 crc kubenswrapper[4890]: I1125 16:11:25.424954 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gxfqw_fc5b6ee3-033f-4b8f-82b7-46e28498f625/extract-utilities/0.log" Nov 25 16:11:25 crc kubenswrapper[4890]: I1125 16:11:25.581668 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gxfqw_fc5b6ee3-033f-4b8f-82b7-46e28498f625/extract-utilities/0.log" Nov 25 16:11:25 crc kubenswrapper[4890]: I1125 16:11:25.617467 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gxfqw_fc5b6ee3-033f-4b8f-82b7-46e28498f625/extract-content/0.log" Nov 25 16:11:25 crc kubenswrapper[4890]: I1125 16:11:25.688578 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gxfqw_fc5b6ee3-033f-4b8f-82b7-46e28498f625/extract-content/0.log" Nov 25 16:11:25 crc kubenswrapper[4890]: I1125 16:11:25.855969 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gxfqw_fc5b6ee3-033f-4b8f-82b7-46e28498f625/extract-content/0.log" Nov 25 16:11:25 crc kubenswrapper[4890]: I1125 16:11:25.856817 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gxfqw_fc5b6ee3-033f-4b8f-82b7-46e28498f625/extract-utilities/0.log" Nov 25 16:11:26 crc kubenswrapper[4890]: I1125 16:11:26.118011 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mhgqc_76fcec1c-1c27-4e7f-9f3e-a2527b72581f/util/0.log" Nov 25 16:11:26 crc kubenswrapper[4890]: I1125 16:11:26.447194 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 16:11:26 crc kubenswrapper[4890]: I1125 16:11:26.447251 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 16:11:26 crc kubenswrapper[4890]: I1125 16:11:26.447304 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" Nov 25 16:11:26 crc kubenswrapper[4890]: I1125 16:11:26.448074 4890 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0ee5f289dd255f61c3f8c4e86c84d7044d423518d51541add4da1a581b273eb1"} pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 16:11:26 crc kubenswrapper[4890]: I1125 16:11:26.448120 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" containerID="cri-o://0ee5f289dd255f61c3f8c4e86c84d7044d423518d51541add4da1a581b273eb1" gracePeriod=600 Nov 25 16:11:26 crc kubenswrapper[4890]: I1125 16:11:26.512704 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mhgqc_76fcec1c-1c27-4e7f-9f3e-a2527b72581f/util/0.log" Nov 25 16:11:26 crc kubenswrapper[4890]: I1125 16:11:26.622252 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mhgqc_76fcec1c-1c27-4e7f-9f3e-a2527b72581f/pull/0.log" Nov 25 16:11:26 crc kubenswrapper[4890]: I1125 16:11:26.663356 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-fksn8_6afbc261-80f8-4f4a-814f-5df9ab43502e/registry-server/0.log" Nov 25 16:11:26 crc kubenswrapper[4890]: I1125 16:11:26.663757 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gxfqw_fc5b6ee3-033f-4b8f-82b7-46e28498f625/registry-server/0.log" Nov 25 16:11:26 crc kubenswrapper[4890]: I1125 16:11:26.930936 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mhgqc_76fcec1c-1c27-4e7f-9f3e-a2527b72581f/util/0.log" Nov 25 16:11:26 crc kubenswrapper[4890]: I1125 16:11:26.983246 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mhgqc_76fcec1c-1c27-4e7f-9f3e-a2527b72581f/extract/0.log" Nov 25 16:11:27 crc kubenswrapper[4890]: I1125 16:11:27.164400 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mhgqc_76fcec1c-1c27-4e7f-9f3e-a2527b72581f/pull/0.log" Nov 25 16:11:27 crc kubenswrapper[4890]: I1125 16:11:27.196323 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mhgqc_76fcec1c-1c27-4e7f-9f3e-a2527b72581f/pull/0.log" Nov 25 16:11:27 crc kubenswrapper[4890]: I1125 16:11:27.235495 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-2xzsf_4bbe88e6-b859-452a-baf1-8a963b475014/marketplace-operator/1.log" Nov 25 16:11:27 crc kubenswrapper[4890]: I1125 16:11:27.291052 4890 generic.go:334] "Generic (PLEG): container finished" podID="4e4f849d-f239-4727-a73e-18327856929a" containerID="0ee5f289dd255f61c3f8c4e86c84d7044d423518d51541add4da1a581b273eb1" exitCode=0 Nov 25 16:11:27 crc kubenswrapper[4890]: I1125 16:11:27.291095 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" event={"ID":"4e4f849d-f239-4727-a73e-18327856929a","Type":"ContainerDied","Data":"0ee5f289dd255f61c3f8c4e86c84d7044d423518d51541add4da1a581b273eb1"} Nov 25 16:11:27 crc kubenswrapper[4890]: I1125 16:11:27.291135 4890 scope.go:117] "RemoveContainer" containerID="a0bac43b712e038c240449718f3b53c0d88f1d20e0f5e92b7dfb0385a9de4f68" Nov 25 16:11:27 crc kubenswrapper[4890]: I1125 16:11:27.423222 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-tjd2v_f444e25c-b3bf-4046-a1ff-eba1a422a88e/extract-utilities/0.log" Nov 25 16:11:27 crc kubenswrapper[4890]: I1125 16:11:27.614051 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-tjd2v_f444e25c-b3bf-4046-a1ff-eba1a422a88e/extract-utilities/0.log" Nov 25 16:11:27 crc kubenswrapper[4890]: I1125 16:11:27.627288 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-tjd2v_f444e25c-b3bf-4046-a1ff-eba1a422a88e/extract-content/0.log" Nov 25 16:11:27 crc kubenswrapper[4890]: I1125 16:11:27.661112 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-2xzsf_4bbe88e6-b859-452a-baf1-8a963b475014/marketplace-operator/0.log" Nov 25 16:11:27 crc kubenswrapper[4890]: I1125 16:11:27.682929 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-tjd2v_f444e25c-b3bf-4046-a1ff-eba1a422a88e/extract-content/0.log" Nov 25 16:11:27 crc kubenswrapper[4890]: I1125 16:11:27.867403 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-tjd2v_f444e25c-b3bf-4046-a1ff-eba1a422a88e/extract-content/0.log" Nov 25 16:11:27 crc kubenswrapper[4890]: I1125 16:11:27.875381 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-tjd2v_f444e25c-b3bf-4046-a1ff-eba1a422a88e/extract-utilities/0.log" Nov 25 16:11:28 crc kubenswrapper[4890]: I1125 16:11:28.073693 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-tjd2v_f444e25c-b3bf-4046-a1ff-eba1a422a88e/registry-server/0.log" Nov 25 16:11:28 crc kubenswrapper[4890]: I1125 16:11:28.090830 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jcmgp_c83811ae-c669-4268-a58d-e4b44d18f26c/extract-utilities/0.log" Nov 25 16:11:28 crc kubenswrapper[4890]: I1125 16:11:28.287961 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jcmgp_c83811ae-c669-4268-a58d-e4b44d18f26c/extract-content/0.log" Nov 25 16:11:28 crc kubenswrapper[4890]: I1125 16:11:28.288112 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jcmgp_c83811ae-c669-4268-a58d-e4b44d18f26c/extract-utilities/0.log" Nov 25 16:11:28 crc kubenswrapper[4890]: I1125 16:11:28.304193 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jcmgp_c83811ae-c669-4268-a58d-e4b44d18f26c/extract-content/0.log" Nov 25 16:11:28 crc kubenswrapper[4890]: I1125 16:11:28.500088 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jcmgp_c83811ae-c669-4268-a58d-e4b44d18f26c/extract-content/0.log" Nov 25 16:11:28 crc kubenswrapper[4890]: I1125 16:11:28.565097 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jcmgp_c83811ae-c669-4268-a58d-e4b44d18f26c/extract-utilities/0.log" Nov 25 16:11:28 crc kubenswrapper[4890]: I1125 16:11:28.584370 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-sz4z6_1deef159-1a05-493e-aad2-c3467c5bd03b/extract-utilities/0.log" Nov 25 16:11:28 crc kubenswrapper[4890]: I1125 16:11:28.705537 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-sz4z6_1deef159-1a05-493e-aad2-c3467c5bd03b/extract-utilities/0.log" Nov 25 16:11:28 crc kubenswrapper[4890]: I1125 16:11:28.740378 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-sz4z6_1deef159-1a05-493e-aad2-c3467c5bd03b/extract-content/0.log" Nov 25 16:11:28 crc kubenswrapper[4890]: I1125 16:11:28.753009 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-sz4z6_1deef159-1a05-493e-aad2-c3467c5bd03b/extract-content/0.log" Nov 25 16:11:28 crc kubenswrapper[4890]: I1125 16:11:28.916636 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-sz4z6_1deef159-1a05-493e-aad2-c3467c5bd03b/extract-content/0.log" Nov 25 16:11:28 crc kubenswrapper[4890]: I1125 16:11:28.920271 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-sz4z6_1deef159-1a05-493e-aad2-c3467c5bd03b/extract-utilities/0.log" Nov 25 16:11:30 crc kubenswrapper[4890]: I1125 16:11:30.509280 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jcmgp_c83811ae-c669-4268-a58d-e4b44d18f26c/registry-server/0.log" Nov 25 16:11:31 crc kubenswrapper[4890]: I1125 16:11:31.336808 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" event={"ID":"4e4f849d-f239-4727-a73e-18327856929a","Type":"ContainerStarted","Data":"28fed48468d92f7b1fe5d54d4505526fd8660c4d76009eee3215a28f9eb145da"} Nov 25 16:11:37 crc kubenswrapper[4890]: I1125 16:11:37.405985 4890 generic.go:334] "Generic (PLEG): container finished" podID="1deef159-1a05-493e-aad2-c3467c5bd03b" containerID="a5770458e17dc975fd913d5757b0dbe1183c0c4dd2effb731d5700e65ac3ff06" exitCode=0 Nov 25 16:11:37 crc kubenswrapper[4890]: I1125 16:11:37.406318 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sz4z6" event={"ID":"1deef159-1a05-493e-aad2-c3467c5bd03b","Type":"ContainerDied","Data":"a5770458e17dc975fd913d5757b0dbe1183c0c4dd2effb731d5700e65ac3ff06"} Nov 25 16:11:40 crc kubenswrapper[4890]: I1125 16:11:40.435802 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sz4z6" event={"ID":"1deef159-1a05-493e-aad2-c3467c5bd03b","Type":"ContainerStarted","Data":"d0f1841c29efa70ab155228d2655705a698312600bf8b1c4a93d77fe04a08606"} Nov 25 16:11:40 crc kubenswrapper[4890]: I1125 16:11:40.463247 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-sz4z6" podStartSLOduration=4.499522186 podStartE2EDuration="31.463220062s" podCreationTimestamp="2025-11-25 16:11:09 +0000 UTC" firstStartedPulling="2025-11-25 16:11:12.14136029 +0000 UTC m=+4130.583822900" lastFinishedPulling="2025-11-25 16:11:39.105058176 +0000 UTC m=+4157.547520776" observedRunningTime="2025-11-25 16:11:40.451573532 +0000 UTC m=+4158.894036142" watchObservedRunningTime="2025-11-25 16:11:40.463220062 +0000 UTC m=+4158.905682672" Nov 25 16:11:49 crc kubenswrapper[4890]: I1125 16:11:49.621828 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-sz4z6" Nov 25 16:11:49 crc kubenswrapper[4890]: I1125 16:11:49.622512 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-sz4z6" Nov 25 16:11:50 crc kubenswrapper[4890]: I1125 16:11:50.672185 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-sz4z6" podUID="1deef159-1a05-493e-aad2-c3467c5bd03b" containerName="registry-server" probeResult="failure" output=< Nov 25 16:11:50 crc kubenswrapper[4890]: timeout: failed to connect service ":50051" within 1s Nov 25 16:11:50 crc kubenswrapper[4890]: > Nov 25 16:12:00 crc kubenswrapper[4890]: I1125 16:12:00.680806 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-sz4z6" podUID="1deef159-1a05-493e-aad2-c3467c5bd03b" containerName="registry-server" probeResult="failure" output=< Nov 25 16:12:00 crc kubenswrapper[4890]: timeout: failed to connect service ":50051" within 1s Nov 25 16:12:00 crc kubenswrapper[4890]: > Nov 25 16:12:09 crc kubenswrapper[4890]: I1125 16:12:09.682520 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-sz4z6" Nov 25 16:12:09 crc kubenswrapper[4890]: I1125 16:12:09.728685 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-sz4z6" Nov 25 16:12:09 crc kubenswrapper[4890]: I1125 16:12:09.916695 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sz4z6"] Nov 25 16:12:10 crc kubenswrapper[4890]: I1125 16:12:10.771782 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-sz4z6" podUID="1deef159-1a05-493e-aad2-c3467c5bd03b" containerName="registry-server" containerID="cri-o://d0f1841c29efa70ab155228d2655705a698312600bf8b1c4a93d77fe04a08606" gracePeriod=2 Nov 25 16:12:11 crc kubenswrapper[4890]: I1125 16:12:11.375951 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sz4z6" Nov 25 16:12:11 crc kubenswrapper[4890]: I1125 16:12:11.585340 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1deef159-1a05-493e-aad2-c3467c5bd03b-utilities\") pod \"1deef159-1a05-493e-aad2-c3467c5bd03b\" (UID: \"1deef159-1a05-493e-aad2-c3467c5bd03b\") " Nov 25 16:12:11 crc kubenswrapper[4890]: I1125 16:12:11.585658 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1deef159-1a05-493e-aad2-c3467c5bd03b-catalog-content\") pod \"1deef159-1a05-493e-aad2-c3467c5bd03b\" (UID: \"1deef159-1a05-493e-aad2-c3467c5bd03b\") " Nov 25 16:12:11 crc kubenswrapper[4890]: I1125 16:12:11.585736 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fvxc2\" (UniqueName: \"kubernetes.io/projected/1deef159-1a05-493e-aad2-c3467c5bd03b-kube-api-access-fvxc2\") pod \"1deef159-1a05-493e-aad2-c3467c5bd03b\" (UID: \"1deef159-1a05-493e-aad2-c3467c5bd03b\") " Nov 25 16:12:11 crc kubenswrapper[4890]: I1125 16:12:11.591282 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1deef159-1a05-493e-aad2-c3467c5bd03b-utilities" (OuterVolumeSpecName: "utilities") pod "1deef159-1a05-493e-aad2-c3467c5bd03b" (UID: "1deef159-1a05-493e-aad2-c3467c5bd03b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:12:11 crc kubenswrapper[4890]: I1125 16:12:11.597260 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1deef159-1a05-493e-aad2-c3467c5bd03b-kube-api-access-fvxc2" (OuterVolumeSpecName: "kube-api-access-fvxc2") pod "1deef159-1a05-493e-aad2-c3467c5bd03b" (UID: "1deef159-1a05-493e-aad2-c3467c5bd03b"). InnerVolumeSpecName "kube-api-access-fvxc2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:12:11 crc kubenswrapper[4890]: I1125 16:12:11.688648 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fvxc2\" (UniqueName: \"kubernetes.io/projected/1deef159-1a05-493e-aad2-c3467c5bd03b-kube-api-access-fvxc2\") on node \"crc\" DevicePath \"\"" Nov 25 16:12:11 crc kubenswrapper[4890]: I1125 16:12:11.688693 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1deef159-1a05-493e-aad2-c3467c5bd03b-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 16:12:11 crc kubenswrapper[4890]: I1125 16:12:11.722962 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1deef159-1a05-493e-aad2-c3467c5bd03b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1deef159-1a05-493e-aad2-c3467c5bd03b" (UID: "1deef159-1a05-493e-aad2-c3467c5bd03b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:12:11 crc kubenswrapper[4890]: I1125 16:12:11.784042 4890 generic.go:334] "Generic (PLEG): container finished" podID="1deef159-1a05-493e-aad2-c3467c5bd03b" containerID="d0f1841c29efa70ab155228d2655705a698312600bf8b1c4a93d77fe04a08606" exitCode=0 Nov 25 16:12:11 crc kubenswrapper[4890]: I1125 16:12:11.784327 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sz4z6" event={"ID":"1deef159-1a05-493e-aad2-c3467c5bd03b","Type":"ContainerDied","Data":"d0f1841c29efa70ab155228d2655705a698312600bf8b1c4a93d77fe04a08606"} Nov 25 16:12:11 crc kubenswrapper[4890]: I1125 16:12:11.786261 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sz4z6" event={"ID":"1deef159-1a05-493e-aad2-c3467c5bd03b","Type":"ContainerDied","Data":"9b12495830bbdcf36a4bcba983fa10108d80ca8d040c5476019a8caf9a425ac5"} Nov 25 16:12:11 crc kubenswrapper[4890]: I1125 16:12:11.786302 4890 scope.go:117] "RemoveContainer" containerID="d0f1841c29efa70ab155228d2655705a698312600bf8b1c4a93d77fe04a08606" Nov 25 16:12:11 crc kubenswrapper[4890]: I1125 16:12:11.784520 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sz4z6" Nov 25 16:12:11 crc kubenswrapper[4890]: I1125 16:12:11.799796 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1deef159-1a05-493e-aad2-c3467c5bd03b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 16:12:11 crc kubenswrapper[4890]: I1125 16:12:11.832113 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sz4z6"] Nov 25 16:12:11 crc kubenswrapper[4890]: I1125 16:12:11.843028 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-sz4z6"] Nov 25 16:12:11 crc kubenswrapper[4890]: I1125 16:12:11.879230 4890 scope.go:117] "RemoveContainer" containerID="a5770458e17dc975fd913d5757b0dbe1183c0c4dd2effb731d5700e65ac3ff06" Nov 25 16:12:11 crc kubenswrapper[4890]: I1125 16:12:11.901548 4890 scope.go:117] "RemoveContainer" containerID="aa80cf31a849e912874a4937d448b34ee93bfeab1d6e97a94100845d839d3bcb" Nov 25 16:12:11 crc kubenswrapper[4890]: I1125 16:12:11.950619 4890 scope.go:117] "RemoveContainer" containerID="d0f1841c29efa70ab155228d2655705a698312600bf8b1c4a93d77fe04a08606" Nov 25 16:12:11 crc kubenswrapper[4890]: E1125 16:12:11.951105 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0f1841c29efa70ab155228d2655705a698312600bf8b1c4a93d77fe04a08606\": container with ID starting with d0f1841c29efa70ab155228d2655705a698312600bf8b1c4a93d77fe04a08606 not found: ID does not exist" containerID="d0f1841c29efa70ab155228d2655705a698312600bf8b1c4a93d77fe04a08606" Nov 25 16:12:11 crc kubenswrapper[4890]: I1125 16:12:11.951285 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0f1841c29efa70ab155228d2655705a698312600bf8b1c4a93d77fe04a08606"} err="failed to get container status \"d0f1841c29efa70ab155228d2655705a698312600bf8b1c4a93d77fe04a08606\": rpc error: code = NotFound desc = could not find container \"d0f1841c29efa70ab155228d2655705a698312600bf8b1c4a93d77fe04a08606\": container with ID starting with d0f1841c29efa70ab155228d2655705a698312600bf8b1c4a93d77fe04a08606 not found: ID does not exist" Nov 25 16:12:11 crc kubenswrapper[4890]: I1125 16:12:11.951382 4890 scope.go:117] "RemoveContainer" containerID="a5770458e17dc975fd913d5757b0dbe1183c0c4dd2effb731d5700e65ac3ff06" Nov 25 16:12:11 crc kubenswrapper[4890]: E1125 16:12:11.951676 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5770458e17dc975fd913d5757b0dbe1183c0c4dd2effb731d5700e65ac3ff06\": container with ID starting with a5770458e17dc975fd913d5757b0dbe1183c0c4dd2effb731d5700e65ac3ff06 not found: ID does not exist" containerID="a5770458e17dc975fd913d5757b0dbe1183c0c4dd2effb731d5700e65ac3ff06" Nov 25 16:12:11 crc kubenswrapper[4890]: I1125 16:12:11.951770 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5770458e17dc975fd913d5757b0dbe1183c0c4dd2effb731d5700e65ac3ff06"} err="failed to get container status \"a5770458e17dc975fd913d5757b0dbe1183c0c4dd2effb731d5700e65ac3ff06\": rpc error: code = NotFound desc = could not find container \"a5770458e17dc975fd913d5757b0dbe1183c0c4dd2effb731d5700e65ac3ff06\": container with ID starting with a5770458e17dc975fd913d5757b0dbe1183c0c4dd2effb731d5700e65ac3ff06 not found: ID does not exist" Nov 25 16:12:11 crc kubenswrapper[4890]: I1125 16:12:11.951838 4890 scope.go:117] "RemoveContainer" containerID="aa80cf31a849e912874a4937d448b34ee93bfeab1d6e97a94100845d839d3bcb" Nov 25 16:12:11 crc kubenswrapper[4890]: E1125 16:12:11.952063 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa80cf31a849e912874a4937d448b34ee93bfeab1d6e97a94100845d839d3bcb\": container with ID starting with aa80cf31a849e912874a4937d448b34ee93bfeab1d6e97a94100845d839d3bcb not found: ID does not exist" containerID="aa80cf31a849e912874a4937d448b34ee93bfeab1d6e97a94100845d839d3bcb" Nov 25 16:12:11 crc kubenswrapper[4890]: I1125 16:12:11.952149 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa80cf31a849e912874a4937d448b34ee93bfeab1d6e97a94100845d839d3bcb"} err="failed to get container status \"aa80cf31a849e912874a4937d448b34ee93bfeab1d6e97a94100845d839d3bcb\": rpc error: code = NotFound desc = could not find container \"aa80cf31a849e912874a4937d448b34ee93bfeab1d6e97a94100845d839d3bcb\": container with ID starting with aa80cf31a849e912874a4937d448b34ee93bfeab1d6e97a94100845d839d3bcb not found: ID does not exist" Nov 25 16:12:12 crc kubenswrapper[4890]: I1125 16:12:12.188029 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1deef159-1a05-493e-aad2-c3467c5bd03b" path="/var/lib/kubelet/pods/1deef159-1a05-493e-aad2-c3467c5bd03b/volumes" Nov 25 16:13:28 crc kubenswrapper[4890]: I1125 16:13:28.612881 4890 generic.go:334] "Generic (PLEG): container finished" podID="745efd4c-6be9-4f7e-84d0-81154833dd25" containerID="bd5b95d54d1173336682277661b67f4c6ebde09df93ffb10b569c5c18ec42c4e" exitCode=0 Nov 25 16:13:28 crc kubenswrapper[4890]: I1125 16:13:28.612957 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-sv4x5/must-gather-t6wvd" event={"ID":"745efd4c-6be9-4f7e-84d0-81154833dd25","Type":"ContainerDied","Data":"bd5b95d54d1173336682277661b67f4c6ebde09df93ffb10b569c5c18ec42c4e"} Nov 25 16:13:28 crc kubenswrapper[4890]: I1125 16:13:28.613887 4890 scope.go:117] "RemoveContainer" containerID="bd5b95d54d1173336682277661b67f4c6ebde09df93ffb10b569c5c18ec42c4e" Nov 25 16:13:29 crc kubenswrapper[4890]: I1125 16:13:29.441717 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-sv4x5_must-gather-t6wvd_745efd4c-6be9-4f7e-84d0-81154833dd25/gather/0.log" Nov 25 16:13:37 crc kubenswrapper[4890]: I1125 16:13:37.970062 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-sv4x5/must-gather-t6wvd"] Nov 25 16:13:37 crc kubenswrapper[4890]: I1125 16:13:37.972279 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-sv4x5/must-gather-t6wvd" podUID="745efd4c-6be9-4f7e-84d0-81154833dd25" containerName="copy" containerID="cri-o://d5dfa60ad3022889688752f0425048acc7fc507e094c6c6f9fcd6c67ea81c8d6" gracePeriod=2 Nov 25 16:13:37 crc kubenswrapper[4890]: I1125 16:13:37.989368 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-sv4x5/must-gather-t6wvd"] Nov 25 16:13:38 crc kubenswrapper[4890]: I1125 16:13:38.455141 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-sv4x5_must-gather-t6wvd_745efd4c-6be9-4f7e-84d0-81154833dd25/copy/0.log" Nov 25 16:13:38 crc kubenswrapper[4890]: I1125 16:13:38.456136 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sv4x5/must-gather-t6wvd" Nov 25 16:13:38 crc kubenswrapper[4890]: I1125 16:13:38.494465 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cdvhk\" (UniqueName: \"kubernetes.io/projected/745efd4c-6be9-4f7e-84d0-81154833dd25-kube-api-access-cdvhk\") pod \"745efd4c-6be9-4f7e-84d0-81154833dd25\" (UID: \"745efd4c-6be9-4f7e-84d0-81154833dd25\") " Nov 25 16:13:38 crc kubenswrapper[4890]: I1125 16:13:38.494915 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/745efd4c-6be9-4f7e-84d0-81154833dd25-must-gather-output\") pod \"745efd4c-6be9-4f7e-84d0-81154833dd25\" (UID: \"745efd4c-6be9-4f7e-84d0-81154833dd25\") " Nov 25 16:13:38 crc kubenswrapper[4890]: I1125 16:13:38.516628 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/745efd4c-6be9-4f7e-84d0-81154833dd25-kube-api-access-cdvhk" (OuterVolumeSpecName: "kube-api-access-cdvhk") pod "745efd4c-6be9-4f7e-84d0-81154833dd25" (UID: "745efd4c-6be9-4f7e-84d0-81154833dd25"). InnerVolumeSpecName "kube-api-access-cdvhk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:13:38 crc kubenswrapper[4890]: I1125 16:13:38.600724 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cdvhk\" (UniqueName: \"kubernetes.io/projected/745efd4c-6be9-4f7e-84d0-81154833dd25-kube-api-access-cdvhk\") on node \"crc\" DevicePath \"\"" Nov 25 16:13:38 crc kubenswrapper[4890]: I1125 16:13:38.685566 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/745efd4c-6be9-4f7e-84d0-81154833dd25-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "745efd4c-6be9-4f7e-84d0-81154833dd25" (UID: "745efd4c-6be9-4f7e-84d0-81154833dd25"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:13:38 crc kubenswrapper[4890]: I1125 16:13:38.706655 4890 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/745efd4c-6be9-4f7e-84d0-81154833dd25-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 25 16:13:38 crc kubenswrapper[4890]: I1125 16:13:38.716101 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-sv4x5_must-gather-t6wvd_745efd4c-6be9-4f7e-84d0-81154833dd25/copy/0.log" Nov 25 16:13:38 crc kubenswrapper[4890]: I1125 16:13:38.716567 4890 generic.go:334] "Generic (PLEG): container finished" podID="745efd4c-6be9-4f7e-84d0-81154833dd25" containerID="d5dfa60ad3022889688752f0425048acc7fc507e094c6c6f9fcd6c67ea81c8d6" exitCode=143 Nov 25 16:13:38 crc kubenswrapper[4890]: I1125 16:13:38.716658 4890 scope.go:117] "RemoveContainer" containerID="d5dfa60ad3022889688752f0425048acc7fc507e094c6c6f9fcd6c67ea81c8d6" Nov 25 16:13:38 crc kubenswrapper[4890]: I1125 16:13:38.716947 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sv4x5/must-gather-t6wvd" Nov 25 16:13:38 crc kubenswrapper[4890]: I1125 16:13:38.771198 4890 scope.go:117] "RemoveContainer" containerID="bd5b95d54d1173336682277661b67f4c6ebde09df93ffb10b569c5c18ec42c4e" Nov 25 16:13:38 crc kubenswrapper[4890]: I1125 16:13:38.887907 4890 scope.go:117] "RemoveContainer" containerID="d5dfa60ad3022889688752f0425048acc7fc507e094c6c6f9fcd6c67ea81c8d6" Nov 25 16:13:38 crc kubenswrapper[4890]: E1125 16:13:38.888422 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5dfa60ad3022889688752f0425048acc7fc507e094c6c6f9fcd6c67ea81c8d6\": container with ID starting with d5dfa60ad3022889688752f0425048acc7fc507e094c6c6f9fcd6c67ea81c8d6 not found: ID does not exist" containerID="d5dfa60ad3022889688752f0425048acc7fc507e094c6c6f9fcd6c67ea81c8d6" Nov 25 16:13:38 crc kubenswrapper[4890]: I1125 16:13:38.888463 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5dfa60ad3022889688752f0425048acc7fc507e094c6c6f9fcd6c67ea81c8d6"} err="failed to get container status \"d5dfa60ad3022889688752f0425048acc7fc507e094c6c6f9fcd6c67ea81c8d6\": rpc error: code = NotFound desc = could not find container \"d5dfa60ad3022889688752f0425048acc7fc507e094c6c6f9fcd6c67ea81c8d6\": container with ID starting with d5dfa60ad3022889688752f0425048acc7fc507e094c6c6f9fcd6c67ea81c8d6 not found: ID does not exist" Nov 25 16:13:38 crc kubenswrapper[4890]: I1125 16:13:38.888492 4890 scope.go:117] "RemoveContainer" containerID="bd5b95d54d1173336682277661b67f4c6ebde09df93ffb10b569c5c18ec42c4e" Nov 25 16:13:38 crc kubenswrapper[4890]: E1125 16:13:38.888900 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd5b95d54d1173336682277661b67f4c6ebde09df93ffb10b569c5c18ec42c4e\": container with ID starting with bd5b95d54d1173336682277661b67f4c6ebde09df93ffb10b569c5c18ec42c4e not found: ID does not exist" containerID="bd5b95d54d1173336682277661b67f4c6ebde09df93ffb10b569c5c18ec42c4e" Nov 25 16:13:38 crc kubenswrapper[4890]: I1125 16:13:38.888933 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd5b95d54d1173336682277661b67f4c6ebde09df93ffb10b569c5c18ec42c4e"} err="failed to get container status \"bd5b95d54d1173336682277661b67f4c6ebde09df93ffb10b569c5c18ec42c4e\": rpc error: code = NotFound desc = could not find container \"bd5b95d54d1173336682277661b67f4c6ebde09df93ffb10b569c5c18ec42c4e\": container with ID starting with bd5b95d54d1173336682277661b67f4c6ebde09df93ffb10b569c5c18ec42c4e not found: ID does not exist" Nov 25 16:13:40 crc kubenswrapper[4890]: I1125 16:13:40.184888 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="745efd4c-6be9-4f7e-84d0-81154833dd25" path="/var/lib/kubelet/pods/745efd4c-6be9-4f7e-84d0-81154833dd25/volumes" Nov 25 16:13:56 crc kubenswrapper[4890]: I1125 16:13:56.447921 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 16:13:56 crc kubenswrapper[4890]: I1125 16:13:56.448633 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 16:14:26 crc kubenswrapper[4890]: I1125 16:14:26.447981 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 16:14:26 crc kubenswrapper[4890]: I1125 16:14:26.448796 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 16:14:56 crc kubenswrapper[4890]: I1125 16:14:56.447269 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 16:14:56 crc kubenswrapper[4890]: I1125 16:14:56.447844 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 16:14:56 crc kubenswrapper[4890]: I1125 16:14:56.447905 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" Nov 25 16:14:56 crc kubenswrapper[4890]: I1125 16:14:56.448959 4890 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"28fed48468d92f7b1fe5d54d4505526fd8660c4d76009eee3215a28f9eb145da"} pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 16:14:56 crc kubenswrapper[4890]: I1125 16:14:56.449038 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" containerID="cri-o://28fed48468d92f7b1fe5d54d4505526fd8660c4d76009eee3215a28f9eb145da" gracePeriod=600 Nov 25 16:14:56 crc kubenswrapper[4890]: E1125 16:14:56.574802 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 16:14:57 crc kubenswrapper[4890]: I1125 16:14:57.550218 4890 generic.go:334] "Generic (PLEG): container finished" podID="4e4f849d-f239-4727-a73e-18327856929a" containerID="28fed48468d92f7b1fe5d54d4505526fd8660c4d76009eee3215a28f9eb145da" exitCode=0 Nov 25 16:14:57 crc kubenswrapper[4890]: I1125 16:14:57.550386 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" event={"ID":"4e4f849d-f239-4727-a73e-18327856929a","Type":"ContainerDied","Data":"28fed48468d92f7b1fe5d54d4505526fd8660c4d76009eee3215a28f9eb145da"} Nov 25 16:14:57 crc kubenswrapper[4890]: I1125 16:14:57.550586 4890 scope.go:117] "RemoveContainer" containerID="0ee5f289dd255f61c3f8c4e86c84d7044d423518d51541add4da1a581b273eb1" Nov 25 16:14:57 crc kubenswrapper[4890]: I1125 16:14:57.551382 4890 scope.go:117] "RemoveContainer" containerID="28fed48468d92f7b1fe5d54d4505526fd8660c4d76009eee3215a28f9eb145da" Nov 25 16:14:57 crc kubenswrapper[4890]: E1125 16:14:57.551633 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 16:15:00 crc kubenswrapper[4890]: I1125 16:15:00.156431 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401455-8j5qm"] Nov 25 16:15:00 crc kubenswrapper[4890]: E1125 16:15:00.157448 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1deef159-1a05-493e-aad2-c3467c5bd03b" containerName="extract-utilities" Nov 25 16:15:00 crc kubenswrapper[4890]: I1125 16:15:00.157465 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="1deef159-1a05-493e-aad2-c3467c5bd03b" containerName="extract-utilities" Nov 25 16:15:00 crc kubenswrapper[4890]: E1125 16:15:00.157495 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="745efd4c-6be9-4f7e-84d0-81154833dd25" containerName="gather" Nov 25 16:15:00 crc kubenswrapper[4890]: I1125 16:15:00.157504 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="745efd4c-6be9-4f7e-84d0-81154833dd25" containerName="gather" Nov 25 16:15:00 crc kubenswrapper[4890]: E1125 16:15:00.157518 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1deef159-1a05-493e-aad2-c3467c5bd03b" containerName="extract-content" Nov 25 16:15:00 crc kubenswrapper[4890]: I1125 16:15:00.157526 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="1deef159-1a05-493e-aad2-c3467c5bd03b" containerName="extract-content" Nov 25 16:15:00 crc kubenswrapper[4890]: E1125 16:15:00.157549 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="745efd4c-6be9-4f7e-84d0-81154833dd25" containerName="copy" Nov 25 16:15:00 crc kubenswrapper[4890]: I1125 16:15:00.157556 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="745efd4c-6be9-4f7e-84d0-81154833dd25" containerName="copy" Nov 25 16:15:00 crc kubenswrapper[4890]: E1125 16:15:00.157594 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1deef159-1a05-493e-aad2-c3467c5bd03b" containerName="registry-server" Nov 25 16:15:00 crc kubenswrapper[4890]: I1125 16:15:00.157601 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="1deef159-1a05-493e-aad2-c3467c5bd03b" containerName="registry-server" Nov 25 16:15:00 crc kubenswrapper[4890]: I1125 16:15:00.157840 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="745efd4c-6be9-4f7e-84d0-81154833dd25" containerName="copy" Nov 25 16:15:00 crc kubenswrapper[4890]: I1125 16:15:00.157867 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="745efd4c-6be9-4f7e-84d0-81154833dd25" containerName="gather" Nov 25 16:15:00 crc kubenswrapper[4890]: I1125 16:15:00.157877 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="1deef159-1a05-493e-aad2-c3467c5bd03b" containerName="registry-server" Nov 25 16:15:00 crc kubenswrapper[4890]: I1125 16:15:00.158731 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401455-8j5qm" Nov 25 16:15:00 crc kubenswrapper[4890]: I1125 16:15:00.168598 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 16:15:00 crc kubenswrapper[4890]: I1125 16:15:00.168656 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 16:15:00 crc kubenswrapper[4890]: I1125 16:15:00.169171 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401455-8j5qm"] Nov 25 16:15:00 crc kubenswrapper[4890]: I1125 16:15:00.217626 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgnhp\" (UniqueName: \"kubernetes.io/projected/0374e2f5-9fd1-406b-8b0c-cff235ef4946-kube-api-access-mgnhp\") pod \"collect-profiles-29401455-8j5qm\" (UID: \"0374e2f5-9fd1-406b-8b0c-cff235ef4946\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401455-8j5qm" Nov 25 16:15:00 crc kubenswrapper[4890]: I1125 16:15:00.217760 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0374e2f5-9fd1-406b-8b0c-cff235ef4946-secret-volume\") pod \"collect-profiles-29401455-8j5qm\" (UID: \"0374e2f5-9fd1-406b-8b0c-cff235ef4946\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401455-8j5qm" Nov 25 16:15:00 crc kubenswrapper[4890]: I1125 16:15:00.217989 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0374e2f5-9fd1-406b-8b0c-cff235ef4946-config-volume\") pod \"collect-profiles-29401455-8j5qm\" (UID: \"0374e2f5-9fd1-406b-8b0c-cff235ef4946\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401455-8j5qm" Nov 25 16:15:00 crc kubenswrapper[4890]: I1125 16:15:00.320229 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0374e2f5-9fd1-406b-8b0c-cff235ef4946-secret-volume\") pod \"collect-profiles-29401455-8j5qm\" (UID: \"0374e2f5-9fd1-406b-8b0c-cff235ef4946\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401455-8j5qm" Nov 25 16:15:00 crc kubenswrapper[4890]: I1125 16:15:00.320377 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0374e2f5-9fd1-406b-8b0c-cff235ef4946-config-volume\") pod \"collect-profiles-29401455-8j5qm\" (UID: \"0374e2f5-9fd1-406b-8b0c-cff235ef4946\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401455-8j5qm" Nov 25 16:15:00 crc kubenswrapper[4890]: I1125 16:15:00.320493 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgnhp\" (UniqueName: \"kubernetes.io/projected/0374e2f5-9fd1-406b-8b0c-cff235ef4946-kube-api-access-mgnhp\") pod \"collect-profiles-29401455-8j5qm\" (UID: \"0374e2f5-9fd1-406b-8b0c-cff235ef4946\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401455-8j5qm" Nov 25 16:15:00 crc kubenswrapper[4890]: I1125 16:15:00.321937 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0374e2f5-9fd1-406b-8b0c-cff235ef4946-config-volume\") pod \"collect-profiles-29401455-8j5qm\" (UID: \"0374e2f5-9fd1-406b-8b0c-cff235ef4946\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401455-8j5qm" Nov 25 16:15:00 crc kubenswrapper[4890]: I1125 16:15:00.327222 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0374e2f5-9fd1-406b-8b0c-cff235ef4946-secret-volume\") pod \"collect-profiles-29401455-8j5qm\" (UID: \"0374e2f5-9fd1-406b-8b0c-cff235ef4946\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401455-8j5qm" Nov 25 16:15:00 crc kubenswrapper[4890]: I1125 16:15:00.338381 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgnhp\" (UniqueName: \"kubernetes.io/projected/0374e2f5-9fd1-406b-8b0c-cff235ef4946-kube-api-access-mgnhp\") pod \"collect-profiles-29401455-8j5qm\" (UID: \"0374e2f5-9fd1-406b-8b0c-cff235ef4946\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401455-8j5qm" Nov 25 16:15:00 crc kubenswrapper[4890]: I1125 16:15:00.485096 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401455-8j5qm" Nov 25 16:15:00 crc kubenswrapper[4890]: I1125 16:15:00.989829 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401455-8j5qm"] Nov 25 16:15:01 crc kubenswrapper[4890]: I1125 16:15:01.597944 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401455-8j5qm" event={"ID":"0374e2f5-9fd1-406b-8b0c-cff235ef4946","Type":"ContainerStarted","Data":"3d1fdf14c1eb41868296d2f5c2b8fe9f71eefda85d5f94a8bcf66c1092f084e8"} Nov 25 16:15:01 crc kubenswrapper[4890]: I1125 16:15:01.597999 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401455-8j5qm" event={"ID":"0374e2f5-9fd1-406b-8b0c-cff235ef4946","Type":"ContainerStarted","Data":"b504a03e77542663deededb6aee002636053cefda15daa8ce97d94c1364d324a"} Nov 25 16:15:01 crc kubenswrapper[4890]: I1125 16:15:01.618593 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29401455-8j5qm" podStartSLOduration=1.6185750269999999 podStartE2EDuration="1.618575027s" podCreationTimestamp="2025-11-25 16:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:15:01.614571697 +0000 UTC m=+4360.057034307" watchObservedRunningTime="2025-11-25 16:15:01.618575027 +0000 UTC m=+4360.061037637" Nov 25 16:15:02 crc kubenswrapper[4890]: I1125 16:15:02.609170 4890 generic.go:334] "Generic (PLEG): container finished" podID="0374e2f5-9fd1-406b-8b0c-cff235ef4946" containerID="3d1fdf14c1eb41868296d2f5c2b8fe9f71eefda85d5f94a8bcf66c1092f084e8" exitCode=0 Nov 25 16:15:02 crc kubenswrapper[4890]: I1125 16:15:02.609261 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401455-8j5qm" event={"ID":"0374e2f5-9fd1-406b-8b0c-cff235ef4946","Type":"ContainerDied","Data":"3d1fdf14c1eb41868296d2f5c2b8fe9f71eefda85d5f94a8bcf66c1092f084e8"} Nov 25 16:15:04 crc kubenswrapper[4890]: I1125 16:15:04.429765 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401455-8j5qm" Nov 25 16:15:04 crc kubenswrapper[4890]: I1125 16:15:04.510423 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0374e2f5-9fd1-406b-8b0c-cff235ef4946-config-volume\") pod \"0374e2f5-9fd1-406b-8b0c-cff235ef4946\" (UID: \"0374e2f5-9fd1-406b-8b0c-cff235ef4946\") " Nov 25 16:15:04 crc kubenswrapper[4890]: I1125 16:15:04.510994 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0374e2f5-9fd1-406b-8b0c-cff235ef4946-secret-volume\") pod \"0374e2f5-9fd1-406b-8b0c-cff235ef4946\" (UID: \"0374e2f5-9fd1-406b-8b0c-cff235ef4946\") " Nov 25 16:15:04 crc kubenswrapper[4890]: I1125 16:15:04.511127 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mgnhp\" (UniqueName: \"kubernetes.io/projected/0374e2f5-9fd1-406b-8b0c-cff235ef4946-kube-api-access-mgnhp\") pod \"0374e2f5-9fd1-406b-8b0c-cff235ef4946\" (UID: \"0374e2f5-9fd1-406b-8b0c-cff235ef4946\") " Nov 25 16:15:04 crc kubenswrapper[4890]: I1125 16:15:04.512018 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0374e2f5-9fd1-406b-8b0c-cff235ef4946-config-volume" (OuterVolumeSpecName: "config-volume") pod "0374e2f5-9fd1-406b-8b0c-cff235ef4946" (UID: "0374e2f5-9fd1-406b-8b0c-cff235ef4946"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 16:15:04 crc kubenswrapper[4890]: I1125 16:15:04.522562 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0374e2f5-9fd1-406b-8b0c-cff235ef4946-kube-api-access-mgnhp" (OuterVolumeSpecName: "kube-api-access-mgnhp") pod "0374e2f5-9fd1-406b-8b0c-cff235ef4946" (UID: "0374e2f5-9fd1-406b-8b0c-cff235ef4946"). InnerVolumeSpecName "kube-api-access-mgnhp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:15:04 crc kubenswrapper[4890]: I1125 16:15:04.522557 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0374e2f5-9fd1-406b-8b0c-cff235ef4946-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0374e2f5-9fd1-406b-8b0c-cff235ef4946" (UID: "0374e2f5-9fd1-406b-8b0c-cff235ef4946"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 16:15:04 crc kubenswrapper[4890]: I1125 16:15:04.613677 4890 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0374e2f5-9fd1-406b-8b0c-cff235ef4946-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 16:15:04 crc kubenswrapper[4890]: I1125 16:15:04.613729 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mgnhp\" (UniqueName: \"kubernetes.io/projected/0374e2f5-9fd1-406b-8b0c-cff235ef4946-kube-api-access-mgnhp\") on node \"crc\" DevicePath \"\"" Nov 25 16:15:04 crc kubenswrapper[4890]: I1125 16:15:04.613741 4890 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0374e2f5-9fd1-406b-8b0c-cff235ef4946-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 16:15:04 crc kubenswrapper[4890]: I1125 16:15:04.632980 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401455-8j5qm" event={"ID":"0374e2f5-9fd1-406b-8b0c-cff235ef4946","Type":"ContainerDied","Data":"b504a03e77542663deededb6aee002636053cefda15daa8ce97d94c1364d324a"} Nov 25 16:15:04 crc kubenswrapper[4890]: I1125 16:15:04.633057 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b504a03e77542663deededb6aee002636053cefda15daa8ce97d94c1364d324a" Nov 25 16:15:04 crc kubenswrapper[4890]: I1125 16:15:04.633412 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401455-8j5qm" Nov 25 16:15:04 crc kubenswrapper[4890]: I1125 16:15:04.711364 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401410-ngdkg"] Nov 25 16:15:04 crc kubenswrapper[4890]: I1125 16:15:04.724651 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401410-ngdkg"] Nov 25 16:15:06 crc kubenswrapper[4890]: I1125 16:15:06.184826 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95656df6-ba37-41ef-a281-87240f41c544" path="/var/lib/kubelet/pods/95656df6-ba37-41ef-a281-87240f41c544/volumes" Nov 25 16:15:12 crc kubenswrapper[4890]: I1125 16:15:12.181798 4890 scope.go:117] "RemoveContainer" containerID="28fed48468d92f7b1fe5d54d4505526fd8660c4d76009eee3215a28f9eb145da" Nov 25 16:15:12 crc kubenswrapper[4890]: E1125 16:15:12.183320 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 16:15:26 crc kubenswrapper[4890]: I1125 16:15:26.173591 4890 scope.go:117] "RemoveContainer" containerID="28fed48468d92f7b1fe5d54d4505526fd8660c4d76009eee3215a28f9eb145da" Nov 25 16:15:26 crc kubenswrapper[4890]: E1125 16:15:26.174401 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 16:15:40 crc kubenswrapper[4890]: I1125 16:15:40.178533 4890 scope.go:117] "RemoveContainer" containerID="28fed48468d92f7b1fe5d54d4505526fd8660c4d76009eee3215a28f9eb145da" Nov 25 16:15:40 crc kubenswrapper[4890]: E1125 16:15:40.179337 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 16:15:40 crc kubenswrapper[4890]: I1125 16:15:40.823779 4890 scope.go:117] "RemoveContainer" containerID="586f0cfcafc29ab77e85b46b9543b5054cba15609dc50e05be6900565ee49a9f" Nov 25 16:15:52 crc kubenswrapper[4890]: I1125 16:15:52.181860 4890 scope.go:117] "RemoveContainer" containerID="28fed48468d92f7b1fe5d54d4505526fd8660c4d76009eee3215a28f9eb145da" Nov 25 16:15:52 crc kubenswrapper[4890]: E1125 16:15:52.182722 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 16:16:07 crc kubenswrapper[4890]: I1125 16:16:07.173411 4890 scope.go:117] "RemoveContainer" containerID="28fed48468d92f7b1fe5d54d4505526fd8660c4d76009eee3215a28f9eb145da" Nov 25 16:16:07 crc kubenswrapper[4890]: E1125 16:16:07.174323 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 16:16:20 crc kubenswrapper[4890]: I1125 16:16:20.173829 4890 scope.go:117] "RemoveContainer" containerID="28fed48468d92f7b1fe5d54d4505526fd8660c4d76009eee3215a28f9eb145da" Nov 25 16:16:20 crc kubenswrapper[4890]: E1125 16:16:20.174677 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 16:16:35 crc kubenswrapper[4890]: I1125 16:16:35.173678 4890 scope.go:117] "RemoveContainer" containerID="28fed48468d92f7b1fe5d54d4505526fd8660c4d76009eee3215a28f9eb145da" Nov 25 16:16:35 crc kubenswrapper[4890]: E1125 16:16:35.174467 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 16:16:40 crc kubenswrapper[4890]: I1125 16:16:40.886343 4890 scope.go:117] "RemoveContainer" containerID="803b18f295753dbaca654c84cfd05616cce379be16a3587b2f20c531368ada81" Nov 25 16:16:42 crc kubenswrapper[4890]: I1125 16:16:42.553443 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-t78gs/must-gather-5gfdb"] Nov 25 16:16:42 crc kubenswrapper[4890]: E1125 16:16:42.554231 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0374e2f5-9fd1-406b-8b0c-cff235ef4946" containerName="collect-profiles" Nov 25 16:16:42 crc kubenswrapper[4890]: I1125 16:16:42.554243 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="0374e2f5-9fd1-406b-8b0c-cff235ef4946" containerName="collect-profiles" Nov 25 16:16:42 crc kubenswrapper[4890]: I1125 16:16:42.554457 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="0374e2f5-9fd1-406b-8b0c-cff235ef4946" containerName="collect-profiles" Nov 25 16:16:42 crc kubenswrapper[4890]: I1125 16:16:42.555449 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t78gs/must-gather-5gfdb" Nov 25 16:16:42 crc kubenswrapper[4890]: I1125 16:16:42.558087 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-t78gs"/"openshift-service-ca.crt" Nov 25 16:16:42 crc kubenswrapper[4890]: I1125 16:16:42.564423 4890 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-t78gs"/"kube-root-ca.crt" Nov 25 16:16:42 crc kubenswrapper[4890]: I1125 16:16:42.577476 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-t78gs/must-gather-5gfdb"] Nov 25 16:16:42 crc kubenswrapper[4890]: I1125 16:16:42.582562 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cf177367-00de-4aa9-a1ff-124727968f00-must-gather-output\") pod \"must-gather-5gfdb\" (UID: \"cf177367-00de-4aa9-a1ff-124727968f00\") " pod="openshift-must-gather-t78gs/must-gather-5gfdb" Nov 25 16:16:42 crc kubenswrapper[4890]: I1125 16:16:42.582652 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vn59\" (UniqueName: \"kubernetes.io/projected/cf177367-00de-4aa9-a1ff-124727968f00-kube-api-access-4vn59\") pod \"must-gather-5gfdb\" (UID: \"cf177367-00de-4aa9-a1ff-124727968f00\") " pod="openshift-must-gather-t78gs/must-gather-5gfdb" Nov 25 16:16:42 crc kubenswrapper[4890]: I1125 16:16:42.683881 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vn59\" (UniqueName: \"kubernetes.io/projected/cf177367-00de-4aa9-a1ff-124727968f00-kube-api-access-4vn59\") pod \"must-gather-5gfdb\" (UID: \"cf177367-00de-4aa9-a1ff-124727968f00\") " pod="openshift-must-gather-t78gs/must-gather-5gfdb" Nov 25 16:16:42 crc kubenswrapper[4890]: I1125 16:16:42.684009 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cf177367-00de-4aa9-a1ff-124727968f00-must-gather-output\") pod \"must-gather-5gfdb\" (UID: \"cf177367-00de-4aa9-a1ff-124727968f00\") " pod="openshift-must-gather-t78gs/must-gather-5gfdb" Nov 25 16:16:42 crc kubenswrapper[4890]: I1125 16:16:42.684476 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cf177367-00de-4aa9-a1ff-124727968f00-must-gather-output\") pod \"must-gather-5gfdb\" (UID: \"cf177367-00de-4aa9-a1ff-124727968f00\") " pod="openshift-must-gather-t78gs/must-gather-5gfdb" Nov 25 16:16:42 crc kubenswrapper[4890]: I1125 16:16:42.710975 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vn59\" (UniqueName: \"kubernetes.io/projected/cf177367-00de-4aa9-a1ff-124727968f00-kube-api-access-4vn59\") pod \"must-gather-5gfdb\" (UID: \"cf177367-00de-4aa9-a1ff-124727968f00\") " pod="openshift-must-gather-t78gs/must-gather-5gfdb" Nov 25 16:16:42 crc kubenswrapper[4890]: I1125 16:16:42.886074 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t78gs/must-gather-5gfdb" Nov 25 16:16:43 crc kubenswrapper[4890]: I1125 16:16:43.381837 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-t78gs/must-gather-5gfdb"] Nov 25 16:16:43 crc kubenswrapper[4890]: I1125 16:16:43.607979 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t78gs/must-gather-5gfdb" event={"ID":"cf177367-00de-4aa9-a1ff-124727968f00","Type":"ContainerStarted","Data":"48ec0fbdd65b5fa5cc2537c18c6659c899b0cfb9be6955a82405193671e7838c"} Nov 25 16:16:44 crc kubenswrapper[4890]: I1125 16:16:44.619003 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t78gs/must-gather-5gfdb" event={"ID":"cf177367-00de-4aa9-a1ff-124727968f00","Type":"ContainerStarted","Data":"9badd2e0e1ceb14d744cf5207ffbc0d863dff27bf2b9ee43bb0c325621e13a94"} Nov 25 16:16:44 crc kubenswrapper[4890]: I1125 16:16:44.621001 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t78gs/must-gather-5gfdb" event={"ID":"cf177367-00de-4aa9-a1ff-124727968f00","Type":"ContainerStarted","Data":"01d2f28946f5e78f665617387571ce72aae378ec04839e128c4e8c9af3d58924"} Nov 25 16:16:44 crc kubenswrapper[4890]: I1125 16:16:44.648418 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-t78gs/must-gather-5gfdb" podStartSLOduration=2.648396839 podStartE2EDuration="2.648396839s" podCreationTimestamp="2025-11-25 16:16:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:16:44.644281407 +0000 UTC m=+4463.086744017" watchObservedRunningTime="2025-11-25 16:16:44.648396839 +0000 UTC m=+4463.090859459" Nov 25 16:16:48 crc kubenswrapper[4890]: I1125 16:16:48.381488 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-t78gs/crc-debug-cqdzv"] Nov 25 16:16:48 crc kubenswrapper[4890]: I1125 16:16:48.383467 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t78gs/crc-debug-cqdzv" Nov 25 16:16:48 crc kubenswrapper[4890]: I1125 16:16:48.385217 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-t78gs"/"default-dockercfg-jgmdk" Nov 25 16:16:48 crc kubenswrapper[4890]: I1125 16:16:48.411475 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/80e91bdf-ca56-4a0f-8090-0262b5ddd621-host\") pod \"crc-debug-cqdzv\" (UID: \"80e91bdf-ca56-4a0f-8090-0262b5ddd621\") " pod="openshift-must-gather-t78gs/crc-debug-cqdzv" Nov 25 16:16:48 crc kubenswrapper[4890]: I1125 16:16:48.411877 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2lnt\" (UniqueName: \"kubernetes.io/projected/80e91bdf-ca56-4a0f-8090-0262b5ddd621-kube-api-access-c2lnt\") pod \"crc-debug-cqdzv\" (UID: \"80e91bdf-ca56-4a0f-8090-0262b5ddd621\") " pod="openshift-must-gather-t78gs/crc-debug-cqdzv" Nov 25 16:16:48 crc kubenswrapper[4890]: I1125 16:16:48.514131 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/80e91bdf-ca56-4a0f-8090-0262b5ddd621-host\") pod \"crc-debug-cqdzv\" (UID: \"80e91bdf-ca56-4a0f-8090-0262b5ddd621\") " pod="openshift-must-gather-t78gs/crc-debug-cqdzv" Nov 25 16:16:48 crc kubenswrapper[4890]: I1125 16:16:48.514242 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/80e91bdf-ca56-4a0f-8090-0262b5ddd621-host\") pod \"crc-debug-cqdzv\" (UID: \"80e91bdf-ca56-4a0f-8090-0262b5ddd621\") " pod="openshift-must-gather-t78gs/crc-debug-cqdzv" Nov 25 16:16:48 crc kubenswrapper[4890]: I1125 16:16:48.514355 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2lnt\" (UniqueName: \"kubernetes.io/projected/80e91bdf-ca56-4a0f-8090-0262b5ddd621-kube-api-access-c2lnt\") pod \"crc-debug-cqdzv\" (UID: \"80e91bdf-ca56-4a0f-8090-0262b5ddd621\") " pod="openshift-must-gather-t78gs/crc-debug-cqdzv" Nov 25 16:16:48 crc kubenswrapper[4890]: I1125 16:16:48.536879 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2lnt\" (UniqueName: \"kubernetes.io/projected/80e91bdf-ca56-4a0f-8090-0262b5ddd621-kube-api-access-c2lnt\") pod \"crc-debug-cqdzv\" (UID: \"80e91bdf-ca56-4a0f-8090-0262b5ddd621\") " pod="openshift-must-gather-t78gs/crc-debug-cqdzv" Nov 25 16:16:48 crc kubenswrapper[4890]: I1125 16:16:48.704927 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t78gs/crc-debug-cqdzv" Nov 25 16:16:49 crc kubenswrapper[4890]: I1125 16:16:49.173480 4890 scope.go:117] "RemoveContainer" containerID="28fed48468d92f7b1fe5d54d4505526fd8660c4d76009eee3215a28f9eb145da" Nov 25 16:16:49 crc kubenswrapper[4890]: E1125 16:16:49.173771 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 16:16:49 crc kubenswrapper[4890]: I1125 16:16:49.669708 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t78gs/crc-debug-cqdzv" event={"ID":"80e91bdf-ca56-4a0f-8090-0262b5ddd621","Type":"ContainerStarted","Data":"8d6c037730429a01e68fa7622e1b7c5fdbc0ffaa7e00575d34401c31641afed1"} Nov 25 16:16:49 crc kubenswrapper[4890]: I1125 16:16:49.670089 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t78gs/crc-debug-cqdzv" event={"ID":"80e91bdf-ca56-4a0f-8090-0262b5ddd621","Type":"ContainerStarted","Data":"3710a8cead586d987c0176d36b8eb16c0fdb9625f9a54527c9caaedc099b79c5"} Nov 25 16:16:49 crc kubenswrapper[4890]: I1125 16:16:49.691194 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-t78gs/crc-debug-cqdzv" podStartSLOduration=1.691173705 podStartE2EDuration="1.691173705s" podCreationTimestamp="2025-11-25 16:16:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 16:16:49.682363826 +0000 UTC m=+4468.124826456" watchObservedRunningTime="2025-11-25 16:16:49.691173705 +0000 UTC m=+4468.133636315" Nov 25 16:17:01 crc kubenswrapper[4890]: I1125 16:17:01.174474 4890 scope.go:117] "RemoveContainer" containerID="28fed48468d92f7b1fe5d54d4505526fd8660c4d76009eee3215a28f9eb145da" Nov 25 16:17:01 crc kubenswrapper[4890]: E1125 16:17:01.175492 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 16:17:09 crc kubenswrapper[4890]: I1125 16:17:09.814382 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-t7wg4"] Nov 25 16:17:09 crc kubenswrapper[4890]: I1125 16:17:09.818057 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t7wg4" Nov 25 16:17:09 crc kubenswrapper[4890]: I1125 16:17:09.829802 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-t7wg4"] Nov 25 16:17:09 crc kubenswrapper[4890]: I1125 16:17:09.944283 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/204f9c68-d5a6-4e95-9b16-21747c2ab45b-catalog-content\") pod \"redhat-marketplace-t7wg4\" (UID: \"204f9c68-d5a6-4e95-9b16-21747c2ab45b\") " pod="openshift-marketplace/redhat-marketplace-t7wg4" Nov 25 16:17:09 crc kubenswrapper[4890]: I1125 16:17:09.944674 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfnnr\" (UniqueName: \"kubernetes.io/projected/204f9c68-d5a6-4e95-9b16-21747c2ab45b-kube-api-access-tfnnr\") pod \"redhat-marketplace-t7wg4\" (UID: \"204f9c68-d5a6-4e95-9b16-21747c2ab45b\") " pod="openshift-marketplace/redhat-marketplace-t7wg4" Nov 25 16:17:09 crc kubenswrapper[4890]: I1125 16:17:09.944832 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/204f9c68-d5a6-4e95-9b16-21747c2ab45b-utilities\") pod \"redhat-marketplace-t7wg4\" (UID: \"204f9c68-d5a6-4e95-9b16-21747c2ab45b\") " pod="openshift-marketplace/redhat-marketplace-t7wg4" Nov 25 16:17:10 crc kubenswrapper[4890]: I1125 16:17:10.046738 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfnnr\" (UniqueName: \"kubernetes.io/projected/204f9c68-d5a6-4e95-9b16-21747c2ab45b-kube-api-access-tfnnr\") pod \"redhat-marketplace-t7wg4\" (UID: \"204f9c68-d5a6-4e95-9b16-21747c2ab45b\") " pod="openshift-marketplace/redhat-marketplace-t7wg4" Nov 25 16:17:10 crc kubenswrapper[4890]: I1125 16:17:10.046832 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/204f9c68-d5a6-4e95-9b16-21747c2ab45b-utilities\") pod \"redhat-marketplace-t7wg4\" (UID: \"204f9c68-d5a6-4e95-9b16-21747c2ab45b\") " pod="openshift-marketplace/redhat-marketplace-t7wg4" Nov 25 16:17:10 crc kubenswrapper[4890]: I1125 16:17:10.046895 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/204f9c68-d5a6-4e95-9b16-21747c2ab45b-catalog-content\") pod \"redhat-marketplace-t7wg4\" (UID: \"204f9c68-d5a6-4e95-9b16-21747c2ab45b\") " pod="openshift-marketplace/redhat-marketplace-t7wg4" Nov 25 16:17:10 crc kubenswrapper[4890]: I1125 16:17:10.047568 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/204f9c68-d5a6-4e95-9b16-21747c2ab45b-catalog-content\") pod \"redhat-marketplace-t7wg4\" (UID: \"204f9c68-d5a6-4e95-9b16-21747c2ab45b\") " pod="openshift-marketplace/redhat-marketplace-t7wg4" Nov 25 16:17:10 crc kubenswrapper[4890]: I1125 16:17:10.047623 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/204f9c68-d5a6-4e95-9b16-21747c2ab45b-utilities\") pod \"redhat-marketplace-t7wg4\" (UID: \"204f9c68-d5a6-4e95-9b16-21747c2ab45b\") " pod="openshift-marketplace/redhat-marketplace-t7wg4" Nov 25 16:17:10 crc kubenswrapper[4890]: I1125 16:17:10.071955 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfnnr\" (UniqueName: \"kubernetes.io/projected/204f9c68-d5a6-4e95-9b16-21747c2ab45b-kube-api-access-tfnnr\") pod \"redhat-marketplace-t7wg4\" (UID: \"204f9c68-d5a6-4e95-9b16-21747c2ab45b\") " pod="openshift-marketplace/redhat-marketplace-t7wg4" Nov 25 16:17:10 crc kubenswrapper[4890]: I1125 16:17:10.143323 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t7wg4" Nov 25 16:17:10 crc kubenswrapper[4890]: I1125 16:17:10.694691 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-t7wg4"] Nov 25 16:17:10 crc kubenswrapper[4890]: I1125 16:17:10.895051 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t7wg4" event={"ID":"204f9c68-d5a6-4e95-9b16-21747c2ab45b","Type":"ContainerStarted","Data":"1ecd97ad69653420e469a8d4786cc9cf0136a117ff110868727a372b5ee31565"} Nov 25 16:17:11 crc kubenswrapper[4890]: I1125 16:17:11.906192 4890 generic.go:334] "Generic (PLEG): container finished" podID="204f9c68-d5a6-4e95-9b16-21747c2ab45b" containerID="e1bd6a1774d26552f8525f3eafe25f1496786ff5a7ae110f9d14a1fde4c1dd87" exitCode=0 Nov 25 16:17:11 crc kubenswrapper[4890]: I1125 16:17:11.906277 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t7wg4" event={"ID":"204f9c68-d5a6-4e95-9b16-21747c2ab45b","Type":"ContainerDied","Data":"e1bd6a1774d26552f8525f3eafe25f1496786ff5a7ae110f9d14a1fde4c1dd87"} Nov 25 16:17:11 crc kubenswrapper[4890]: I1125 16:17:11.908327 4890 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 16:17:12 crc kubenswrapper[4890]: I1125 16:17:12.179780 4890 scope.go:117] "RemoveContainer" containerID="28fed48468d92f7b1fe5d54d4505526fd8660c4d76009eee3215a28f9eb145da" Nov 25 16:17:12 crc kubenswrapper[4890]: E1125 16:17:12.180045 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 16:17:12 crc kubenswrapper[4890]: I1125 16:17:12.917715 4890 generic.go:334] "Generic (PLEG): container finished" podID="204f9c68-d5a6-4e95-9b16-21747c2ab45b" containerID="7ec89a16e4bbe11c32036d6192d3bada4ca09b93f7b59d6888c29fbfa91fed48" exitCode=0 Nov 25 16:17:12 crc kubenswrapper[4890]: I1125 16:17:12.917757 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t7wg4" event={"ID":"204f9c68-d5a6-4e95-9b16-21747c2ab45b","Type":"ContainerDied","Data":"7ec89a16e4bbe11c32036d6192d3bada4ca09b93f7b59d6888c29fbfa91fed48"} Nov 25 16:17:13 crc kubenswrapper[4890]: I1125 16:17:13.930677 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t7wg4" event={"ID":"204f9c68-d5a6-4e95-9b16-21747c2ab45b","Type":"ContainerStarted","Data":"9ef2f0bf5a16e7be7cc093221b39a02f81a5436a30127f021a743e8757932863"} Nov 25 16:17:13 crc kubenswrapper[4890]: I1125 16:17:13.968182 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-t7wg4" podStartSLOduration=3.531487947 podStartE2EDuration="4.968140618s" podCreationTimestamp="2025-11-25 16:17:09 +0000 UTC" firstStartedPulling="2025-11-25 16:17:11.907931372 +0000 UTC m=+4490.350393992" lastFinishedPulling="2025-11-25 16:17:13.344584053 +0000 UTC m=+4491.787046663" observedRunningTime="2025-11-25 16:17:13.957440182 +0000 UTC m=+4492.399902802" watchObservedRunningTime="2025-11-25 16:17:13.968140618 +0000 UTC m=+4492.410603238" Nov 25 16:17:20 crc kubenswrapper[4890]: I1125 16:17:20.143515 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-t7wg4" Nov 25 16:17:20 crc kubenswrapper[4890]: I1125 16:17:20.144529 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-t7wg4" Nov 25 16:17:20 crc kubenswrapper[4890]: I1125 16:17:20.208078 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-t7wg4" Nov 25 16:17:21 crc kubenswrapper[4890]: I1125 16:17:21.051913 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-t7wg4" Nov 25 16:17:21 crc kubenswrapper[4890]: I1125 16:17:21.132581 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-t7wg4"] Nov 25 16:17:23 crc kubenswrapper[4890]: I1125 16:17:23.030101 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-t7wg4" podUID="204f9c68-d5a6-4e95-9b16-21747c2ab45b" containerName="registry-server" containerID="cri-o://9ef2f0bf5a16e7be7cc093221b39a02f81a5436a30127f021a743e8757932863" gracePeriod=2 Nov 25 16:17:23 crc kubenswrapper[4890]: I1125 16:17:23.571882 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t7wg4" Nov 25 16:17:23 crc kubenswrapper[4890]: I1125 16:17:23.740782 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tfnnr\" (UniqueName: \"kubernetes.io/projected/204f9c68-d5a6-4e95-9b16-21747c2ab45b-kube-api-access-tfnnr\") pod \"204f9c68-d5a6-4e95-9b16-21747c2ab45b\" (UID: \"204f9c68-d5a6-4e95-9b16-21747c2ab45b\") " Nov 25 16:17:23 crc kubenswrapper[4890]: I1125 16:17:23.740922 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/204f9c68-d5a6-4e95-9b16-21747c2ab45b-catalog-content\") pod \"204f9c68-d5a6-4e95-9b16-21747c2ab45b\" (UID: \"204f9c68-d5a6-4e95-9b16-21747c2ab45b\") " Nov 25 16:17:23 crc kubenswrapper[4890]: I1125 16:17:23.741178 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/204f9c68-d5a6-4e95-9b16-21747c2ab45b-utilities\") pod \"204f9c68-d5a6-4e95-9b16-21747c2ab45b\" (UID: \"204f9c68-d5a6-4e95-9b16-21747c2ab45b\") " Nov 25 16:17:23 crc kubenswrapper[4890]: I1125 16:17:23.742463 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/204f9c68-d5a6-4e95-9b16-21747c2ab45b-utilities" (OuterVolumeSpecName: "utilities") pod "204f9c68-d5a6-4e95-9b16-21747c2ab45b" (UID: "204f9c68-d5a6-4e95-9b16-21747c2ab45b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:17:23 crc kubenswrapper[4890]: I1125 16:17:23.753404 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/204f9c68-d5a6-4e95-9b16-21747c2ab45b-kube-api-access-tfnnr" (OuterVolumeSpecName: "kube-api-access-tfnnr") pod "204f9c68-d5a6-4e95-9b16-21747c2ab45b" (UID: "204f9c68-d5a6-4e95-9b16-21747c2ab45b"). InnerVolumeSpecName "kube-api-access-tfnnr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:17:23 crc kubenswrapper[4890]: I1125 16:17:23.771793 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/204f9c68-d5a6-4e95-9b16-21747c2ab45b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "204f9c68-d5a6-4e95-9b16-21747c2ab45b" (UID: "204f9c68-d5a6-4e95-9b16-21747c2ab45b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:17:23 crc kubenswrapper[4890]: I1125 16:17:23.843344 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/204f9c68-d5a6-4e95-9b16-21747c2ab45b-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 16:17:23 crc kubenswrapper[4890]: I1125 16:17:23.843387 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tfnnr\" (UniqueName: \"kubernetes.io/projected/204f9c68-d5a6-4e95-9b16-21747c2ab45b-kube-api-access-tfnnr\") on node \"crc\" DevicePath \"\"" Nov 25 16:17:23 crc kubenswrapper[4890]: I1125 16:17:23.843401 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/204f9c68-d5a6-4e95-9b16-21747c2ab45b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 16:17:24 crc kubenswrapper[4890]: I1125 16:17:24.041604 4890 generic.go:334] "Generic (PLEG): container finished" podID="204f9c68-d5a6-4e95-9b16-21747c2ab45b" containerID="9ef2f0bf5a16e7be7cc093221b39a02f81a5436a30127f021a743e8757932863" exitCode=0 Nov 25 16:17:24 crc kubenswrapper[4890]: I1125 16:17:24.041648 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t7wg4" event={"ID":"204f9c68-d5a6-4e95-9b16-21747c2ab45b","Type":"ContainerDied","Data":"9ef2f0bf5a16e7be7cc093221b39a02f81a5436a30127f021a743e8757932863"} Nov 25 16:17:24 crc kubenswrapper[4890]: I1125 16:17:24.041675 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t7wg4" event={"ID":"204f9c68-d5a6-4e95-9b16-21747c2ab45b","Type":"ContainerDied","Data":"1ecd97ad69653420e469a8d4786cc9cf0136a117ff110868727a372b5ee31565"} Nov 25 16:17:24 crc kubenswrapper[4890]: I1125 16:17:24.041727 4890 scope.go:117] "RemoveContainer" containerID="9ef2f0bf5a16e7be7cc093221b39a02f81a5436a30127f021a743e8757932863" Nov 25 16:17:24 crc kubenswrapper[4890]: I1125 16:17:24.041737 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t7wg4" Nov 25 16:17:24 crc kubenswrapper[4890]: I1125 16:17:24.099092 4890 scope.go:117] "RemoveContainer" containerID="7ec89a16e4bbe11c32036d6192d3bada4ca09b93f7b59d6888c29fbfa91fed48" Nov 25 16:17:24 crc kubenswrapper[4890]: I1125 16:17:24.104220 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-t7wg4"] Nov 25 16:17:24 crc kubenswrapper[4890]: I1125 16:17:24.113299 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-t7wg4"] Nov 25 16:17:24 crc kubenswrapper[4890]: I1125 16:17:24.172366 4890 scope.go:117] "RemoveContainer" containerID="e1bd6a1774d26552f8525f3eafe25f1496786ff5a7ae110f9d14a1fde4c1dd87" Nov 25 16:17:24 crc kubenswrapper[4890]: I1125 16:17:24.233124 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="204f9c68-d5a6-4e95-9b16-21747c2ab45b" path="/var/lib/kubelet/pods/204f9c68-d5a6-4e95-9b16-21747c2ab45b/volumes" Nov 25 16:17:24 crc kubenswrapper[4890]: I1125 16:17:24.242370 4890 scope.go:117] "RemoveContainer" containerID="9ef2f0bf5a16e7be7cc093221b39a02f81a5436a30127f021a743e8757932863" Nov 25 16:17:24 crc kubenswrapper[4890]: E1125 16:17:24.246294 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ef2f0bf5a16e7be7cc093221b39a02f81a5436a30127f021a743e8757932863\": container with ID starting with 9ef2f0bf5a16e7be7cc093221b39a02f81a5436a30127f021a743e8757932863 not found: ID does not exist" containerID="9ef2f0bf5a16e7be7cc093221b39a02f81a5436a30127f021a743e8757932863" Nov 25 16:17:24 crc kubenswrapper[4890]: I1125 16:17:24.246344 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ef2f0bf5a16e7be7cc093221b39a02f81a5436a30127f021a743e8757932863"} err="failed to get container status \"9ef2f0bf5a16e7be7cc093221b39a02f81a5436a30127f021a743e8757932863\": rpc error: code = NotFound desc = could not find container \"9ef2f0bf5a16e7be7cc093221b39a02f81a5436a30127f021a743e8757932863\": container with ID starting with 9ef2f0bf5a16e7be7cc093221b39a02f81a5436a30127f021a743e8757932863 not found: ID does not exist" Nov 25 16:17:24 crc kubenswrapper[4890]: I1125 16:17:24.246371 4890 scope.go:117] "RemoveContainer" containerID="7ec89a16e4bbe11c32036d6192d3bada4ca09b93f7b59d6888c29fbfa91fed48" Nov 25 16:17:24 crc kubenswrapper[4890]: E1125 16:17:24.249439 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ec89a16e4bbe11c32036d6192d3bada4ca09b93f7b59d6888c29fbfa91fed48\": container with ID starting with 7ec89a16e4bbe11c32036d6192d3bada4ca09b93f7b59d6888c29fbfa91fed48 not found: ID does not exist" containerID="7ec89a16e4bbe11c32036d6192d3bada4ca09b93f7b59d6888c29fbfa91fed48" Nov 25 16:17:24 crc kubenswrapper[4890]: I1125 16:17:24.249478 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ec89a16e4bbe11c32036d6192d3bada4ca09b93f7b59d6888c29fbfa91fed48"} err="failed to get container status \"7ec89a16e4bbe11c32036d6192d3bada4ca09b93f7b59d6888c29fbfa91fed48\": rpc error: code = NotFound desc = could not find container \"7ec89a16e4bbe11c32036d6192d3bada4ca09b93f7b59d6888c29fbfa91fed48\": container with ID starting with 7ec89a16e4bbe11c32036d6192d3bada4ca09b93f7b59d6888c29fbfa91fed48 not found: ID does not exist" Nov 25 16:17:24 crc kubenswrapper[4890]: I1125 16:17:24.249505 4890 scope.go:117] "RemoveContainer" containerID="e1bd6a1774d26552f8525f3eafe25f1496786ff5a7ae110f9d14a1fde4c1dd87" Nov 25 16:17:24 crc kubenswrapper[4890]: E1125 16:17:24.249961 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1bd6a1774d26552f8525f3eafe25f1496786ff5a7ae110f9d14a1fde4c1dd87\": container with ID starting with e1bd6a1774d26552f8525f3eafe25f1496786ff5a7ae110f9d14a1fde4c1dd87 not found: ID does not exist" containerID="e1bd6a1774d26552f8525f3eafe25f1496786ff5a7ae110f9d14a1fde4c1dd87" Nov 25 16:17:24 crc kubenswrapper[4890]: I1125 16:17:24.249983 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1bd6a1774d26552f8525f3eafe25f1496786ff5a7ae110f9d14a1fde4c1dd87"} err="failed to get container status \"e1bd6a1774d26552f8525f3eafe25f1496786ff5a7ae110f9d14a1fde4c1dd87\": rpc error: code = NotFound desc = could not find container \"e1bd6a1774d26552f8525f3eafe25f1496786ff5a7ae110f9d14a1fde4c1dd87\": container with ID starting with e1bd6a1774d26552f8525f3eafe25f1496786ff5a7ae110f9d14a1fde4c1dd87 not found: ID does not exist" Nov 25 16:17:26 crc kubenswrapper[4890]: I1125 16:17:26.062104 4890 generic.go:334] "Generic (PLEG): container finished" podID="80e91bdf-ca56-4a0f-8090-0262b5ddd621" containerID="8d6c037730429a01e68fa7622e1b7c5fdbc0ffaa7e00575d34401c31641afed1" exitCode=0 Nov 25 16:17:26 crc kubenswrapper[4890]: I1125 16:17:26.062188 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t78gs/crc-debug-cqdzv" event={"ID":"80e91bdf-ca56-4a0f-8090-0262b5ddd621","Type":"ContainerDied","Data":"8d6c037730429a01e68fa7622e1b7c5fdbc0ffaa7e00575d34401c31641afed1"} Nov 25 16:17:26 crc kubenswrapper[4890]: I1125 16:17:26.173192 4890 scope.go:117] "RemoveContainer" containerID="28fed48468d92f7b1fe5d54d4505526fd8660c4d76009eee3215a28f9eb145da" Nov 25 16:17:26 crc kubenswrapper[4890]: E1125 16:17:26.173458 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 16:17:27 crc kubenswrapper[4890]: I1125 16:17:27.184371 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t78gs/crc-debug-cqdzv" Nov 25 16:17:27 crc kubenswrapper[4890]: I1125 16:17:27.237582 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-t78gs/crc-debug-cqdzv"] Nov 25 16:17:27 crc kubenswrapper[4890]: I1125 16:17:27.246860 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-t78gs/crc-debug-cqdzv"] Nov 25 16:17:27 crc kubenswrapper[4890]: I1125 16:17:27.312378 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2lnt\" (UniqueName: \"kubernetes.io/projected/80e91bdf-ca56-4a0f-8090-0262b5ddd621-kube-api-access-c2lnt\") pod \"80e91bdf-ca56-4a0f-8090-0262b5ddd621\" (UID: \"80e91bdf-ca56-4a0f-8090-0262b5ddd621\") " Nov 25 16:17:27 crc kubenswrapper[4890]: I1125 16:17:27.312854 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/80e91bdf-ca56-4a0f-8090-0262b5ddd621-host\") pod \"80e91bdf-ca56-4a0f-8090-0262b5ddd621\" (UID: \"80e91bdf-ca56-4a0f-8090-0262b5ddd621\") " Nov 25 16:17:27 crc kubenswrapper[4890]: I1125 16:17:27.312957 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/80e91bdf-ca56-4a0f-8090-0262b5ddd621-host" (OuterVolumeSpecName: "host") pod "80e91bdf-ca56-4a0f-8090-0262b5ddd621" (UID: "80e91bdf-ca56-4a0f-8090-0262b5ddd621"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 16:17:27 crc kubenswrapper[4890]: I1125 16:17:27.314507 4890 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/80e91bdf-ca56-4a0f-8090-0262b5ddd621-host\") on node \"crc\" DevicePath \"\"" Nov 25 16:17:27 crc kubenswrapper[4890]: I1125 16:17:27.320915 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80e91bdf-ca56-4a0f-8090-0262b5ddd621-kube-api-access-c2lnt" (OuterVolumeSpecName: "kube-api-access-c2lnt") pod "80e91bdf-ca56-4a0f-8090-0262b5ddd621" (UID: "80e91bdf-ca56-4a0f-8090-0262b5ddd621"). InnerVolumeSpecName "kube-api-access-c2lnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:17:27 crc kubenswrapper[4890]: I1125 16:17:27.416071 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2lnt\" (UniqueName: \"kubernetes.io/projected/80e91bdf-ca56-4a0f-8090-0262b5ddd621-kube-api-access-c2lnt\") on node \"crc\" DevicePath \"\"" Nov 25 16:17:28 crc kubenswrapper[4890]: I1125 16:17:28.085192 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3710a8cead586d987c0176d36b8eb16c0fdb9625f9a54527c9caaedc099b79c5" Nov 25 16:17:28 crc kubenswrapper[4890]: I1125 16:17:28.085915 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t78gs/crc-debug-cqdzv" Nov 25 16:17:28 crc kubenswrapper[4890]: I1125 16:17:28.201209 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80e91bdf-ca56-4a0f-8090-0262b5ddd621" path="/var/lib/kubelet/pods/80e91bdf-ca56-4a0f-8090-0262b5ddd621/volumes" Nov 25 16:17:28 crc kubenswrapper[4890]: I1125 16:17:28.437871 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-t78gs/crc-debug-bjmnz"] Nov 25 16:17:28 crc kubenswrapper[4890]: E1125 16:17:28.438309 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80e91bdf-ca56-4a0f-8090-0262b5ddd621" containerName="container-00" Nov 25 16:17:28 crc kubenswrapper[4890]: I1125 16:17:28.438328 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="80e91bdf-ca56-4a0f-8090-0262b5ddd621" containerName="container-00" Nov 25 16:17:28 crc kubenswrapper[4890]: E1125 16:17:28.438360 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="204f9c68-d5a6-4e95-9b16-21747c2ab45b" containerName="extract-content" Nov 25 16:17:28 crc kubenswrapper[4890]: I1125 16:17:28.438370 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="204f9c68-d5a6-4e95-9b16-21747c2ab45b" containerName="extract-content" Nov 25 16:17:28 crc kubenswrapper[4890]: E1125 16:17:28.438381 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="204f9c68-d5a6-4e95-9b16-21747c2ab45b" containerName="registry-server" Nov 25 16:17:28 crc kubenswrapper[4890]: I1125 16:17:28.438387 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="204f9c68-d5a6-4e95-9b16-21747c2ab45b" containerName="registry-server" Nov 25 16:17:28 crc kubenswrapper[4890]: E1125 16:17:28.438409 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="204f9c68-d5a6-4e95-9b16-21747c2ab45b" containerName="extract-utilities" Nov 25 16:17:28 crc kubenswrapper[4890]: I1125 16:17:28.438415 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="204f9c68-d5a6-4e95-9b16-21747c2ab45b" containerName="extract-utilities" Nov 25 16:17:28 crc kubenswrapper[4890]: I1125 16:17:28.438601 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="204f9c68-d5a6-4e95-9b16-21747c2ab45b" containerName="registry-server" Nov 25 16:17:28 crc kubenswrapper[4890]: I1125 16:17:28.438614 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="80e91bdf-ca56-4a0f-8090-0262b5ddd621" containerName="container-00" Nov 25 16:17:28 crc kubenswrapper[4890]: I1125 16:17:28.439268 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t78gs/crc-debug-bjmnz" Nov 25 16:17:28 crc kubenswrapper[4890]: I1125 16:17:28.445040 4890 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-t78gs"/"default-dockercfg-jgmdk" Nov 25 16:17:28 crc kubenswrapper[4890]: I1125 16:17:28.537053 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k46jq\" (UniqueName: \"kubernetes.io/projected/d5c406fd-5a7e-4d3c-865a-9613d267bbb4-kube-api-access-k46jq\") pod \"crc-debug-bjmnz\" (UID: \"d5c406fd-5a7e-4d3c-865a-9613d267bbb4\") " pod="openshift-must-gather-t78gs/crc-debug-bjmnz" Nov 25 16:17:28 crc kubenswrapper[4890]: I1125 16:17:28.537311 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d5c406fd-5a7e-4d3c-865a-9613d267bbb4-host\") pod \"crc-debug-bjmnz\" (UID: \"d5c406fd-5a7e-4d3c-865a-9613d267bbb4\") " pod="openshift-must-gather-t78gs/crc-debug-bjmnz" Nov 25 16:17:28 crc kubenswrapper[4890]: I1125 16:17:28.638465 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k46jq\" (UniqueName: \"kubernetes.io/projected/d5c406fd-5a7e-4d3c-865a-9613d267bbb4-kube-api-access-k46jq\") pod \"crc-debug-bjmnz\" (UID: \"d5c406fd-5a7e-4d3c-865a-9613d267bbb4\") " pod="openshift-must-gather-t78gs/crc-debug-bjmnz" Nov 25 16:17:28 crc kubenswrapper[4890]: I1125 16:17:28.638964 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d5c406fd-5a7e-4d3c-865a-9613d267bbb4-host\") pod \"crc-debug-bjmnz\" (UID: \"d5c406fd-5a7e-4d3c-865a-9613d267bbb4\") " pod="openshift-must-gather-t78gs/crc-debug-bjmnz" Nov 25 16:17:28 crc kubenswrapper[4890]: I1125 16:17:28.639091 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d5c406fd-5a7e-4d3c-865a-9613d267bbb4-host\") pod \"crc-debug-bjmnz\" (UID: \"d5c406fd-5a7e-4d3c-865a-9613d267bbb4\") " pod="openshift-must-gather-t78gs/crc-debug-bjmnz" Nov 25 16:17:28 crc kubenswrapper[4890]: I1125 16:17:28.916130 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k46jq\" (UniqueName: \"kubernetes.io/projected/d5c406fd-5a7e-4d3c-865a-9613d267bbb4-kube-api-access-k46jq\") pod \"crc-debug-bjmnz\" (UID: \"d5c406fd-5a7e-4d3c-865a-9613d267bbb4\") " pod="openshift-must-gather-t78gs/crc-debug-bjmnz" Nov 25 16:17:29 crc kubenswrapper[4890]: I1125 16:17:29.063844 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t78gs/crc-debug-bjmnz" Nov 25 16:17:30 crc kubenswrapper[4890]: I1125 16:17:30.105121 4890 generic.go:334] "Generic (PLEG): container finished" podID="d5c406fd-5a7e-4d3c-865a-9613d267bbb4" containerID="01a40daf86169b385ee69dcdf255d666eed7a7c14b963c23599beed322d2ac31" exitCode=0 Nov 25 16:17:30 crc kubenswrapper[4890]: I1125 16:17:30.105221 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t78gs/crc-debug-bjmnz" event={"ID":"d5c406fd-5a7e-4d3c-865a-9613d267bbb4","Type":"ContainerDied","Data":"01a40daf86169b385ee69dcdf255d666eed7a7c14b963c23599beed322d2ac31"} Nov 25 16:17:30 crc kubenswrapper[4890]: I1125 16:17:30.106042 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t78gs/crc-debug-bjmnz" event={"ID":"d5c406fd-5a7e-4d3c-865a-9613d267bbb4","Type":"ContainerStarted","Data":"c8a2ef05939aabbad054dfd6f37bf9ca39d09ebc05f8e7f8d800553450ec0f80"} Nov 25 16:17:30 crc kubenswrapper[4890]: I1125 16:17:30.471799 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-t78gs/crc-debug-bjmnz"] Nov 25 16:17:30 crc kubenswrapper[4890]: I1125 16:17:30.481538 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-t78gs/crc-debug-bjmnz"] Nov 25 16:17:31 crc kubenswrapper[4890]: I1125 16:17:31.227195 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t78gs/crc-debug-bjmnz" Nov 25 16:17:31 crc kubenswrapper[4890]: I1125 16:17:31.293627 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d5c406fd-5a7e-4d3c-865a-9613d267bbb4-host\") pod \"d5c406fd-5a7e-4d3c-865a-9613d267bbb4\" (UID: \"d5c406fd-5a7e-4d3c-865a-9613d267bbb4\") " Nov 25 16:17:31 crc kubenswrapper[4890]: I1125 16:17:31.294053 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k46jq\" (UniqueName: \"kubernetes.io/projected/d5c406fd-5a7e-4d3c-865a-9613d267bbb4-kube-api-access-k46jq\") pod \"d5c406fd-5a7e-4d3c-865a-9613d267bbb4\" (UID: \"d5c406fd-5a7e-4d3c-865a-9613d267bbb4\") " Nov 25 16:17:31 crc kubenswrapper[4890]: I1125 16:17:31.296624 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d5c406fd-5a7e-4d3c-865a-9613d267bbb4-host" (OuterVolumeSpecName: "host") pod "d5c406fd-5a7e-4d3c-865a-9613d267bbb4" (UID: "d5c406fd-5a7e-4d3c-865a-9613d267bbb4"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 16:17:31 crc kubenswrapper[4890]: I1125 16:17:31.303304 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5c406fd-5a7e-4d3c-865a-9613d267bbb4-kube-api-access-k46jq" (OuterVolumeSpecName: "kube-api-access-k46jq") pod "d5c406fd-5a7e-4d3c-865a-9613d267bbb4" (UID: "d5c406fd-5a7e-4d3c-865a-9613d267bbb4"). InnerVolumeSpecName "kube-api-access-k46jq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:17:31 crc kubenswrapper[4890]: I1125 16:17:31.396195 4890 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d5c406fd-5a7e-4d3c-865a-9613d267bbb4-host\") on node \"crc\" DevicePath \"\"" Nov 25 16:17:31 crc kubenswrapper[4890]: I1125 16:17:31.396232 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k46jq\" (UniqueName: \"kubernetes.io/projected/d5c406fd-5a7e-4d3c-865a-9613d267bbb4-kube-api-access-k46jq\") on node \"crc\" DevicePath \"\"" Nov 25 16:17:31 crc kubenswrapper[4890]: I1125 16:17:31.687595 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-t78gs/crc-debug-s99s2"] Nov 25 16:17:31 crc kubenswrapper[4890]: E1125 16:17:31.688042 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5c406fd-5a7e-4d3c-865a-9613d267bbb4" containerName="container-00" Nov 25 16:17:31 crc kubenswrapper[4890]: I1125 16:17:31.688064 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5c406fd-5a7e-4d3c-865a-9613d267bbb4" containerName="container-00" Nov 25 16:17:31 crc kubenswrapper[4890]: I1125 16:17:31.688370 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5c406fd-5a7e-4d3c-865a-9613d267bbb4" containerName="container-00" Nov 25 16:17:31 crc kubenswrapper[4890]: I1125 16:17:31.689093 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t78gs/crc-debug-s99s2" Nov 25 16:17:31 crc kubenswrapper[4890]: I1125 16:17:31.804840 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xr5xn\" (UniqueName: \"kubernetes.io/projected/3f5b3c8b-7416-48c3-b0e7-d8daa3598faf-kube-api-access-xr5xn\") pod \"crc-debug-s99s2\" (UID: \"3f5b3c8b-7416-48c3-b0e7-d8daa3598faf\") " pod="openshift-must-gather-t78gs/crc-debug-s99s2" Nov 25 16:17:31 crc kubenswrapper[4890]: I1125 16:17:31.805109 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3f5b3c8b-7416-48c3-b0e7-d8daa3598faf-host\") pod \"crc-debug-s99s2\" (UID: \"3f5b3c8b-7416-48c3-b0e7-d8daa3598faf\") " pod="openshift-must-gather-t78gs/crc-debug-s99s2" Nov 25 16:17:31 crc kubenswrapper[4890]: I1125 16:17:31.907019 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3f5b3c8b-7416-48c3-b0e7-d8daa3598faf-host\") pod \"crc-debug-s99s2\" (UID: \"3f5b3c8b-7416-48c3-b0e7-d8daa3598faf\") " pod="openshift-must-gather-t78gs/crc-debug-s99s2" Nov 25 16:17:31 crc kubenswrapper[4890]: I1125 16:17:31.907180 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xr5xn\" (UniqueName: \"kubernetes.io/projected/3f5b3c8b-7416-48c3-b0e7-d8daa3598faf-kube-api-access-xr5xn\") pod \"crc-debug-s99s2\" (UID: \"3f5b3c8b-7416-48c3-b0e7-d8daa3598faf\") " pod="openshift-must-gather-t78gs/crc-debug-s99s2" Nov 25 16:17:31 crc kubenswrapper[4890]: I1125 16:17:31.907151 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3f5b3c8b-7416-48c3-b0e7-d8daa3598faf-host\") pod \"crc-debug-s99s2\" (UID: \"3f5b3c8b-7416-48c3-b0e7-d8daa3598faf\") " pod="openshift-must-gather-t78gs/crc-debug-s99s2" Nov 25 16:17:31 crc kubenswrapper[4890]: I1125 16:17:31.924185 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xr5xn\" (UniqueName: \"kubernetes.io/projected/3f5b3c8b-7416-48c3-b0e7-d8daa3598faf-kube-api-access-xr5xn\") pod \"crc-debug-s99s2\" (UID: \"3f5b3c8b-7416-48c3-b0e7-d8daa3598faf\") " pod="openshift-must-gather-t78gs/crc-debug-s99s2" Nov 25 16:17:32 crc kubenswrapper[4890]: I1125 16:17:32.006861 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t78gs/crc-debug-s99s2" Nov 25 16:17:32 crc kubenswrapper[4890]: W1125 16:17:32.037419 4890 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3f5b3c8b_7416_48c3_b0e7_d8daa3598faf.slice/crio-c8091c6e4af0b5ae36f5306fab5a450549ff3b6eb961c9341d102b17ee0b3a6a WatchSource:0}: Error finding container c8091c6e4af0b5ae36f5306fab5a450549ff3b6eb961c9341d102b17ee0b3a6a: Status 404 returned error can't find the container with id c8091c6e4af0b5ae36f5306fab5a450549ff3b6eb961c9341d102b17ee0b3a6a Nov 25 16:17:32 crc kubenswrapper[4890]: I1125 16:17:32.125642 4890 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c8a2ef05939aabbad054dfd6f37bf9ca39d09ebc05f8e7f8d800553450ec0f80" Nov 25 16:17:32 crc kubenswrapper[4890]: I1125 16:17:32.125744 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t78gs/crc-debug-bjmnz" Nov 25 16:17:32 crc kubenswrapper[4890]: I1125 16:17:32.127113 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t78gs/crc-debug-s99s2" event={"ID":"3f5b3c8b-7416-48c3-b0e7-d8daa3598faf","Type":"ContainerStarted","Data":"c8091c6e4af0b5ae36f5306fab5a450549ff3b6eb961c9341d102b17ee0b3a6a"} Nov 25 16:17:32 crc kubenswrapper[4890]: I1125 16:17:32.188393 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5c406fd-5a7e-4d3c-865a-9613d267bbb4" path="/var/lib/kubelet/pods/d5c406fd-5a7e-4d3c-865a-9613d267bbb4/volumes" Nov 25 16:17:33 crc kubenswrapper[4890]: I1125 16:17:33.136320 4890 generic.go:334] "Generic (PLEG): container finished" podID="3f5b3c8b-7416-48c3-b0e7-d8daa3598faf" containerID="f31611a1514ea1b10c6fa68f40b12cac9415243502dbb75175cc37a4cf672ecb" exitCode=0 Nov 25 16:17:33 crc kubenswrapper[4890]: I1125 16:17:33.136429 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t78gs/crc-debug-s99s2" event={"ID":"3f5b3c8b-7416-48c3-b0e7-d8daa3598faf","Type":"ContainerDied","Data":"f31611a1514ea1b10c6fa68f40b12cac9415243502dbb75175cc37a4cf672ecb"} Nov 25 16:17:33 crc kubenswrapper[4890]: I1125 16:17:33.176268 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-t78gs/crc-debug-s99s2"] Nov 25 16:17:33 crc kubenswrapper[4890]: I1125 16:17:33.185899 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-t78gs/crc-debug-s99s2"] Nov 25 16:17:34 crc kubenswrapper[4890]: I1125 16:17:34.266451 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t78gs/crc-debug-s99s2" Nov 25 16:17:34 crc kubenswrapper[4890]: I1125 16:17:34.358045 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xr5xn\" (UniqueName: \"kubernetes.io/projected/3f5b3c8b-7416-48c3-b0e7-d8daa3598faf-kube-api-access-xr5xn\") pod \"3f5b3c8b-7416-48c3-b0e7-d8daa3598faf\" (UID: \"3f5b3c8b-7416-48c3-b0e7-d8daa3598faf\") " Nov 25 16:17:34 crc kubenswrapper[4890]: I1125 16:17:34.358255 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3f5b3c8b-7416-48c3-b0e7-d8daa3598faf-host\") pod \"3f5b3c8b-7416-48c3-b0e7-d8daa3598faf\" (UID: \"3f5b3c8b-7416-48c3-b0e7-d8daa3598faf\") " Nov 25 16:17:34 crc kubenswrapper[4890]: I1125 16:17:34.358366 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3f5b3c8b-7416-48c3-b0e7-d8daa3598faf-host" (OuterVolumeSpecName: "host") pod "3f5b3c8b-7416-48c3-b0e7-d8daa3598faf" (UID: "3f5b3c8b-7416-48c3-b0e7-d8daa3598faf"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 16:17:34 crc kubenswrapper[4890]: I1125 16:17:34.358666 4890 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3f5b3c8b-7416-48c3-b0e7-d8daa3598faf-host\") on node \"crc\" DevicePath \"\"" Nov 25 16:17:34 crc kubenswrapper[4890]: I1125 16:17:34.364104 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f5b3c8b-7416-48c3-b0e7-d8daa3598faf-kube-api-access-xr5xn" (OuterVolumeSpecName: "kube-api-access-xr5xn") pod "3f5b3c8b-7416-48c3-b0e7-d8daa3598faf" (UID: "3f5b3c8b-7416-48c3-b0e7-d8daa3598faf"). InnerVolumeSpecName "kube-api-access-xr5xn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:17:34 crc kubenswrapper[4890]: I1125 16:17:34.459941 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xr5xn\" (UniqueName: \"kubernetes.io/projected/3f5b3c8b-7416-48c3-b0e7-d8daa3598faf-kube-api-access-xr5xn\") on node \"crc\" DevicePath \"\"" Nov 25 16:17:35 crc kubenswrapper[4890]: I1125 16:17:35.155577 4890 scope.go:117] "RemoveContainer" containerID="f31611a1514ea1b10c6fa68f40b12cac9415243502dbb75175cc37a4cf672ecb" Nov 25 16:17:35 crc kubenswrapper[4890]: I1125 16:17:35.155632 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t78gs/crc-debug-s99s2" Nov 25 16:17:36 crc kubenswrapper[4890]: I1125 16:17:36.190305 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f5b3c8b-7416-48c3-b0e7-d8daa3598faf" path="/var/lib/kubelet/pods/3f5b3c8b-7416-48c3-b0e7-d8daa3598faf/volumes" Nov 25 16:17:40 crc kubenswrapper[4890]: I1125 16:17:40.936053 4890 scope.go:117] "RemoveContainer" containerID="a64f294aa64d8b49d90a4e3a9d93b29d2a6455721ac18032c93addaeea49c32f" Nov 25 16:17:40 crc kubenswrapper[4890]: I1125 16:17:40.966440 4890 scope.go:117] "RemoveContainer" containerID="e7d09b0568a74792250a27d04030cc57d5e5bcd897339facdac174ee0e39df63" Nov 25 16:17:41 crc kubenswrapper[4890]: I1125 16:17:41.173222 4890 scope.go:117] "RemoveContainer" containerID="28fed48468d92f7b1fe5d54d4505526fd8660c4d76009eee3215a28f9eb145da" Nov 25 16:17:41 crc kubenswrapper[4890]: E1125 16:17:41.173747 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 16:17:54 crc kubenswrapper[4890]: I1125 16:17:54.173668 4890 scope.go:117] "RemoveContainer" containerID="28fed48468d92f7b1fe5d54d4505526fd8660c4d76009eee3215a28f9eb145da" Nov 25 16:17:54 crc kubenswrapper[4890]: E1125 16:17:54.174500 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 16:17:55 crc kubenswrapper[4890]: I1125 16:17:55.165637 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8rglr"] Nov 25 16:17:55 crc kubenswrapper[4890]: E1125 16:17:55.166333 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f5b3c8b-7416-48c3-b0e7-d8daa3598faf" containerName="container-00" Nov 25 16:17:55 crc kubenswrapper[4890]: I1125 16:17:55.166350 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f5b3c8b-7416-48c3-b0e7-d8daa3598faf" containerName="container-00" Nov 25 16:17:55 crc kubenswrapper[4890]: I1125 16:17:55.166546 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f5b3c8b-7416-48c3-b0e7-d8daa3598faf" containerName="container-00" Nov 25 16:17:55 crc kubenswrapper[4890]: I1125 16:17:55.167973 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8rglr" Nov 25 16:17:55 crc kubenswrapper[4890]: I1125 16:17:55.179652 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8rglr"] Nov 25 16:17:55 crc kubenswrapper[4890]: I1125 16:17:55.262783 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/924a90dd-2aec-4eff-beee-1018698908f6-utilities\") pod \"certified-operators-8rglr\" (UID: \"924a90dd-2aec-4eff-beee-1018698908f6\") " pod="openshift-marketplace/certified-operators-8rglr" Nov 25 16:17:55 crc kubenswrapper[4890]: I1125 16:17:55.262951 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/924a90dd-2aec-4eff-beee-1018698908f6-catalog-content\") pod \"certified-operators-8rglr\" (UID: \"924a90dd-2aec-4eff-beee-1018698908f6\") " pod="openshift-marketplace/certified-operators-8rglr" Nov 25 16:17:55 crc kubenswrapper[4890]: I1125 16:17:55.263016 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjfpm\" (UniqueName: \"kubernetes.io/projected/924a90dd-2aec-4eff-beee-1018698908f6-kube-api-access-wjfpm\") pod \"certified-operators-8rglr\" (UID: \"924a90dd-2aec-4eff-beee-1018698908f6\") " pod="openshift-marketplace/certified-operators-8rglr" Nov 25 16:17:55 crc kubenswrapper[4890]: I1125 16:17:55.364340 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/924a90dd-2aec-4eff-beee-1018698908f6-utilities\") pod \"certified-operators-8rglr\" (UID: \"924a90dd-2aec-4eff-beee-1018698908f6\") " pod="openshift-marketplace/certified-operators-8rglr" Nov 25 16:17:55 crc kubenswrapper[4890]: I1125 16:17:55.364450 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/924a90dd-2aec-4eff-beee-1018698908f6-catalog-content\") pod \"certified-operators-8rglr\" (UID: \"924a90dd-2aec-4eff-beee-1018698908f6\") " pod="openshift-marketplace/certified-operators-8rglr" Nov 25 16:17:55 crc kubenswrapper[4890]: I1125 16:17:55.364492 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjfpm\" (UniqueName: \"kubernetes.io/projected/924a90dd-2aec-4eff-beee-1018698908f6-kube-api-access-wjfpm\") pod \"certified-operators-8rglr\" (UID: \"924a90dd-2aec-4eff-beee-1018698908f6\") " pod="openshift-marketplace/certified-operators-8rglr" Nov 25 16:17:55 crc kubenswrapper[4890]: I1125 16:17:55.364969 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/924a90dd-2aec-4eff-beee-1018698908f6-utilities\") pod \"certified-operators-8rglr\" (UID: \"924a90dd-2aec-4eff-beee-1018698908f6\") " pod="openshift-marketplace/certified-operators-8rglr" Nov 25 16:17:55 crc kubenswrapper[4890]: I1125 16:17:55.365023 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/924a90dd-2aec-4eff-beee-1018698908f6-catalog-content\") pod \"certified-operators-8rglr\" (UID: \"924a90dd-2aec-4eff-beee-1018698908f6\") " pod="openshift-marketplace/certified-operators-8rglr" Nov 25 16:17:55 crc kubenswrapper[4890]: I1125 16:17:55.396944 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjfpm\" (UniqueName: \"kubernetes.io/projected/924a90dd-2aec-4eff-beee-1018698908f6-kube-api-access-wjfpm\") pod \"certified-operators-8rglr\" (UID: \"924a90dd-2aec-4eff-beee-1018698908f6\") " pod="openshift-marketplace/certified-operators-8rglr" Nov 25 16:17:55 crc kubenswrapper[4890]: I1125 16:17:55.498509 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8rglr" Nov 25 16:17:56 crc kubenswrapper[4890]: I1125 16:17:56.089525 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8rglr"] Nov 25 16:17:56 crc kubenswrapper[4890]: I1125 16:17:56.354326 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8rglr" event={"ID":"924a90dd-2aec-4eff-beee-1018698908f6","Type":"ContainerStarted","Data":"0e3236bf6a00b892ce8a84e0efe1cb4995d5bf319a35c2762a1d0c33117fbce7"} Nov 25 16:17:56 crc kubenswrapper[4890]: I1125 16:17:56.355076 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8rglr" event={"ID":"924a90dd-2aec-4eff-beee-1018698908f6","Type":"ContainerStarted","Data":"ab6b9ca93412a6d05f724236acff5134bdbb62e94972d37c6d71f03e12348e07"} Nov 25 16:17:57 crc kubenswrapper[4890]: I1125 16:17:57.366839 4890 generic.go:334] "Generic (PLEG): container finished" podID="924a90dd-2aec-4eff-beee-1018698908f6" containerID="0e3236bf6a00b892ce8a84e0efe1cb4995d5bf319a35c2762a1d0c33117fbce7" exitCode=0 Nov 25 16:17:57 crc kubenswrapper[4890]: I1125 16:17:57.366978 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8rglr" event={"ID":"924a90dd-2aec-4eff-beee-1018698908f6","Type":"ContainerDied","Data":"0e3236bf6a00b892ce8a84e0efe1cb4995d5bf319a35c2762a1d0c33117fbce7"} Nov 25 16:18:00 crc kubenswrapper[4890]: E1125 16:18:00.383987 4890 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod924a90dd_2aec_4eff_beee_1018698908f6.slice/crio-19cdfcdd86091811de8da714474faf844b4ff19d6d02e515ecb400b29dcec3de.scope\": RecentStats: unable to find data in memory cache]" Nov 25 16:18:00 crc kubenswrapper[4890]: I1125 16:18:00.395833 4890 generic.go:334] "Generic (PLEG): container finished" podID="924a90dd-2aec-4eff-beee-1018698908f6" containerID="19cdfcdd86091811de8da714474faf844b4ff19d6d02e515ecb400b29dcec3de" exitCode=0 Nov 25 16:18:00 crc kubenswrapper[4890]: I1125 16:18:00.395876 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8rglr" event={"ID":"924a90dd-2aec-4eff-beee-1018698908f6","Type":"ContainerDied","Data":"19cdfcdd86091811de8da714474faf844b4ff19d6d02e515ecb400b29dcec3de"} Nov 25 16:18:02 crc kubenswrapper[4890]: I1125 16:18:02.416904 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8rglr" event={"ID":"924a90dd-2aec-4eff-beee-1018698908f6","Type":"ContainerStarted","Data":"2512e55cd3e885cddac6a0558d9c4a70e88076a4e368ee7ecb3ff29f2e2b5351"} Nov 25 16:18:02 crc kubenswrapper[4890]: I1125 16:18:02.439295 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8rglr" podStartSLOduration=3.601828598 podStartE2EDuration="7.439266303s" podCreationTimestamp="2025-11-25 16:17:55 +0000 UTC" firstStartedPulling="2025-11-25 16:17:57.369520576 +0000 UTC m=+4535.811983176" lastFinishedPulling="2025-11-25 16:18:01.206958271 +0000 UTC m=+4539.649420881" observedRunningTime="2025-11-25 16:18:02.436886453 +0000 UTC m=+4540.879349083" watchObservedRunningTime="2025-11-25 16:18:02.439266303 +0000 UTC m=+4540.881728913" Nov 25 16:18:03 crc kubenswrapper[4890]: I1125 16:18:03.307678 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7d4dcdcc94-g9drz_117b4c5d-110a-472c-adc1-80846e0d37cb/barbican-api/0.log" Nov 25 16:18:03 crc kubenswrapper[4890]: I1125 16:18:03.409201 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7d4dcdcc94-g9drz_117b4c5d-110a-472c-adc1-80846e0d37cb/barbican-api-log/0.log" Nov 25 16:18:03 crc kubenswrapper[4890]: I1125 16:18:03.517452 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-54dc864df6-phkw8_0b033caa-b23d-43c7-b7d4-a5465d13588b/barbican-keystone-listener/0.log" Nov 25 16:18:03 crc kubenswrapper[4890]: I1125 16:18:03.581511 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-54dc864df6-phkw8_0b033caa-b23d-43c7-b7d4-a5465d13588b/barbican-keystone-listener-log/0.log" Nov 25 16:18:03 crc kubenswrapper[4890]: I1125 16:18:03.729502 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-77c9dbb88c-b285m_56f2ab61-294c-4099-99d8-e5177128adec/barbican-worker/0.log" Nov 25 16:18:03 crc kubenswrapper[4890]: I1125 16:18:03.755219 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-77c9dbb88c-b285m_56f2ab61-294c-4099-99d8-e5177128adec/barbican-worker-log/0.log" Nov 25 16:18:03 crc kubenswrapper[4890]: I1125 16:18:03.881951 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-ddcgt_a2e80fb3-23be-4b01-b663-554207d5d538/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 16:18:04 crc kubenswrapper[4890]: I1125 16:18:04.039136 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f27f06e7-28d2-403a-92c8-b1d858d90619/ceilometer-central-agent/0.log" Nov 25 16:18:04 crc kubenswrapper[4890]: I1125 16:18:04.077540 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f27f06e7-28d2-403a-92c8-b1d858d90619/proxy-httpd/0.log" Nov 25 16:18:04 crc kubenswrapper[4890]: I1125 16:18:04.137681 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f27f06e7-28d2-403a-92c8-b1d858d90619/ceilometer-notification-agent/0.log" Nov 25 16:18:04 crc kubenswrapper[4890]: I1125 16:18:04.279945 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f27f06e7-28d2-403a-92c8-b1d858d90619/sg-core/0.log" Nov 25 16:18:04 crc kubenswrapper[4890]: I1125 16:18:04.374709 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_309052f6-3a50-436c-89e0-b97d5454b935/cinder-api-log/0.log" Nov 25 16:18:04 crc kubenswrapper[4890]: I1125 16:18:04.503551 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_309052f6-3a50-436c-89e0-b97d5454b935/cinder-api/0.log" Nov 25 16:18:04 crc kubenswrapper[4890]: I1125 16:18:04.662840 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_bb86ab7c-e51d-46ad-a1d6-6a7f5ebabc41/cinder-scheduler/0.log" Nov 25 16:18:04 crc kubenswrapper[4890]: I1125 16:18:04.689956 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_bb86ab7c-e51d-46ad-a1d6-6a7f5ebabc41/probe/0.log" Nov 25 16:18:04 crc kubenswrapper[4890]: I1125 16:18:04.768008 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-9h2zh_35c7409b-9251-4ebb-9a04-8de78ea0fde0/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 16:18:04 crc kubenswrapper[4890]: I1125 16:18:04.922128 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-wdv5b_2875b0db-cd24-42e1-91b7-3308bec9d7d1/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 16:18:05 crc kubenswrapper[4890]: I1125 16:18:05.041562 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-j8b46_2d2eef50-1148-4cd4-8b76-67dc29bf8846/init/0.log" Nov 25 16:18:05 crc kubenswrapper[4890]: I1125 16:18:05.167378 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-j8b46_2d2eef50-1148-4cd4-8b76-67dc29bf8846/init/0.log" Nov 25 16:18:05 crc kubenswrapper[4890]: I1125 16:18:05.261427 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-sgklf_ef539e05-0227-4553-b6a8-0a70f5eadf8c/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 16:18:05 crc kubenswrapper[4890]: I1125 16:18:05.275251 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-j8b46_2d2eef50-1148-4cd4-8b76-67dc29bf8846/dnsmasq-dns/0.log" Nov 25 16:18:05 crc kubenswrapper[4890]: I1125 16:18:05.500511 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8rglr" Nov 25 16:18:05 crc kubenswrapper[4890]: I1125 16:18:05.501069 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8rglr" Nov 25 16:18:05 crc kubenswrapper[4890]: I1125 16:18:05.513151 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_9c617481-223e-4a84-8ecc-087c98dbd459/glance-log/0.log" Nov 25 16:18:05 crc kubenswrapper[4890]: I1125 16:18:05.539092 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_9c617481-223e-4a84-8ecc-087c98dbd459/glance-httpd/0.log" Nov 25 16:18:05 crc kubenswrapper[4890]: I1125 16:18:05.566620 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8rglr" Nov 25 16:18:05 crc kubenswrapper[4890]: I1125 16:18:05.703297 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_31454d1f-69db-4da1-a219-8c8ded147188/glance-log/0.log" Nov 25 16:18:05 crc kubenswrapper[4890]: I1125 16:18:05.770574 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_31454d1f-69db-4da1-a219-8c8ded147188/glance-httpd/0.log" Nov 25 16:18:05 crc kubenswrapper[4890]: I1125 16:18:05.925470 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7d75bbdb6d-5dvcn_2bf94c7c-4031-4868-bb7a-39c84abc96ea/horizon/0.log" Nov 25 16:18:06 crc kubenswrapper[4890]: I1125 16:18:06.060386 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-j96lg_2a24b42b-12c8-43bc-9dc7-f07991224497/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 16:18:06 crc kubenswrapper[4890]: I1125 16:18:06.238688 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-nn65s_6679d2f8-52a9-4539-bb89-f4c9c986bc6b/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 16:18:06 crc kubenswrapper[4890]: I1125 16:18:06.431433 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7d75bbdb6d-5dvcn_2bf94c7c-4031-4868-bb7a-39c84abc96ea/horizon-log/0.log" Nov 25 16:18:06 crc kubenswrapper[4890]: I1125 16:18:06.489374 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29401441-rfltq_f146ff30-8207-40d9-a579-ef8f314c41e4/keystone-cron/0.log" Nov 25 16:18:06 crc kubenswrapper[4890]: I1125 16:18:06.524133 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8rglr" Nov 25 16:18:06 crc kubenswrapper[4890]: I1125 16:18:06.579717 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-64fdb55f6-22c8w_e127320b-04a5-48f0-b434-ba6bfe1c2696/keystone-api/0.log" Nov 25 16:18:06 crc kubenswrapper[4890]: I1125 16:18:06.586147 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8rglr"] Nov 25 16:18:06 crc kubenswrapper[4890]: I1125 16:18:06.643585 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_a47203e9-a88a-42f7-86e1-164c6ad532b5/kube-state-metrics/3.log" Nov 25 16:18:06 crc kubenswrapper[4890]: I1125 16:18:06.690011 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_a47203e9-a88a-42f7-86e1-164c6ad532b5/kube-state-metrics/2.log" Nov 25 16:18:06 crc kubenswrapper[4890]: I1125 16:18:06.871836 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-4fkzh_ca9e616d-0ce2-4784-945c-9e99ba4f437b/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 16:18:07 crc kubenswrapper[4890]: I1125 16:18:07.166103 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-59d57fcc85-nkqp6_5baff989-9642-4963-816c-83904cab7811/neutron-api/0.log" Nov 25 16:18:07 crc kubenswrapper[4890]: I1125 16:18:07.173451 4890 scope.go:117] "RemoveContainer" containerID="28fed48468d92f7b1fe5d54d4505526fd8660c4d76009eee3215a28f9eb145da" Nov 25 16:18:07 crc kubenswrapper[4890]: E1125 16:18:07.173705 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 16:18:07 crc kubenswrapper[4890]: I1125 16:18:07.197867 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-59d57fcc85-nkqp6_5baff989-9642-4963-816c-83904cab7811/neutron-httpd/0.log" Nov 25 16:18:07 crc kubenswrapper[4890]: I1125 16:18:07.218427 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-dvxc6_cb4c455d-9d31-430d-b225-9bdb0b9690e0/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 16:18:07 crc kubenswrapper[4890]: I1125 16:18:07.769808 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_3c6b60ed-19f6-472a-92d6-2a1e06c92ea6/nova-api-log/0.log" Nov 25 16:18:07 crc kubenswrapper[4890]: I1125 16:18:07.869236 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_f7929bde-463b-4dff-b685-db4bf8368909/nova-cell0-conductor-conductor/0.log" Nov 25 16:18:08 crc kubenswrapper[4890]: I1125 16:18:08.204760 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_59c66e1f-c917-441e-99d4-10980ac6891e/nova-cell1-conductor-conductor/0.log" Nov 25 16:18:08 crc kubenswrapper[4890]: I1125 16:18:08.351207 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_3c6b60ed-19f6-472a-92d6-2a1e06c92ea6/nova-api-api/0.log" Nov 25 16:18:08 crc kubenswrapper[4890]: I1125 16:18:08.369597 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_85887546-bd08-4f9d-826a-ec30d6dd3dee/nova-cell1-novncproxy-novncproxy/0.log" Nov 25 16:18:08 crc kubenswrapper[4890]: I1125 16:18:08.464112 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-49nlj_e890bd11-32e9-4d73-9dd0-ae5d69a1fe58/nova-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 16:18:08 crc kubenswrapper[4890]: I1125 16:18:08.485361 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8rglr" podUID="924a90dd-2aec-4eff-beee-1018698908f6" containerName="registry-server" containerID="cri-o://2512e55cd3e885cddac6a0558d9c4a70e88076a4e368ee7ecb3ff29f2e2b5351" gracePeriod=2 Nov 25 16:18:08 crc kubenswrapper[4890]: I1125 16:18:08.727243 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_ab6971fb-880d-4e68-96f6-bb23533105cd/nova-metadata-log/0.log" Nov 25 16:18:09 crc kubenswrapper[4890]: I1125 16:18:09.016915 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8rglr" Nov 25 16:18:09 crc kubenswrapper[4890]: I1125 16:18:09.067054 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab/mysql-bootstrap/0.log" Nov 25 16:18:09 crc kubenswrapper[4890]: I1125 16:18:09.074627 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/924a90dd-2aec-4eff-beee-1018698908f6-catalog-content\") pod \"924a90dd-2aec-4eff-beee-1018698908f6\" (UID: \"924a90dd-2aec-4eff-beee-1018698908f6\") " Nov 25 16:18:09 crc kubenswrapper[4890]: I1125 16:18:09.074709 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjfpm\" (UniqueName: \"kubernetes.io/projected/924a90dd-2aec-4eff-beee-1018698908f6-kube-api-access-wjfpm\") pod \"924a90dd-2aec-4eff-beee-1018698908f6\" (UID: \"924a90dd-2aec-4eff-beee-1018698908f6\") " Nov 25 16:18:09 crc kubenswrapper[4890]: I1125 16:18:09.075573 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/924a90dd-2aec-4eff-beee-1018698908f6-utilities\") pod \"924a90dd-2aec-4eff-beee-1018698908f6\" (UID: \"924a90dd-2aec-4eff-beee-1018698908f6\") " Nov 25 16:18:09 crc kubenswrapper[4890]: I1125 16:18:09.077141 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/924a90dd-2aec-4eff-beee-1018698908f6-utilities" (OuterVolumeSpecName: "utilities") pod "924a90dd-2aec-4eff-beee-1018698908f6" (UID: "924a90dd-2aec-4eff-beee-1018698908f6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:18:09 crc kubenswrapper[4890]: I1125 16:18:09.085668 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/924a90dd-2aec-4eff-beee-1018698908f6-kube-api-access-wjfpm" (OuterVolumeSpecName: "kube-api-access-wjfpm") pod "924a90dd-2aec-4eff-beee-1018698908f6" (UID: "924a90dd-2aec-4eff-beee-1018698908f6"). InnerVolumeSpecName "kube-api-access-wjfpm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:18:09 crc kubenswrapper[4890]: I1125 16:18:09.136290 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/924a90dd-2aec-4eff-beee-1018698908f6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "924a90dd-2aec-4eff-beee-1018698908f6" (UID: "924a90dd-2aec-4eff-beee-1018698908f6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:18:09 crc kubenswrapper[4890]: I1125 16:18:09.178050 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjfpm\" (UniqueName: \"kubernetes.io/projected/924a90dd-2aec-4eff-beee-1018698908f6-kube-api-access-wjfpm\") on node \"crc\" DevicePath \"\"" Nov 25 16:18:09 crc kubenswrapper[4890]: I1125 16:18:09.178088 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/924a90dd-2aec-4eff-beee-1018698908f6-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 16:18:09 crc kubenswrapper[4890]: I1125 16:18:09.178106 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/924a90dd-2aec-4eff-beee-1018698908f6-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 16:18:09 crc kubenswrapper[4890]: I1125 16:18:09.230879 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_8df24e08-7452-4c77-9360-b555fdecf971/nova-scheduler-scheduler/0.log" Nov 25 16:18:09 crc kubenswrapper[4890]: I1125 16:18:09.251738 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab/mysql-bootstrap/0.log" Nov 25 16:18:09 crc kubenswrapper[4890]: I1125 16:18:09.323413 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_9a500d8a-0929-4a91-b7bf-3f2ccb40e3ab/galera/0.log" Nov 25 16:18:09 crc kubenswrapper[4890]: I1125 16:18:09.498227 4890 generic.go:334] "Generic (PLEG): container finished" podID="924a90dd-2aec-4eff-beee-1018698908f6" containerID="2512e55cd3e885cddac6a0558d9c4a70e88076a4e368ee7ecb3ff29f2e2b5351" exitCode=0 Nov 25 16:18:09 crc kubenswrapper[4890]: I1125 16:18:09.498294 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8rglr" event={"ID":"924a90dd-2aec-4eff-beee-1018698908f6","Type":"ContainerDied","Data":"2512e55cd3e885cddac6a0558d9c4a70e88076a4e368ee7ecb3ff29f2e2b5351"} Nov 25 16:18:09 crc kubenswrapper[4890]: I1125 16:18:09.498343 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8rglr" event={"ID":"924a90dd-2aec-4eff-beee-1018698908f6","Type":"ContainerDied","Data":"ab6b9ca93412a6d05f724236acff5134bdbb62e94972d37c6d71f03e12348e07"} Nov 25 16:18:09 crc kubenswrapper[4890]: I1125 16:18:09.498371 4890 scope.go:117] "RemoveContainer" containerID="2512e55cd3e885cddac6a0558d9c4a70e88076a4e368ee7ecb3ff29f2e2b5351" Nov 25 16:18:09 crc kubenswrapper[4890]: I1125 16:18:09.498376 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8rglr" Nov 25 16:18:09 crc kubenswrapper[4890]: I1125 16:18:09.520688 4890 scope.go:117] "RemoveContainer" containerID="19cdfcdd86091811de8da714474faf844b4ff19d6d02e515ecb400b29dcec3de" Nov 25 16:18:09 crc kubenswrapper[4890]: I1125 16:18:09.528382 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_90881a25-bc18-4b43-a226-a44e3165cb3a/mysql-bootstrap/0.log" Nov 25 16:18:09 crc kubenswrapper[4890]: I1125 16:18:09.563462 4890 scope.go:117] "RemoveContainer" containerID="0e3236bf6a00b892ce8a84e0efe1cb4995d5bf319a35c2762a1d0c33117fbce7" Nov 25 16:18:09 crc kubenswrapper[4890]: I1125 16:18:09.573572 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8rglr"] Nov 25 16:18:09 crc kubenswrapper[4890]: I1125 16:18:09.587404 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8rglr"] Nov 25 16:18:09 crc kubenswrapper[4890]: I1125 16:18:09.629252 4890 scope.go:117] "RemoveContainer" containerID="2512e55cd3e885cddac6a0558d9c4a70e88076a4e368ee7ecb3ff29f2e2b5351" Nov 25 16:18:09 crc kubenswrapper[4890]: E1125 16:18:09.631038 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2512e55cd3e885cddac6a0558d9c4a70e88076a4e368ee7ecb3ff29f2e2b5351\": container with ID starting with 2512e55cd3e885cddac6a0558d9c4a70e88076a4e368ee7ecb3ff29f2e2b5351 not found: ID does not exist" containerID="2512e55cd3e885cddac6a0558d9c4a70e88076a4e368ee7ecb3ff29f2e2b5351" Nov 25 16:18:09 crc kubenswrapper[4890]: I1125 16:18:09.631081 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2512e55cd3e885cddac6a0558d9c4a70e88076a4e368ee7ecb3ff29f2e2b5351"} err="failed to get container status \"2512e55cd3e885cddac6a0558d9c4a70e88076a4e368ee7ecb3ff29f2e2b5351\": rpc error: code = NotFound desc = could not find container \"2512e55cd3e885cddac6a0558d9c4a70e88076a4e368ee7ecb3ff29f2e2b5351\": container with ID starting with 2512e55cd3e885cddac6a0558d9c4a70e88076a4e368ee7ecb3ff29f2e2b5351 not found: ID does not exist" Nov 25 16:18:09 crc kubenswrapper[4890]: I1125 16:18:09.631108 4890 scope.go:117] "RemoveContainer" containerID="19cdfcdd86091811de8da714474faf844b4ff19d6d02e515ecb400b29dcec3de" Nov 25 16:18:09 crc kubenswrapper[4890]: E1125 16:18:09.631624 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19cdfcdd86091811de8da714474faf844b4ff19d6d02e515ecb400b29dcec3de\": container with ID starting with 19cdfcdd86091811de8da714474faf844b4ff19d6d02e515ecb400b29dcec3de not found: ID does not exist" containerID="19cdfcdd86091811de8da714474faf844b4ff19d6d02e515ecb400b29dcec3de" Nov 25 16:18:09 crc kubenswrapper[4890]: I1125 16:18:09.631655 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19cdfcdd86091811de8da714474faf844b4ff19d6d02e515ecb400b29dcec3de"} err="failed to get container status \"19cdfcdd86091811de8da714474faf844b4ff19d6d02e515ecb400b29dcec3de\": rpc error: code = NotFound desc = could not find container \"19cdfcdd86091811de8da714474faf844b4ff19d6d02e515ecb400b29dcec3de\": container with ID starting with 19cdfcdd86091811de8da714474faf844b4ff19d6d02e515ecb400b29dcec3de not found: ID does not exist" Nov 25 16:18:09 crc kubenswrapper[4890]: I1125 16:18:09.631675 4890 scope.go:117] "RemoveContainer" containerID="0e3236bf6a00b892ce8a84e0efe1cb4995d5bf319a35c2762a1d0c33117fbce7" Nov 25 16:18:09 crc kubenswrapper[4890]: E1125 16:18:09.635105 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e3236bf6a00b892ce8a84e0efe1cb4995d5bf319a35c2762a1d0c33117fbce7\": container with ID starting with 0e3236bf6a00b892ce8a84e0efe1cb4995d5bf319a35c2762a1d0c33117fbce7 not found: ID does not exist" containerID="0e3236bf6a00b892ce8a84e0efe1cb4995d5bf319a35c2762a1d0c33117fbce7" Nov 25 16:18:09 crc kubenswrapper[4890]: I1125 16:18:09.635140 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e3236bf6a00b892ce8a84e0efe1cb4995d5bf319a35c2762a1d0c33117fbce7"} err="failed to get container status \"0e3236bf6a00b892ce8a84e0efe1cb4995d5bf319a35c2762a1d0c33117fbce7\": rpc error: code = NotFound desc = could not find container \"0e3236bf6a00b892ce8a84e0efe1cb4995d5bf319a35c2762a1d0c33117fbce7\": container with ID starting with 0e3236bf6a00b892ce8a84e0efe1cb4995d5bf319a35c2762a1d0c33117fbce7 not found: ID does not exist" Nov 25 16:18:09 crc kubenswrapper[4890]: I1125 16:18:09.726245 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_90881a25-bc18-4b43-a226-a44e3165cb3a/mysql-bootstrap/0.log" Nov 25 16:18:09 crc kubenswrapper[4890]: I1125 16:18:09.808917 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_90881a25-bc18-4b43-a226-a44e3165cb3a/galera/0.log" Nov 25 16:18:09 crc kubenswrapper[4890]: I1125 16:18:09.975481 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_1d4df90d-315a-4483-a6e5-c6e5b322db54/openstackclient/0.log" Nov 25 16:18:10 crc kubenswrapper[4890]: I1125 16:18:10.024009 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-v59z7_f282d61d-d8c2-4933-99c1-d95a9d4f40e2/openstack-network-exporter/0.log" Nov 25 16:18:10 crc kubenswrapper[4890]: I1125 16:18:10.184645 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="924a90dd-2aec-4eff-beee-1018698908f6" path="/var/lib/kubelet/pods/924a90dd-2aec-4eff-beee-1018698908f6/volumes" Nov 25 16:18:10 crc kubenswrapper[4890]: I1125 16:18:10.206857 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-nrbtc_524ff107-da15-4e2f-a0c8-acca0e91b27a/ovn-controller/0.log" Nov 25 16:18:10 crc kubenswrapper[4890]: I1125 16:18:10.420359 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-5tg6b_289df7bb-0b8e-4411-ab2f-935629ed4019/ovsdb-server-init/0.log" Nov 25 16:18:10 crc kubenswrapper[4890]: I1125 16:18:10.558783 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-5tg6b_289df7bb-0b8e-4411-ab2f-935629ed4019/ovs-vswitchd/0.log" Nov 25 16:18:10 crc kubenswrapper[4890]: I1125 16:18:10.597379 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-5tg6b_289df7bb-0b8e-4411-ab2f-935629ed4019/ovsdb-server-init/0.log" Nov 25 16:18:10 crc kubenswrapper[4890]: I1125 16:18:10.619459 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-5tg6b_289df7bb-0b8e-4411-ab2f-935629ed4019/ovsdb-server/0.log" Nov 25 16:18:10 crc kubenswrapper[4890]: I1125 16:18:10.658400 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_ab6971fb-880d-4e68-96f6-bb23533105cd/nova-metadata-metadata/0.log" Nov 25 16:18:10 crc kubenswrapper[4890]: I1125 16:18:10.819927 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-llz86_995132c2-f199-4d15-b3ef-3c71fed454f5/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 16:18:11 crc kubenswrapper[4890]: I1125 16:18:11.028338 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_effbba39-45db-4d3b-a883-03f38d0b49f8/openstack-network-exporter/0.log" Nov 25 16:18:11 crc kubenswrapper[4890]: I1125 16:18:11.553650 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_effbba39-45db-4d3b-a883-03f38d0b49f8/ovn-northd/0.log" Nov 25 16:18:11 crc kubenswrapper[4890]: I1125 16:18:11.559823 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_49e51424-feba-4038-88a6-1cd56637984c/ovsdbserver-nb/0.log" Nov 25 16:18:11 crc kubenswrapper[4890]: I1125 16:18:11.598725 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_49e51424-feba-4038-88a6-1cd56637984c/openstack-network-exporter/0.log" Nov 25 16:18:11 crc kubenswrapper[4890]: I1125 16:18:11.788538 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_c9667a27-2ed9-4b9a-a75a-4901a93cab5e/openstack-network-exporter/0.log" Nov 25 16:18:11 crc kubenswrapper[4890]: I1125 16:18:11.831480 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_c9667a27-2ed9-4b9a-a75a-4901a93cab5e/ovsdbserver-sb/0.log" Nov 25 16:18:12 crc kubenswrapper[4890]: I1125 16:18:12.108137 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-74c777b4cb-m7k57_46d00836-60a2-459f-b4c9-ed2d5e3b1fb1/placement-api/0.log" Nov 25 16:18:12 crc kubenswrapper[4890]: I1125 16:18:12.126278 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_d08ea973-c3f2-4192-8bfc-b025f0c5391b/setup-container/0.log" Nov 25 16:18:12 crc kubenswrapper[4890]: I1125 16:18:12.166832 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-74c777b4cb-m7k57_46d00836-60a2-459f-b4c9-ed2d5e3b1fb1/placement-log/0.log" Nov 25 16:18:12 crc kubenswrapper[4890]: I1125 16:18:12.281724 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_d08ea973-c3f2-4192-8bfc-b025f0c5391b/setup-container/0.log" Nov 25 16:18:12 crc kubenswrapper[4890]: I1125 16:18:12.334673 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_d08ea973-c3f2-4192-8bfc-b025f0c5391b/rabbitmq/0.log" Nov 25 16:18:12 crc kubenswrapper[4890]: I1125 16:18:12.454929 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_9bab37b3-681b-4cd5-9fee-503afa0d6d22/setup-container/0.log" Nov 25 16:18:12 crc kubenswrapper[4890]: I1125 16:18:12.582629 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_9bab37b3-681b-4cd5-9fee-503afa0d6d22/setup-container/0.log" Nov 25 16:18:12 crc kubenswrapper[4890]: I1125 16:18:12.721911 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-7bjzl_2ba10718-5ae6-4670-bbc6-b02e9c5562e0/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 16:18:12 crc kubenswrapper[4890]: I1125 16:18:12.726217 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_9bab37b3-681b-4cd5-9fee-503afa0d6d22/rabbitmq/0.log" Nov 25 16:18:12 crc kubenswrapper[4890]: I1125 16:18:12.947351 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-w4tcq_4f0ccf80-fad3-4920-83e3-1c68111ed1ac/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 16:18:13 crc kubenswrapper[4890]: I1125 16:18:13.095624 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-cr58b_6d15a3f2-bbd4-43ae-b620-93251f74369e/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 16:18:13 crc kubenswrapper[4890]: I1125 16:18:13.703550 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-l94p8_90b03f38-bd28-4c62-ae05-5cc72f4e6ee7/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 16:18:13 crc kubenswrapper[4890]: I1125 16:18:13.761693 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-59sbs_788b5bd1-adf2-4416-a8f0-975895de101c/ssh-known-hosts-edpm-deployment/0.log" Nov 25 16:18:14 crc kubenswrapper[4890]: I1125 16:18:14.065732 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5f45cd8b8f-8wpds_66cfb07e-ac1a-4aee-8fbd-7871e2e30d88/proxy-server/0.log" Nov 25 16:18:14 crc kubenswrapper[4890]: I1125 16:18:14.080414 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5f45cd8b8f-8wpds_66cfb07e-ac1a-4aee-8fbd-7871e2e30d88/proxy-httpd/0.log" Nov 25 16:18:14 crc kubenswrapper[4890]: I1125 16:18:14.110528 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-vs9dw_cf34ad11-7dd6-4572-b3d2-4c14127a1d76/swift-ring-rebalance/0.log" Nov 25 16:18:14 crc kubenswrapper[4890]: I1125 16:18:14.313251 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6517688d-0723-4809-ba51-bcaf48524157/account-reaper/0.log" Nov 25 16:18:14 crc kubenswrapper[4890]: I1125 16:18:14.332847 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6517688d-0723-4809-ba51-bcaf48524157/account-replicator/0.log" Nov 25 16:18:14 crc kubenswrapper[4890]: I1125 16:18:14.357909 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6517688d-0723-4809-ba51-bcaf48524157/account-auditor/0.log" Nov 25 16:18:14 crc kubenswrapper[4890]: I1125 16:18:14.561106 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6517688d-0723-4809-ba51-bcaf48524157/container-auditor/0.log" Nov 25 16:18:14 crc kubenswrapper[4890]: I1125 16:18:14.561262 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6517688d-0723-4809-ba51-bcaf48524157/account-server/0.log" Nov 25 16:18:14 crc kubenswrapper[4890]: I1125 16:18:14.561451 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6517688d-0723-4809-ba51-bcaf48524157/container-replicator/0.log" Nov 25 16:18:14 crc kubenswrapper[4890]: I1125 16:18:14.598501 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6517688d-0723-4809-ba51-bcaf48524157/container-server/0.log" Nov 25 16:18:14 crc kubenswrapper[4890]: I1125 16:18:14.758452 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6517688d-0723-4809-ba51-bcaf48524157/object-expirer/0.log" Nov 25 16:18:14 crc kubenswrapper[4890]: I1125 16:18:14.801223 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6517688d-0723-4809-ba51-bcaf48524157/container-updater/0.log" Nov 25 16:18:14 crc kubenswrapper[4890]: I1125 16:18:14.822421 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6517688d-0723-4809-ba51-bcaf48524157/object-auditor/0.log" Nov 25 16:18:14 crc kubenswrapper[4890]: I1125 16:18:14.857988 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6517688d-0723-4809-ba51-bcaf48524157/object-replicator/0.log" Nov 25 16:18:14 crc kubenswrapper[4890]: I1125 16:18:14.973275 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6517688d-0723-4809-ba51-bcaf48524157/object-server/0.log" Nov 25 16:18:15 crc kubenswrapper[4890]: I1125 16:18:15.029884 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6517688d-0723-4809-ba51-bcaf48524157/rsync/0.log" Nov 25 16:18:15 crc kubenswrapper[4890]: I1125 16:18:15.051408 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6517688d-0723-4809-ba51-bcaf48524157/object-updater/0.log" Nov 25 16:18:15 crc kubenswrapper[4890]: I1125 16:18:15.098944 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6517688d-0723-4809-ba51-bcaf48524157/swift-recon-cron/0.log" Nov 25 16:18:15 crc kubenswrapper[4890]: I1125 16:18:15.366365 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-jm99p_588df9b2-9cfb-449a-9dca-17912f88a4c6/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 16:18:15 crc kubenswrapper[4890]: I1125 16:18:15.388170 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_8b652448-2962-42f9-b0b5-e551b50edc1e/tempest-tests-tempest-tests-runner/0.log" Nov 25 16:18:15 crc kubenswrapper[4890]: I1125 16:18:15.631953 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_a76905fc-a86c-4ed6-982d-0a50bada561f/test-operator-logs-container/0.log" Nov 25 16:18:15 crc kubenswrapper[4890]: I1125 16:18:15.686399 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-6vbxl_8394f488-cfd0-4fdc-aa22-96fe01aec73e/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 16:18:18 crc kubenswrapper[4890]: I1125 16:18:18.173612 4890 scope.go:117] "RemoveContainer" containerID="28fed48468d92f7b1fe5d54d4505526fd8660c4d76009eee3215a28f9eb145da" Nov 25 16:18:18 crc kubenswrapper[4890]: E1125 16:18:18.174666 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 16:18:31 crc kubenswrapper[4890]: I1125 16:18:31.235023 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_c5630332-c335-4b9a-8486-e845c6baaa9c/memcached/0.log" Nov 25 16:18:33 crc kubenswrapper[4890]: I1125 16:18:33.173228 4890 scope.go:117] "RemoveContainer" containerID="28fed48468d92f7b1fe5d54d4505526fd8660c4d76009eee3215a28f9eb145da" Nov 25 16:18:33 crc kubenswrapper[4890]: E1125 16:18:33.173815 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 16:18:43 crc kubenswrapper[4890]: I1125 16:18:43.443736 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9b547be2eac4c3e8f15e2800f340fb44715065f09b75b108844ab9dee67zk4x_4050c336-b097-49f8-bc65-17bb8d191fa4/util/0.log" Nov 25 16:18:43 crc kubenswrapper[4890]: I1125 16:18:43.691097 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9b547be2eac4c3e8f15e2800f340fb44715065f09b75b108844ab9dee67zk4x_4050c336-b097-49f8-bc65-17bb8d191fa4/pull/0.log" Nov 25 16:18:43 crc kubenswrapper[4890]: I1125 16:18:43.694370 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9b547be2eac4c3e8f15e2800f340fb44715065f09b75b108844ab9dee67zk4x_4050c336-b097-49f8-bc65-17bb8d191fa4/util/0.log" Nov 25 16:18:43 crc kubenswrapper[4890]: I1125 16:18:43.719117 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9b547be2eac4c3e8f15e2800f340fb44715065f09b75b108844ab9dee67zk4x_4050c336-b097-49f8-bc65-17bb8d191fa4/pull/0.log" Nov 25 16:18:43 crc kubenswrapper[4890]: I1125 16:18:43.891625 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9b547be2eac4c3e8f15e2800f340fb44715065f09b75b108844ab9dee67zk4x_4050c336-b097-49f8-bc65-17bb8d191fa4/util/0.log" Nov 25 16:18:43 crc kubenswrapper[4890]: I1125 16:18:43.909280 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9b547be2eac4c3e8f15e2800f340fb44715065f09b75b108844ab9dee67zk4x_4050c336-b097-49f8-bc65-17bb8d191fa4/pull/0.log" Nov 25 16:18:43 crc kubenswrapper[4890]: I1125 16:18:43.925039 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9b547be2eac4c3e8f15e2800f340fb44715065f09b75b108844ab9dee67zk4x_4050c336-b097-49f8-bc65-17bb8d191fa4/extract/0.log" Nov 25 16:18:44 crc kubenswrapper[4890]: I1125 16:18:44.077731 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-86dc4d89c8-pmmf2_d175a35b-8b9b-49c0-8fc2-02ef62f131d9/kube-rbac-proxy/0.log" Nov 25 16:18:44 crc kubenswrapper[4890]: I1125 16:18:44.145191 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-86dc4d89c8-pmmf2_d175a35b-8b9b-49c0-8fc2-02ef62f131d9/manager/3.log" Nov 25 16:18:44 crc kubenswrapper[4890]: I1125 16:18:44.247105 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-86dc4d89c8-pmmf2_d175a35b-8b9b-49c0-8fc2-02ef62f131d9/manager/2.log" Nov 25 16:18:44 crc kubenswrapper[4890]: I1125 16:18:44.300898 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-79856dc55c-jcw7x_0bdbada3-e08c-4f8a-bd55-bac955d4370d/kube-rbac-proxy/0.log" Nov 25 16:18:44 crc kubenswrapper[4890]: I1125 16:18:44.408196 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-79856dc55c-jcw7x_0bdbada3-e08c-4f8a-bd55-bac955d4370d/manager/3.log" Nov 25 16:18:44 crc kubenswrapper[4890]: I1125 16:18:44.432126 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-79856dc55c-jcw7x_0bdbada3-e08c-4f8a-bd55-bac955d4370d/manager/2.log" Nov 25 16:18:44 crc kubenswrapper[4890]: I1125 16:18:44.545869 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-7d695c9b56-stwjm_77a5e8f9-92ab-48b5-be19-29a7a8e5df49/kube-rbac-proxy/0.log" Nov 25 16:18:45 crc kubenswrapper[4890]: I1125 16:18:45.174669 4890 scope.go:117] "RemoveContainer" containerID="28fed48468d92f7b1fe5d54d4505526fd8660c4d76009eee3215a28f9eb145da" Nov 25 16:18:45 crc kubenswrapper[4890]: E1125 16:18:45.175891 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 16:18:45 crc kubenswrapper[4890]: I1125 16:18:45.323785 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-7d695c9b56-stwjm_77a5e8f9-92ab-48b5-be19-29a7a8e5df49/manager/3.log" Nov 25 16:18:45 crc kubenswrapper[4890]: I1125 16:18:45.346427 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-7d695c9b56-stwjm_77a5e8f9-92ab-48b5-be19-29a7a8e5df49/manager/2.log" Nov 25 16:18:45 crc kubenswrapper[4890]: I1125 16:18:45.347330 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-68b95954c9-lfnwl_076c2a0e-7fdb-42e1-ab73-e296bba2b7a9/kube-rbac-proxy/0.log" Nov 25 16:18:45 crc kubenswrapper[4890]: I1125 16:18:45.603287 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-68b95954c9-lfnwl_076c2a0e-7fdb-42e1-ab73-e296bba2b7a9/manager/2.log" Nov 25 16:18:45 crc kubenswrapper[4890]: I1125 16:18:45.608054 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-68b95954c9-lfnwl_076c2a0e-7fdb-42e1-ab73-e296bba2b7a9/manager/3.log" Nov 25 16:18:45 crc kubenswrapper[4890]: I1125 16:18:45.664647 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-774b86978c-xvfgd_72c26c5b-3370-4994-a5f0-82128c62c2df/kube-rbac-proxy/0.log" Nov 25 16:18:45 crc kubenswrapper[4890]: I1125 16:18:45.816265 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-774b86978c-xvfgd_72c26c5b-3370-4994-a5f0-82128c62c2df/manager/3.log" Nov 25 16:18:45 crc kubenswrapper[4890]: I1125 16:18:45.882756 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-774b86978c-xvfgd_72c26c5b-3370-4994-a5f0-82128c62c2df/manager/2.log" Nov 25 16:18:45 crc kubenswrapper[4890]: I1125 16:18:45.888839 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c9694994-j6pqs_c528348f-0907-423a-a0c2-f15cf18ff99a/kube-rbac-proxy/0.log" Nov 25 16:18:46 crc kubenswrapper[4890]: I1125 16:18:46.028327 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c9694994-j6pqs_c528348f-0907-423a-a0c2-f15cf18ff99a/manager/3.log" Nov 25 16:18:46 crc kubenswrapper[4890]: I1125 16:18:46.090440 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-d5cc86f4b-stflb_7e576602-19f5-4656-84dd-5ab80abca895/kube-rbac-proxy/0.log" Nov 25 16:18:46 crc kubenswrapper[4890]: I1125 16:18:46.105310 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c9694994-j6pqs_c528348f-0907-423a-a0c2-f15cf18ff99a/manager/2.log" Nov 25 16:18:46 crc kubenswrapper[4890]: I1125 16:18:46.288769 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-d5cc86f4b-stflb_7e576602-19f5-4656-84dd-5ab80abca895/manager/3.log" Nov 25 16:18:46 crc kubenswrapper[4890]: I1125 16:18:46.324267 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-d5cc86f4b-stflb_7e576602-19f5-4656-84dd-5ab80abca895/manager/2.log" Nov 25 16:18:46 crc kubenswrapper[4890]: I1125 16:18:46.355731 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5bfcdc958c-hxzf4_679b3bf3-1ebc-4d19-9ad2-588bd41565dd/kube-rbac-proxy/0.log" Nov 25 16:18:46 crc kubenswrapper[4890]: I1125 16:18:46.470625 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5bfcdc958c-hxzf4_679b3bf3-1ebc-4d19-9ad2-588bd41565dd/manager/3.log" Nov 25 16:18:46 crc kubenswrapper[4890]: I1125 16:18:46.514151 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5bfcdc958c-hxzf4_679b3bf3-1ebc-4d19-9ad2-588bd41565dd/manager/2.log" Nov 25 16:18:46 crc kubenswrapper[4890]: I1125 16:18:46.563863 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-748dc6576f-rs6hx_a5d8a8ba-50c0-4fb6-823f-f28cf0d0f125/kube-rbac-proxy/0.log" Nov 25 16:18:46 crc kubenswrapper[4890]: I1125 16:18:46.688336 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-748dc6576f-rs6hx_a5d8a8ba-50c0-4fb6-823f-f28cf0d0f125/manager/3.log" Nov 25 16:18:46 crc kubenswrapper[4890]: I1125 16:18:46.712327 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-748dc6576f-rs6hx_a5d8a8ba-50c0-4fb6-823f-f28cf0d0f125/manager/2.log" Nov 25 16:18:46 crc kubenswrapper[4890]: I1125 16:18:46.833212 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58bb8d67cc-slpt4_f272410f-230d-43f2-946e-e34e4d9c05ea/kube-rbac-proxy/0.log" Nov 25 16:18:46 crc kubenswrapper[4890]: I1125 16:18:46.883808 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58bb8d67cc-slpt4_f272410f-230d-43f2-946e-e34e4d9c05ea/manager/3.log" Nov 25 16:18:47 crc kubenswrapper[4890]: I1125 16:18:47.287411 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58bb8d67cc-slpt4_f272410f-230d-43f2-946e-e34e4d9c05ea/manager/2.log" Nov 25 16:18:47 crc kubenswrapper[4890]: I1125 16:18:47.337069 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-cb6c4fdb7-wvjw7_03353c8b-91f4-41ae-afc1-c522b1c73bb4/kube-rbac-proxy/0.log" Nov 25 16:18:47 crc kubenswrapper[4890]: I1125 16:18:47.359984 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-cb6c4fdb7-wvjw7_03353c8b-91f4-41ae-afc1-c522b1c73bb4/manager/3.log" Nov 25 16:18:47 crc kubenswrapper[4890]: I1125 16:18:47.422316 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-cb6c4fdb7-wvjw7_03353c8b-91f4-41ae-afc1-c522b1c73bb4/manager/2.log" Nov 25 16:18:47 crc kubenswrapper[4890]: I1125 16:18:47.535999 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-7c57c8bbc4-829nk_4364eec5-444a-41f6-af1c-fe16b24bd044/kube-rbac-proxy/0.log" Nov 25 16:18:47 crc kubenswrapper[4890]: I1125 16:18:47.570203 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-7c57c8bbc4-829nk_4364eec5-444a-41f6-af1c-fe16b24bd044/manager/2.log" Nov 25 16:18:47 crc kubenswrapper[4890]: I1125 16:18:47.590300 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-7c57c8bbc4-829nk_4364eec5-444a-41f6-af1c-fe16b24bd044/manager/3.log" Nov 25 16:18:47 crc kubenswrapper[4890]: I1125 16:18:47.730070 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-6jq7q_0769cde4-b49f-41c5-b720-e0dbc667eb4a/kube-rbac-proxy/0.log" Nov 25 16:18:47 crc kubenswrapper[4890]: I1125 16:18:47.750806 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-6jq7q_0769cde4-b49f-41c5-b720-e0dbc667eb4a/manager/2.log" Nov 25 16:18:47 crc kubenswrapper[4890]: I1125 16:18:47.766569 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-6jq7q_0769cde4-b49f-41c5-b720-e0dbc667eb4a/manager/3.log" Nov 25 16:18:47 crc kubenswrapper[4890]: I1125 16:18:47.884819 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-fd75fd47d-56bcr_a9efeabb-bf1c-4de2-8322-961b284e95cc/kube-rbac-proxy/0.log" Nov 25 16:18:47 crc kubenswrapper[4890]: I1125 16:18:47.942088 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-fd75fd47d-56bcr_a9efeabb-bf1c-4de2-8322-961b284e95cc/manager/3.log" Nov 25 16:18:47 crc kubenswrapper[4890]: I1125 16:18:47.984308 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-fd75fd47d-56bcr_a9efeabb-bf1c-4de2-8322-961b284e95cc/manager/2.log" Nov 25 16:18:48 crc kubenswrapper[4890]: I1125 16:18:48.057197 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-544b9bb9-5lr7r_10534635-efe1-4e20-8c9b-9cbea73fbe87/kube-rbac-proxy/0.log" Nov 25 16:18:48 crc kubenswrapper[4890]: I1125 16:18:48.126431 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-544b9bb9-5lr7r_10534635-efe1-4e20-8c9b-9cbea73fbe87/manager/1.log" Nov 25 16:18:48 crc kubenswrapper[4890]: I1125 16:18:48.142179 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-544b9bb9-5lr7r_10534635-efe1-4e20-8c9b-9cbea73fbe87/manager/0.log" Nov 25 16:18:48 crc kubenswrapper[4890]: I1125 16:18:48.273386 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-68587559f4-9b4cf_3e496c3d-9485-44bc-b3b1-dc1682a88434/manager/2.log" Nov 25 16:18:48 crc kubenswrapper[4890]: I1125 16:18:48.453999 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6d88ccc4fc-9smnn_bd60efda-e354-44d0-af56-c110651bdbaf/operator/1.log" Nov 25 16:18:48 crc kubenswrapper[4890]: I1125 16:18:48.479960 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-68587559f4-9b4cf_3e496c3d-9485-44bc-b3b1-dc1682a88434/manager/3.log" Nov 25 16:18:48 crc kubenswrapper[4890]: I1125 16:18:48.670844 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6d88ccc4fc-9smnn_bd60efda-e354-44d0-af56-c110651bdbaf/operator/0.log" Nov 25 16:18:48 crc kubenswrapper[4890]: I1125 16:18:48.686499 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-vtg6p_bd47d0df-89c3-465a-8f11-8f0e7e1117eb/registry-server/0.log" Nov 25 16:18:48 crc kubenswrapper[4890]: I1125 16:18:48.714006 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-66cf5c67ff-vx75t_29f250db-2a96-41ad-bb8f-5b503a9288d8/kube-rbac-proxy/0.log" Nov 25 16:18:48 crc kubenswrapper[4890]: I1125 16:18:48.974191 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-66cf5c67ff-vx75t_29f250db-2a96-41ad-bb8f-5b503a9288d8/manager/3.log" Nov 25 16:18:49 crc kubenswrapper[4890]: I1125 16:18:49.022763 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-66cf5c67ff-vx75t_29f250db-2a96-41ad-bb8f-5b503a9288d8/manager/2.log" Nov 25 16:18:49 crc kubenswrapper[4890]: I1125 16:18:49.061716 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5db546f9d9-nshw7_9f0e3cf9-dcc0-405f-a2de-9148844ec3ae/kube-rbac-proxy/0.log" Nov 25 16:18:49 crc kubenswrapper[4890]: I1125 16:18:49.085310 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5db546f9d9-nshw7_9f0e3cf9-dcc0-405f-a2de-9148844ec3ae/manager/3.log" Nov 25 16:18:49 crc kubenswrapper[4890]: I1125 16:18:49.201340 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5db546f9d9-nshw7_9f0e3cf9-dcc0-405f-a2de-9148844ec3ae/manager/2.log" Nov 25 16:18:49 crc kubenswrapper[4890]: I1125 16:18:49.252607 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-dmlx4_e26594ec-4bc2-4051-8a49-fc490bd3b6c9/operator/3.log" Nov 25 16:18:49 crc kubenswrapper[4890]: I1125 16:18:49.301102 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-dmlx4_e26594ec-4bc2-4051-8a49-fc490bd3b6c9/operator/2.log" Nov 25 16:18:49 crc kubenswrapper[4890]: I1125 16:18:49.405076 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6fdc4fcf86-dzkgt_861317e1-222d-4f14-b931-e9e7d100ebd6/kube-rbac-proxy/0.log" Nov 25 16:18:49 crc kubenswrapper[4890]: I1125 16:18:49.430152 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6fdc4fcf86-dzkgt_861317e1-222d-4f14-b931-e9e7d100ebd6/manager/3.log" Nov 25 16:18:49 crc kubenswrapper[4890]: I1125 16:18:49.475895 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6fdc4fcf86-dzkgt_861317e1-222d-4f14-b931-e9e7d100ebd6/manager/2.log" Nov 25 16:18:49 crc kubenswrapper[4890]: I1125 16:18:49.554527 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-567f98c9d-6zr4x_d50aec9d-1673-4b81-bb57-04cae418843e/kube-rbac-proxy/0.log" Nov 25 16:18:49 crc kubenswrapper[4890]: I1125 16:18:49.612110 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-567f98c9d-6zr4x_d50aec9d-1673-4b81-bb57-04cae418843e/manager/3.log" Nov 25 16:18:49 crc kubenswrapper[4890]: I1125 16:18:49.627359 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-567f98c9d-6zr4x_d50aec9d-1673-4b81-bb57-04cae418843e/manager/2.log" Nov 25 16:18:49 crc kubenswrapper[4890]: I1125 16:18:49.692489 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cb74df96-6mqt2_9637fded-765d-4037-be8f-82e59ffdfc14/kube-rbac-proxy/0.log" Nov 25 16:18:49 crc kubenswrapper[4890]: I1125 16:18:49.781037 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cb74df96-6mqt2_9637fded-765d-4037-be8f-82e59ffdfc14/manager/1.log" Nov 25 16:18:49 crc kubenswrapper[4890]: I1125 16:18:49.844542 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cb74df96-6mqt2_9637fded-765d-4037-be8f-82e59ffdfc14/manager/0.log" Nov 25 16:18:49 crc kubenswrapper[4890]: I1125 16:18:49.900952 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-864885998-62ddz_a9a69828-3f0d-4d55-a840-10c2c791444c/kube-rbac-proxy/0.log" Nov 25 16:18:49 crc kubenswrapper[4890]: I1125 16:18:49.926311 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-864885998-62ddz_a9a69828-3f0d-4d55-a840-10c2c791444c/manager/2.log" Nov 25 16:18:50 crc kubenswrapper[4890]: I1125 16:18:50.006084 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-864885998-62ddz_a9a69828-3f0d-4d55-a840-10c2c791444c/manager/1.log" Nov 25 16:18:58 crc kubenswrapper[4890]: I1125 16:18:58.174194 4890 scope.go:117] "RemoveContainer" containerID="28fed48468d92f7b1fe5d54d4505526fd8660c4d76009eee3215a28f9eb145da" Nov 25 16:18:58 crc kubenswrapper[4890]: E1125 16:18:58.175568 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 16:19:08 crc kubenswrapper[4890]: I1125 16:19:08.780085 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-zs6w7_a3f62807-e422-4d74-82c5-81756d963b77/control-plane-machine-set-operator/0.log" Nov 25 16:19:09 crc kubenswrapper[4890]: I1125 16:19:09.174278 4890 scope.go:117] "RemoveContainer" containerID="28fed48468d92f7b1fe5d54d4505526fd8660c4d76009eee3215a28f9eb145da" Nov 25 16:19:09 crc kubenswrapper[4890]: E1125 16:19:09.174638 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 16:19:09 crc kubenswrapper[4890]: I1125 16:19:09.268111 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-6dmh8_8b5c8177-1153-453a-9110-6023d01e3fb7/kube-rbac-proxy/0.log" Nov 25 16:19:09 crc kubenswrapper[4890]: I1125 16:19:09.305402 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-6dmh8_8b5c8177-1153-453a-9110-6023d01e3fb7/machine-api-operator/0.log" Nov 25 16:19:20 crc kubenswrapper[4890]: I1125 16:19:20.808815 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-9c9wk_e14d1071-6d9a-4543-b593-8e999b7dcad2/cert-manager-controller/1.log" Nov 25 16:19:20 crc kubenswrapper[4890]: I1125 16:19:20.832857 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-9c9wk_e14d1071-6d9a-4543-b593-8e999b7dcad2/cert-manager-controller/0.log" Nov 25 16:19:21 crc kubenswrapper[4890]: I1125 16:19:21.022738 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-x7dwr_d0f0970d-430a-4ea0-b18c-6af23eb40c71/cert-manager-cainjector/1.log" Nov 25 16:19:21 crc kubenswrapper[4890]: I1125 16:19:21.034872 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-x7dwr_d0f0970d-430a-4ea0-b18c-6af23eb40c71/cert-manager-cainjector/0.log" Nov 25 16:19:21 crc kubenswrapper[4890]: I1125 16:19:21.186821 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-r5kd7_3e0ccee5-4080-4973-99df-e4d6e929e471/cert-manager-webhook/0.log" Nov 25 16:19:22 crc kubenswrapper[4890]: I1125 16:19:22.173017 4890 scope.go:117] "RemoveContainer" containerID="28fed48468d92f7b1fe5d54d4505526fd8660c4d76009eee3215a28f9eb145da" Nov 25 16:19:22 crc kubenswrapper[4890]: E1125 16:19:22.173683 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 16:19:33 crc kubenswrapper[4890]: I1125 16:19:33.603919 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5874bd7bc5-mrnt9_85206fda-8548-4b94-82ce-df958e8d4045/nmstate-console-plugin/0.log" Nov 25 16:19:33 crc kubenswrapper[4890]: I1125 16:19:33.780551 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-sk8fz_1e35f1ea-5729-463f-b875-6b897b4d5944/nmstate-handler/0.log" Nov 25 16:19:33 crc kubenswrapper[4890]: I1125 16:19:33.857122 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-8mxxw_9d98268c-c10a-4157-9737-e2df6cbdcfe5/kube-rbac-proxy/0.log" Nov 25 16:19:33 crc kubenswrapper[4890]: I1125 16:19:33.859315 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-8mxxw_9d98268c-c10a-4157-9737-e2df6cbdcfe5/nmstate-metrics/0.log" Nov 25 16:19:34 crc kubenswrapper[4890]: I1125 16:19:34.066423 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-557fdffb88-v4blp_213afbd7-4c85-41e6-94c9-dac6db9da0d0/nmstate-operator/0.log" Nov 25 16:19:34 crc kubenswrapper[4890]: I1125 16:19:34.097898 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6b89b748d8-nhh4b_f6edfb37-872a-4515-ada7-07d552977de0/nmstate-webhook/0.log" Nov 25 16:19:36 crc kubenswrapper[4890]: I1125 16:19:36.174756 4890 scope.go:117] "RemoveContainer" containerID="28fed48468d92f7b1fe5d54d4505526fd8660c4d76009eee3215a28f9eb145da" Nov 25 16:19:36 crc kubenswrapper[4890]: E1125 16:19:36.175562 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 16:19:48 crc kubenswrapper[4890]: I1125 16:19:48.173358 4890 scope.go:117] "RemoveContainer" containerID="28fed48468d92f7b1fe5d54d4505526fd8660c4d76009eee3215a28f9eb145da" Nov 25 16:19:48 crc kubenswrapper[4890]: E1125 16:19:48.174124 4890 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xx2tz_openshift-machine-config-operator(4e4f849d-f239-4727-a73e-18327856929a)\"" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" Nov 25 16:19:48 crc kubenswrapper[4890]: I1125 16:19:48.919333 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-hknzk_d97f8477-4597-4fb0-b2fc-e3faf124ba4e/kube-rbac-proxy/0.log" Nov 25 16:19:49 crc kubenswrapper[4890]: I1125 16:19:49.082749 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-hknzk_d97f8477-4597-4fb0-b2fc-e3faf124ba4e/controller/0.log" Nov 25 16:19:49 crc kubenswrapper[4890]: I1125 16:19:49.187963 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhqzp_60b160c7-c63e-4ffc-b031-9c9fa5c248e2/cp-frr-files/0.log" Nov 25 16:19:49 crc kubenswrapper[4890]: I1125 16:19:49.359743 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhqzp_60b160c7-c63e-4ffc-b031-9c9fa5c248e2/cp-reloader/0.log" Nov 25 16:19:49 crc kubenswrapper[4890]: I1125 16:19:49.364381 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhqzp_60b160c7-c63e-4ffc-b031-9c9fa5c248e2/cp-frr-files/0.log" Nov 25 16:19:49 crc kubenswrapper[4890]: I1125 16:19:49.389283 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhqzp_60b160c7-c63e-4ffc-b031-9c9fa5c248e2/cp-reloader/0.log" Nov 25 16:19:49 crc kubenswrapper[4890]: I1125 16:19:49.415713 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhqzp_60b160c7-c63e-4ffc-b031-9c9fa5c248e2/cp-metrics/0.log" Nov 25 16:19:49 crc kubenswrapper[4890]: I1125 16:19:49.598464 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhqzp_60b160c7-c63e-4ffc-b031-9c9fa5c248e2/cp-frr-files/0.log" Nov 25 16:19:49 crc kubenswrapper[4890]: I1125 16:19:49.632938 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhqzp_60b160c7-c63e-4ffc-b031-9c9fa5c248e2/cp-reloader/0.log" Nov 25 16:19:49 crc kubenswrapper[4890]: I1125 16:19:49.647845 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhqzp_60b160c7-c63e-4ffc-b031-9c9fa5c248e2/cp-metrics/0.log" Nov 25 16:19:49 crc kubenswrapper[4890]: I1125 16:19:49.695332 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhqzp_60b160c7-c63e-4ffc-b031-9c9fa5c248e2/cp-metrics/0.log" Nov 25 16:19:49 crc kubenswrapper[4890]: I1125 16:19:49.864115 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhqzp_60b160c7-c63e-4ffc-b031-9c9fa5c248e2/cp-reloader/0.log" Nov 25 16:19:49 crc kubenswrapper[4890]: I1125 16:19:49.871573 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhqzp_60b160c7-c63e-4ffc-b031-9c9fa5c248e2/cp-frr-files/0.log" Nov 25 16:19:49 crc kubenswrapper[4890]: I1125 16:19:49.897798 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhqzp_60b160c7-c63e-4ffc-b031-9c9fa5c248e2/cp-metrics/0.log" Nov 25 16:19:49 crc kubenswrapper[4890]: I1125 16:19:49.916870 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhqzp_60b160c7-c63e-4ffc-b031-9c9fa5c248e2/controller/0.log" Nov 25 16:19:50 crc kubenswrapper[4890]: I1125 16:19:50.080332 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhqzp_60b160c7-c63e-4ffc-b031-9c9fa5c248e2/kube-rbac-proxy/0.log" Nov 25 16:19:50 crc kubenswrapper[4890]: I1125 16:19:50.096240 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhqzp_60b160c7-c63e-4ffc-b031-9c9fa5c248e2/kube-rbac-proxy-frr/0.log" Nov 25 16:19:50 crc kubenswrapper[4890]: I1125 16:19:50.116223 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhqzp_60b160c7-c63e-4ffc-b031-9c9fa5c248e2/frr-metrics/0.log" Nov 25 16:19:50 crc kubenswrapper[4890]: I1125 16:19:50.403688 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhqzp_60b160c7-c63e-4ffc-b031-9c9fa5c248e2/reloader/0.log" Nov 25 16:19:50 crc kubenswrapper[4890]: I1125 16:19:50.525948 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-6998585d5-9ncz9_742ce93d-d49e-4b12-a1cf-5413ec8475ac/frr-k8s-webhook-server/0.log" Nov 25 16:19:50 crc kubenswrapper[4890]: I1125 16:19:50.709795 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-66fc8c7fbf-lnjxm_dad943fe-2209-40f9-a9bb-0be7ba80cca2/manager/3.log" Nov 25 16:19:50 crc kubenswrapper[4890]: I1125 16:19:50.737819 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-66fc8c7fbf-lnjxm_dad943fe-2209-40f9-a9bb-0be7ba80cca2/manager/2.log" Nov 25 16:19:51 crc kubenswrapper[4890]: I1125 16:19:51.033578 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-6b856c5bd9-slq4g_fb5c6944-bf04-42b5-98f6-550b9b2811d5/webhook-server/0.log" Nov 25 16:19:51 crc kubenswrapper[4890]: I1125 16:19:51.213440 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-ndl5n_9c563abc-f1fc-4c15-9a03-901371600338/kube-rbac-proxy/0.log" Nov 25 16:19:51 crc kubenswrapper[4890]: I1125 16:19:51.669009 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nhqzp_60b160c7-c63e-4ffc-b031-9c9fa5c248e2/frr/0.log" Nov 25 16:19:51 crc kubenswrapper[4890]: I1125 16:19:51.733547 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-ndl5n_9c563abc-f1fc-4c15-9a03-901371600338/speaker/0.log" Nov 25 16:20:03 crc kubenswrapper[4890]: I1125 16:20:03.173944 4890 scope.go:117] "RemoveContainer" containerID="28fed48468d92f7b1fe5d54d4505526fd8660c4d76009eee3215a28f9eb145da" Nov 25 16:20:03 crc kubenswrapper[4890]: I1125 16:20:03.644283 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" event={"ID":"4e4f849d-f239-4727-a73e-18327856929a","Type":"ContainerStarted","Data":"dd26fe0760ee6f253c48428804dee70538aa517abe6734bbe073fce009f96770"} Nov 25 16:20:04 crc kubenswrapper[4890]: I1125 16:20:04.586819 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772er4qx8_d89e1065-7553-4d5b-aef6-38108a5cb8ae/util/0.log" Nov 25 16:20:04 crc kubenswrapper[4890]: I1125 16:20:04.889659 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772er4qx8_d89e1065-7553-4d5b-aef6-38108a5cb8ae/pull/0.log" Nov 25 16:20:04 crc kubenswrapper[4890]: I1125 16:20:04.894461 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772er4qx8_d89e1065-7553-4d5b-aef6-38108a5cb8ae/util/0.log" Nov 25 16:20:04 crc kubenswrapper[4890]: I1125 16:20:04.936420 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772er4qx8_d89e1065-7553-4d5b-aef6-38108a5cb8ae/pull/0.log" Nov 25 16:20:05 crc kubenswrapper[4890]: I1125 16:20:05.286483 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772er4qx8_d89e1065-7553-4d5b-aef6-38108a5cb8ae/pull/0.log" Nov 25 16:20:05 crc kubenswrapper[4890]: I1125 16:20:05.308653 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772er4qx8_d89e1065-7553-4d5b-aef6-38108a5cb8ae/extract/0.log" Nov 25 16:20:05 crc kubenswrapper[4890]: I1125 16:20:05.320376 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772er4qx8_d89e1065-7553-4d5b-aef6-38108a5cb8ae/util/0.log" Nov 25 16:20:05 crc kubenswrapper[4890]: I1125 16:20:05.603659 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-fksn8_6afbc261-80f8-4f4a-814f-5df9ab43502e/extract-utilities/0.log" Nov 25 16:20:05 crc kubenswrapper[4890]: I1125 16:20:05.790265 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-fksn8_6afbc261-80f8-4f4a-814f-5df9ab43502e/extract-content/0.log" Nov 25 16:20:05 crc kubenswrapper[4890]: I1125 16:20:05.882542 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-fksn8_6afbc261-80f8-4f4a-814f-5df9ab43502e/extract-content/0.log" Nov 25 16:20:05 crc kubenswrapper[4890]: I1125 16:20:05.889221 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-fksn8_6afbc261-80f8-4f4a-814f-5df9ab43502e/extract-utilities/0.log" Nov 25 16:20:06 crc kubenswrapper[4890]: I1125 16:20:06.165014 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-fksn8_6afbc261-80f8-4f4a-814f-5df9ab43502e/extract-content/0.log" Nov 25 16:20:06 crc kubenswrapper[4890]: I1125 16:20:06.226439 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-fksn8_6afbc261-80f8-4f4a-814f-5df9ab43502e/extract-utilities/0.log" Nov 25 16:20:06 crc kubenswrapper[4890]: I1125 16:20:06.466608 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gxfqw_fc5b6ee3-033f-4b8f-82b7-46e28498f625/extract-utilities/0.log" Nov 25 16:20:06 crc kubenswrapper[4890]: I1125 16:20:06.676425 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gxfqw_fc5b6ee3-033f-4b8f-82b7-46e28498f625/extract-utilities/0.log" Nov 25 16:20:06 crc kubenswrapper[4890]: I1125 16:20:06.768575 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gxfqw_fc5b6ee3-033f-4b8f-82b7-46e28498f625/extract-content/0.log" Nov 25 16:20:06 crc kubenswrapper[4890]: I1125 16:20:06.844552 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gxfqw_fc5b6ee3-033f-4b8f-82b7-46e28498f625/extract-content/0.log" Nov 25 16:20:06 crc kubenswrapper[4890]: I1125 16:20:06.850469 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-fksn8_6afbc261-80f8-4f4a-814f-5df9ab43502e/registry-server/0.log" Nov 25 16:20:06 crc kubenswrapper[4890]: I1125 16:20:06.946513 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gxfqw_fc5b6ee3-033f-4b8f-82b7-46e28498f625/extract-utilities/0.log" Nov 25 16:20:07 crc kubenswrapper[4890]: I1125 16:20:07.028473 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gxfqw_fc5b6ee3-033f-4b8f-82b7-46e28498f625/extract-content/0.log" Nov 25 16:20:07 crc kubenswrapper[4890]: I1125 16:20:07.232194 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mhgqc_76fcec1c-1c27-4e7f-9f3e-a2527b72581f/util/0.log" Nov 25 16:20:07 crc kubenswrapper[4890]: I1125 16:20:07.423535 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mhgqc_76fcec1c-1c27-4e7f-9f3e-a2527b72581f/util/0.log" Nov 25 16:20:07 crc kubenswrapper[4890]: I1125 16:20:07.427131 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mhgqc_76fcec1c-1c27-4e7f-9f3e-a2527b72581f/pull/0.log" Nov 25 16:20:07 crc kubenswrapper[4890]: I1125 16:20:07.466888 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mhgqc_76fcec1c-1c27-4e7f-9f3e-a2527b72581f/pull/0.log" Nov 25 16:20:07 crc kubenswrapper[4890]: I1125 16:20:07.516153 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gxfqw_fc5b6ee3-033f-4b8f-82b7-46e28498f625/registry-server/0.log" Nov 25 16:20:07 crc kubenswrapper[4890]: I1125 16:20:07.677188 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mhgqc_76fcec1c-1c27-4e7f-9f3e-a2527b72581f/pull/0.log" Nov 25 16:20:07 crc kubenswrapper[4890]: I1125 16:20:07.691039 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mhgqc_76fcec1c-1c27-4e7f-9f3e-a2527b72581f/util/0.log" Nov 25 16:20:07 crc kubenswrapper[4890]: I1125 16:20:07.709747 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6mhgqc_76fcec1c-1c27-4e7f-9f3e-a2527b72581f/extract/0.log" Nov 25 16:20:07 crc kubenswrapper[4890]: I1125 16:20:07.959446 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-2xzsf_4bbe88e6-b859-452a-baf1-8a963b475014/marketplace-operator/1.log" Nov 25 16:20:07 crc kubenswrapper[4890]: I1125 16:20:07.980940 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-2xzsf_4bbe88e6-b859-452a-baf1-8a963b475014/marketplace-operator/0.log" Nov 25 16:20:08 crc kubenswrapper[4890]: I1125 16:20:08.006096 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-tjd2v_f444e25c-b3bf-4046-a1ff-eba1a422a88e/extract-utilities/0.log" Nov 25 16:20:08 crc kubenswrapper[4890]: I1125 16:20:08.220087 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-tjd2v_f444e25c-b3bf-4046-a1ff-eba1a422a88e/extract-utilities/0.log" Nov 25 16:20:08 crc kubenswrapper[4890]: I1125 16:20:08.237374 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-tjd2v_f444e25c-b3bf-4046-a1ff-eba1a422a88e/extract-content/0.log" Nov 25 16:20:08 crc kubenswrapper[4890]: I1125 16:20:08.270903 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-tjd2v_f444e25c-b3bf-4046-a1ff-eba1a422a88e/extract-content/0.log" Nov 25 16:20:08 crc kubenswrapper[4890]: I1125 16:20:08.584601 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-tjd2v_f444e25c-b3bf-4046-a1ff-eba1a422a88e/extract-content/0.log" Nov 25 16:20:08 crc kubenswrapper[4890]: I1125 16:20:08.641300 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-tjd2v_f444e25c-b3bf-4046-a1ff-eba1a422a88e/extract-utilities/0.log" Nov 25 16:20:08 crc kubenswrapper[4890]: I1125 16:20:08.808567 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jcmgp_c83811ae-c669-4268-a58d-e4b44d18f26c/extract-utilities/0.log" Nov 25 16:20:08 crc kubenswrapper[4890]: I1125 16:20:08.852750 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-tjd2v_f444e25c-b3bf-4046-a1ff-eba1a422a88e/registry-server/0.log" Nov 25 16:20:09 crc kubenswrapper[4890]: I1125 16:20:09.067301 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jcmgp_c83811ae-c669-4268-a58d-e4b44d18f26c/extract-content/0.log" Nov 25 16:20:09 crc kubenswrapper[4890]: I1125 16:20:09.125671 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jcmgp_c83811ae-c669-4268-a58d-e4b44d18f26c/extract-utilities/0.log" Nov 25 16:20:09 crc kubenswrapper[4890]: I1125 16:20:09.135028 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jcmgp_c83811ae-c669-4268-a58d-e4b44d18f26c/extract-content/0.log" Nov 25 16:20:09 crc kubenswrapper[4890]: I1125 16:20:09.312751 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jcmgp_c83811ae-c669-4268-a58d-e4b44d18f26c/extract-content/0.log" Nov 25 16:20:09 crc kubenswrapper[4890]: I1125 16:20:09.314783 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jcmgp_c83811ae-c669-4268-a58d-e4b44d18f26c/extract-utilities/0.log" Nov 25 16:20:09 crc kubenswrapper[4890]: I1125 16:20:09.966887 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jcmgp_c83811ae-c669-4268-a58d-e4b44d18f26c/registry-server/0.log" Nov 25 16:20:36 crc kubenswrapper[4890]: E1125 16:20:36.605928 4890 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.12:53200->38.102.83.12:39007: write tcp 38.102.83.12:53200->38.102.83.12:39007: write: broken pipe Nov 25 16:20:42 crc kubenswrapper[4890]: E1125 16:20:42.456743 4890 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.12:53410->38.102.83.12:39007: write tcp 38.102.83.12:53410->38.102.83.12:39007: write: broken pipe Nov 25 16:21:53 crc kubenswrapper[4890]: I1125 16:21:53.758295 4890 generic.go:334] "Generic (PLEG): container finished" podID="cf177367-00de-4aa9-a1ff-124727968f00" containerID="01d2f28946f5e78f665617387571ce72aae378ec04839e128c4e8c9af3d58924" exitCode=0 Nov 25 16:21:53 crc kubenswrapper[4890]: I1125 16:21:53.758404 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t78gs/must-gather-5gfdb" event={"ID":"cf177367-00de-4aa9-a1ff-124727968f00","Type":"ContainerDied","Data":"01d2f28946f5e78f665617387571ce72aae378ec04839e128c4e8c9af3d58924"} Nov 25 16:21:53 crc kubenswrapper[4890]: I1125 16:21:53.759261 4890 scope.go:117] "RemoveContainer" containerID="01d2f28946f5e78f665617387571ce72aae378ec04839e128c4e8c9af3d58924" Nov 25 16:21:54 crc kubenswrapper[4890]: I1125 16:21:54.836855 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-t78gs_must-gather-5gfdb_cf177367-00de-4aa9-a1ff-124727968f00/gather/0.log" Nov 25 16:22:04 crc kubenswrapper[4890]: I1125 16:22:04.985587 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-t78gs/must-gather-5gfdb"] Nov 25 16:22:04 crc kubenswrapper[4890]: I1125 16:22:04.986592 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-t78gs/must-gather-5gfdb" podUID="cf177367-00de-4aa9-a1ff-124727968f00" containerName="copy" containerID="cri-o://9badd2e0e1ceb14d744cf5207ffbc0d863dff27bf2b9ee43bb0c325621e13a94" gracePeriod=2 Nov 25 16:22:04 crc kubenswrapper[4890]: I1125 16:22:04.996283 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-t78gs/must-gather-5gfdb"] Nov 25 16:22:05 crc kubenswrapper[4890]: I1125 16:22:05.573183 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-t78gs_must-gather-5gfdb_cf177367-00de-4aa9-a1ff-124727968f00/copy/0.log" Nov 25 16:22:05 crc kubenswrapper[4890]: I1125 16:22:05.577585 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t78gs/must-gather-5gfdb" Nov 25 16:22:05 crc kubenswrapper[4890]: I1125 16:22:05.740107 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cf177367-00de-4aa9-a1ff-124727968f00-must-gather-output\") pod \"cf177367-00de-4aa9-a1ff-124727968f00\" (UID: \"cf177367-00de-4aa9-a1ff-124727968f00\") " Nov 25 16:22:05 crc kubenswrapper[4890]: I1125 16:22:05.740204 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4vn59\" (UniqueName: \"kubernetes.io/projected/cf177367-00de-4aa9-a1ff-124727968f00-kube-api-access-4vn59\") pod \"cf177367-00de-4aa9-a1ff-124727968f00\" (UID: \"cf177367-00de-4aa9-a1ff-124727968f00\") " Nov 25 16:22:05 crc kubenswrapper[4890]: I1125 16:22:05.750845 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf177367-00de-4aa9-a1ff-124727968f00-kube-api-access-4vn59" (OuterVolumeSpecName: "kube-api-access-4vn59") pod "cf177367-00de-4aa9-a1ff-124727968f00" (UID: "cf177367-00de-4aa9-a1ff-124727968f00"). InnerVolumeSpecName "kube-api-access-4vn59". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:22:05 crc kubenswrapper[4890]: I1125 16:22:05.842077 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4vn59\" (UniqueName: \"kubernetes.io/projected/cf177367-00de-4aa9-a1ff-124727968f00-kube-api-access-4vn59\") on node \"crc\" DevicePath \"\"" Nov 25 16:22:05 crc kubenswrapper[4890]: I1125 16:22:05.887400 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf177367-00de-4aa9-a1ff-124727968f00-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "cf177367-00de-4aa9-a1ff-124727968f00" (UID: "cf177367-00de-4aa9-a1ff-124727968f00"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:22:05 crc kubenswrapper[4890]: I1125 16:22:05.904048 4890 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-t78gs_must-gather-5gfdb_cf177367-00de-4aa9-a1ff-124727968f00/copy/0.log" Nov 25 16:22:05 crc kubenswrapper[4890]: I1125 16:22:05.904753 4890 generic.go:334] "Generic (PLEG): container finished" podID="cf177367-00de-4aa9-a1ff-124727968f00" containerID="9badd2e0e1ceb14d744cf5207ffbc0d863dff27bf2b9ee43bb0c325621e13a94" exitCode=143 Nov 25 16:22:05 crc kubenswrapper[4890]: I1125 16:22:05.904849 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t78gs/must-gather-5gfdb" Nov 25 16:22:05 crc kubenswrapper[4890]: I1125 16:22:05.904950 4890 scope.go:117] "RemoveContainer" containerID="9badd2e0e1ceb14d744cf5207ffbc0d863dff27bf2b9ee43bb0c325621e13a94" Nov 25 16:22:05 crc kubenswrapper[4890]: I1125 16:22:05.932801 4890 scope.go:117] "RemoveContainer" containerID="01d2f28946f5e78f665617387571ce72aae378ec04839e128c4e8c9af3d58924" Nov 25 16:22:05 crc kubenswrapper[4890]: I1125 16:22:05.945018 4890 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cf177367-00de-4aa9-a1ff-124727968f00-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 25 16:22:06 crc kubenswrapper[4890]: I1125 16:22:06.015979 4890 scope.go:117] "RemoveContainer" containerID="9badd2e0e1ceb14d744cf5207ffbc0d863dff27bf2b9ee43bb0c325621e13a94" Nov 25 16:22:06 crc kubenswrapper[4890]: E1125 16:22:06.016709 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9badd2e0e1ceb14d744cf5207ffbc0d863dff27bf2b9ee43bb0c325621e13a94\": container with ID starting with 9badd2e0e1ceb14d744cf5207ffbc0d863dff27bf2b9ee43bb0c325621e13a94 not found: ID does not exist" containerID="9badd2e0e1ceb14d744cf5207ffbc0d863dff27bf2b9ee43bb0c325621e13a94" Nov 25 16:22:06 crc kubenswrapper[4890]: I1125 16:22:06.016776 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9badd2e0e1ceb14d744cf5207ffbc0d863dff27bf2b9ee43bb0c325621e13a94"} err="failed to get container status \"9badd2e0e1ceb14d744cf5207ffbc0d863dff27bf2b9ee43bb0c325621e13a94\": rpc error: code = NotFound desc = could not find container \"9badd2e0e1ceb14d744cf5207ffbc0d863dff27bf2b9ee43bb0c325621e13a94\": container with ID starting with 9badd2e0e1ceb14d744cf5207ffbc0d863dff27bf2b9ee43bb0c325621e13a94 not found: ID does not exist" Nov 25 16:22:06 crc kubenswrapper[4890]: I1125 16:22:06.016815 4890 scope.go:117] "RemoveContainer" containerID="01d2f28946f5e78f665617387571ce72aae378ec04839e128c4e8c9af3d58924" Nov 25 16:22:06 crc kubenswrapper[4890]: E1125 16:22:06.018259 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01d2f28946f5e78f665617387571ce72aae378ec04839e128c4e8c9af3d58924\": container with ID starting with 01d2f28946f5e78f665617387571ce72aae378ec04839e128c4e8c9af3d58924 not found: ID does not exist" containerID="01d2f28946f5e78f665617387571ce72aae378ec04839e128c4e8c9af3d58924" Nov 25 16:22:06 crc kubenswrapper[4890]: I1125 16:22:06.018303 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01d2f28946f5e78f665617387571ce72aae378ec04839e128c4e8c9af3d58924"} err="failed to get container status \"01d2f28946f5e78f665617387571ce72aae378ec04839e128c4e8c9af3d58924\": rpc error: code = NotFound desc = could not find container \"01d2f28946f5e78f665617387571ce72aae378ec04839e128c4e8c9af3d58924\": container with ID starting with 01d2f28946f5e78f665617387571ce72aae378ec04839e128c4e8c9af3d58924 not found: ID does not exist" Nov 25 16:22:06 crc kubenswrapper[4890]: I1125 16:22:06.185472 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf177367-00de-4aa9-a1ff-124727968f00" path="/var/lib/kubelet/pods/cf177367-00de-4aa9-a1ff-124727968f00/volumes" Nov 25 16:22:26 crc kubenswrapper[4890]: I1125 16:22:26.447007 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 16:22:26 crc kubenswrapper[4890]: I1125 16:22:26.447807 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 16:22:56 crc kubenswrapper[4890]: I1125 16:22:56.447984 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 16:22:56 crc kubenswrapper[4890]: I1125 16:22:56.448609 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 16:23:07 crc kubenswrapper[4890]: I1125 16:23:07.465533 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5bgq9"] Nov 25 16:23:07 crc kubenswrapper[4890]: E1125 16:23:07.468567 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf177367-00de-4aa9-a1ff-124727968f00" containerName="copy" Nov 25 16:23:07 crc kubenswrapper[4890]: I1125 16:23:07.468806 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf177367-00de-4aa9-a1ff-124727968f00" containerName="copy" Nov 25 16:23:07 crc kubenswrapper[4890]: E1125 16:23:07.468969 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="924a90dd-2aec-4eff-beee-1018698908f6" containerName="extract-content" Nov 25 16:23:07 crc kubenswrapper[4890]: I1125 16:23:07.469143 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="924a90dd-2aec-4eff-beee-1018698908f6" containerName="extract-content" Nov 25 16:23:07 crc kubenswrapper[4890]: E1125 16:23:07.469335 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf177367-00de-4aa9-a1ff-124727968f00" containerName="gather" Nov 25 16:23:07 crc kubenswrapper[4890]: I1125 16:23:07.469468 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf177367-00de-4aa9-a1ff-124727968f00" containerName="gather" Nov 25 16:23:07 crc kubenswrapper[4890]: E1125 16:23:07.469646 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="924a90dd-2aec-4eff-beee-1018698908f6" containerName="registry-server" Nov 25 16:23:07 crc kubenswrapper[4890]: I1125 16:23:07.469784 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="924a90dd-2aec-4eff-beee-1018698908f6" containerName="registry-server" Nov 25 16:23:07 crc kubenswrapper[4890]: E1125 16:23:07.469932 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="924a90dd-2aec-4eff-beee-1018698908f6" containerName="extract-utilities" Nov 25 16:23:07 crc kubenswrapper[4890]: I1125 16:23:07.470049 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="924a90dd-2aec-4eff-beee-1018698908f6" containerName="extract-utilities" Nov 25 16:23:07 crc kubenswrapper[4890]: I1125 16:23:07.470621 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf177367-00de-4aa9-a1ff-124727968f00" containerName="gather" Nov 25 16:23:07 crc kubenswrapper[4890]: I1125 16:23:07.470812 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf177367-00de-4aa9-a1ff-124727968f00" containerName="copy" Nov 25 16:23:07 crc kubenswrapper[4890]: I1125 16:23:07.470946 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="924a90dd-2aec-4eff-beee-1018698908f6" containerName="registry-server" Nov 25 16:23:07 crc kubenswrapper[4890]: I1125 16:23:07.473675 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5bgq9" Nov 25 16:23:07 crc kubenswrapper[4890]: I1125 16:23:07.480631 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5bgq9"] Nov 25 16:23:07 crc kubenswrapper[4890]: I1125 16:23:07.600497 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73aeebf4-f3e4-481f-997d-b019f7aa8cba-catalog-content\") pod \"redhat-operators-5bgq9\" (UID: \"73aeebf4-f3e4-481f-997d-b019f7aa8cba\") " pod="openshift-marketplace/redhat-operators-5bgq9" Nov 25 16:23:07 crc kubenswrapper[4890]: I1125 16:23:07.600916 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdxdn\" (UniqueName: \"kubernetes.io/projected/73aeebf4-f3e4-481f-997d-b019f7aa8cba-kube-api-access-xdxdn\") pod \"redhat-operators-5bgq9\" (UID: \"73aeebf4-f3e4-481f-997d-b019f7aa8cba\") " pod="openshift-marketplace/redhat-operators-5bgq9" Nov 25 16:23:07 crc kubenswrapper[4890]: I1125 16:23:07.601048 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73aeebf4-f3e4-481f-997d-b019f7aa8cba-utilities\") pod \"redhat-operators-5bgq9\" (UID: \"73aeebf4-f3e4-481f-997d-b019f7aa8cba\") " pod="openshift-marketplace/redhat-operators-5bgq9" Nov 25 16:23:07 crc kubenswrapper[4890]: I1125 16:23:07.702560 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73aeebf4-f3e4-481f-997d-b019f7aa8cba-catalog-content\") pod \"redhat-operators-5bgq9\" (UID: \"73aeebf4-f3e4-481f-997d-b019f7aa8cba\") " pod="openshift-marketplace/redhat-operators-5bgq9" Nov 25 16:23:07 crc kubenswrapper[4890]: I1125 16:23:07.702632 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdxdn\" (UniqueName: \"kubernetes.io/projected/73aeebf4-f3e4-481f-997d-b019f7aa8cba-kube-api-access-xdxdn\") pod \"redhat-operators-5bgq9\" (UID: \"73aeebf4-f3e4-481f-997d-b019f7aa8cba\") " pod="openshift-marketplace/redhat-operators-5bgq9" Nov 25 16:23:07 crc kubenswrapper[4890]: I1125 16:23:07.702710 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73aeebf4-f3e4-481f-997d-b019f7aa8cba-utilities\") pod \"redhat-operators-5bgq9\" (UID: \"73aeebf4-f3e4-481f-997d-b019f7aa8cba\") " pod="openshift-marketplace/redhat-operators-5bgq9" Nov 25 16:23:07 crc kubenswrapper[4890]: I1125 16:23:07.703133 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73aeebf4-f3e4-481f-997d-b019f7aa8cba-catalog-content\") pod \"redhat-operators-5bgq9\" (UID: \"73aeebf4-f3e4-481f-997d-b019f7aa8cba\") " pod="openshift-marketplace/redhat-operators-5bgq9" Nov 25 16:23:07 crc kubenswrapper[4890]: I1125 16:23:07.703209 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73aeebf4-f3e4-481f-997d-b019f7aa8cba-utilities\") pod \"redhat-operators-5bgq9\" (UID: \"73aeebf4-f3e4-481f-997d-b019f7aa8cba\") " pod="openshift-marketplace/redhat-operators-5bgq9" Nov 25 16:23:07 crc kubenswrapper[4890]: I1125 16:23:07.736843 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdxdn\" (UniqueName: \"kubernetes.io/projected/73aeebf4-f3e4-481f-997d-b019f7aa8cba-kube-api-access-xdxdn\") pod \"redhat-operators-5bgq9\" (UID: \"73aeebf4-f3e4-481f-997d-b019f7aa8cba\") " pod="openshift-marketplace/redhat-operators-5bgq9" Nov 25 16:23:07 crc kubenswrapper[4890]: I1125 16:23:07.800211 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5bgq9" Nov 25 16:23:08 crc kubenswrapper[4890]: I1125 16:23:08.321685 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5bgq9"] Nov 25 16:23:08 crc kubenswrapper[4890]: I1125 16:23:08.540965 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5bgq9" event={"ID":"73aeebf4-f3e4-481f-997d-b019f7aa8cba","Type":"ContainerStarted","Data":"ff2d817569c925414fd355a71b5cfa5f2e3a959c5539c787a020c6025d773855"} Nov 25 16:23:08 crc kubenswrapper[4890]: I1125 16:23:08.541437 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5bgq9" event={"ID":"73aeebf4-f3e4-481f-997d-b019f7aa8cba","Type":"ContainerStarted","Data":"705e99af3177a997ce7214817b7a5e0526fe0cf93734969f9d2a095951ba76c3"} Nov 25 16:23:09 crc kubenswrapper[4890]: I1125 16:23:09.553445 4890 generic.go:334] "Generic (PLEG): container finished" podID="73aeebf4-f3e4-481f-997d-b019f7aa8cba" containerID="ff2d817569c925414fd355a71b5cfa5f2e3a959c5539c787a020c6025d773855" exitCode=0 Nov 25 16:23:09 crc kubenswrapper[4890]: I1125 16:23:09.553625 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5bgq9" event={"ID":"73aeebf4-f3e4-481f-997d-b019f7aa8cba","Type":"ContainerDied","Data":"ff2d817569c925414fd355a71b5cfa5f2e3a959c5539c787a020c6025d773855"} Nov 25 16:23:09 crc kubenswrapper[4890]: I1125 16:23:09.557536 4890 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 16:23:10 crc kubenswrapper[4890]: I1125 16:23:10.564248 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5bgq9" event={"ID":"73aeebf4-f3e4-481f-997d-b019f7aa8cba","Type":"ContainerStarted","Data":"6b1abeeae137dab54175bb4bb7d5191f3de344dec47945ae357f42ee79b1ef3c"} Nov 25 16:23:14 crc kubenswrapper[4890]: I1125 16:23:14.603907 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5bgq9" event={"ID":"73aeebf4-f3e4-481f-997d-b019f7aa8cba","Type":"ContainerDied","Data":"6b1abeeae137dab54175bb4bb7d5191f3de344dec47945ae357f42ee79b1ef3c"} Nov 25 16:23:14 crc kubenswrapper[4890]: I1125 16:23:14.603928 4890 generic.go:334] "Generic (PLEG): container finished" podID="73aeebf4-f3e4-481f-997d-b019f7aa8cba" containerID="6b1abeeae137dab54175bb4bb7d5191f3de344dec47945ae357f42ee79b1ef3c" exitCode=0 Nov 25 16:23:16 crc kubenswrapper[4890]: I1125 16:23:16.626914 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5bgq9" event={"ID":"73aeebf4-f3e4-481f-997d-b019f7aa8cba","Type":"ContainerStarted","Data":"790d9e2d2d793c062496e9319cada1c20cb83911fa2faa802932e2b821f53376"} Nov 25 16:23:16 crc kubenswrapper[4890]: I1125 16:23:16.651999 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5bgq9" podStartSLOduration=4.011834141 podStartE2EDuration="9.651979686s" podCreationTimestamp="2025-11-25 16:23:07 +0000 UTC" firstStartedPulling="2025-11-25 16:23:09.557144565 +0000 UTC m=+4847.999607175" lastFinishedPulling="2025-11-25 16:23:15.19729011 +0000 UTC m=+4853.639752720" observedRunningTime="2025-11-25 16:23:16.644903809 +0000 UTC m=+4855.087366429" watchObservedRunningTime="2025-11-25 16:23:16.651979686 +0000 UTC m=+4855.094442296" Nov 25 16:23:17 crc kubenswrapper[4890]: I1125 16:23:17.800427 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5bgq9" Nov 25 16:23:17 crc kubenswrapper[4890]: I1125 16:23:17.800502 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5bgq9" Nov 25 16:23:18 crc kubenswrapper[4890]: I1125 16:23:18.871933 4890 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5bgq9" podUID="73aeebf4-f3e4-481f-997d-b019f7aa8cba" containerName="registry-server" probeResult="failure" output=< Nov 25 16:23:18 crc kubenswrapper[4890]: timeout: failed to connect service ":50051" within 1s Nov 25 16:23:18 crc kubenswrapper[4890]: > Nov 25 16:23:26 crc kubenswrapper[4890]: I1125 16:23:26.448298 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 16:23:26 crc kubenswrapper[4890]: I1125 16:23:26.449268 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 16:23:26 crc kubenswrapper[4890]: I1125 16:23:26.449321 4890 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" Nov 25 16:23:26 crc kubenswrapper[4890]: I1125 16:23:26.450813 4890 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dd26fe0760ee6f253c48428804dee70538aa517abe6734bbe073fce009f96770"} pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 16:23:26 crc kubenswrapper[4890]: I1125 16:23:26.450875 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" containerID="cri-o://dd26fe0760ee6f253c48428804dee70538aa517abe6734bbe073fce009f96770" gracePeriod=600 Nov 25 16:23:27 crc kubenswrapper[4890]: I1125 16:23:27.745569 4890 generic.go:334] "Generic (PLEG): container finished" podID="4e4f849d-f239-4727-a73e-18327856929a" containerID="dd26fe0760ee6f253c48428804dee70538aa517abe6734bbe073fce009f96770" exitCode=0 Nov 25 16:23:27 crc kubenswrapper[4890]: I1125 16:23:27.745665 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" event={"ID":"4e4f849d-f239-4727-a73e-18327856929a","Type":"ContainerDied","Data":"dd26fe0760ee6f253c48428804dee70538aa517abe6734bbe073fce009f96770"} Nov 25 16:23:27 crc kubenswrapper[4890]: I1125 16:23:27.746201 4890 scope.go:117] "RemoveContainer" containerID="28fed48468d92f7b1fe5d54d4505526fd8660c4d76009eee3215a28f9eb145da" Nov 25 16:23:27 crc kubenswrapper[4890]: I1125 16:23:27.857527 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5bgq9" Nov 25 16:23:27 crc kubenswrapper[4890]: I1125 16:23:27.921569 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5bgq9" Nov 25 16:23:28 crc kubenswrapper[4890]: I1125 16:23:28.105413 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5bgq9"] Nov 25 16:23:28 crc kubenswrapper[4890]: I1125 16:23:28.757542 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" event={"ID":"4e4f849d-f239-4727-a73e-18327856929a","Type":"ContainerStarted","Data":"ea715daefc570123e2c4f82d23055f783b8fbf602875e1c1f118f5cf43d7b395"} Nov 25 16:23:29 crc kubenswrapper[4890]: I1125 16:23:29.766766 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5bgq9" podUID="73aeebf4-f3e4-481f-997d-b019f7aa8cba" containerName="registry-server" containerID="cri-o://790d9e2d2d793c062496e9319cada1c20cb83911fa2faa802932e2b821f53376" gracePeriod=2 Nov 25 16:23:30 crc kubenswrapper[4890]: I1125 16:23:30.210273 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5bgq9" Nov 25 16:23:30 crc kubenswrapper[4890]: I1125 16:23:30.309880 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xdxdn\" (UniqueName: \"kubernetes.io/projected/73aeebf4-f3e4-481f-997d-b019f7aa8cba-kube-api-access-xdxdn\") pod \"73aeebf4-f3e4-481f-997d-b019f7aa8cba\" (UID: \"73aeebf4-f3e4-481f-997d-b019f7aa8cba\") " Nov 25 16:23:30 crc kubenswrapper[4890]: I1125 16:23:30.310058 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73aeebf4-f3e4-481f-997d-b019f7aa8cba-utilities\") pod \"73aeebf4-f3e4-481f-997d-b019f7aa8cba\" (UID: \"73aeebf4-f3e4-481f-997d-b019f7aa8cba\") " Nov 25 16:23:30 crc kubenswrapper[4890]: I1125 16:23:30.310089 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73aeebf4-f3e4-481f-997d-b019f7aa8cba-catalog-content\") pod \"73aeebf4-f3e4-481f-997d-b019f7aa8cba\" (UID: \"73aeebf4-f3e4-481f-997d-b019f7aa8cba\") " Nov 25 16:23:30 crc kubenswrapper[4890]: I1125 16:23:30.311007 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73aeebf4-f3e4-481f-997d-b019f7aa8cba-utilities" (OuterVolumeSpecName: "utilities") pod "73aeebf4-f3e4-481f-997d-b019f7aa8cba" (UID: "73aeebf4-f3e4-481f-997d-b019f7aa8cba"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:23:30 crc kubenswrapper[4890]: I1125 16:23:30.318235 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73aeebf4-f3e4-481f-997d-b019f7aa8cba-kube-api-access-xdxdn" (OuterVolumeSpecName: "kube-api-access-xdxdn") pod "73aeebf4-f3e4-481f-997d-b019f7aa8cba" (UID: "73aeebf4-f3e4-481f-997d-b019f7aa8cba"). InnerVolumeSpecName "kube-api-access-xdxdn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:23:30 crc kubenswrapper[4890]: I1125 16:23:30.401855 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73aeebf4-f3e4-481f-997d-b019f7aa8cba-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "73aeebf4-f3e4-481f-997d-b019f7aa8cba" (UID: "73aeebf4-f3e4-481f-997d-b019f7aa8cba"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:23:30 crc kubenswrapper[4890]: I1125 16:23:30.412396 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xdxdn\" (UniqueName: \"kubernetes.io/projected/73aeebf4-f3e4-481f-997d-b019f7aa8cba-kube-api-access-xdxdn\") on node \"crc\" DevicePath \"\"" Nov 25 16:23:30 crc kubenswrapper[4890]: I1125 16:23:30.412418 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73aeebf4-f3e4-481f-997d-b019f7aa8cba-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 16:23:30 crc kubenswrapper[4890]: I1125 16:23:30.412429 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73aeebf4-f3e4-481f-997d-b019f7aa8cba-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 16:23:30 crc kubenswrapper[4890]: I1125 16:23:30.777476 4890 generic.go:334] "Generic (PLEG): container finished" podID="73aeebf4-f3e4-481f-997d-b019f7aa8cba" containerID="790d9e2d2d793c062496e9319cada1c20cb83911fa2faa802932e2b821f53376" exitCode=0 Nov 25 16:23:30 crc kubenswrapper[4890]: I1125 16:23:30.777532 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5bgq9" Nov 25 16:23:30 crc kubenswrapper[4890]: I1125 16:23:30.777575 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5bgq9" event={"ID":"73aeebf4-f3e4-481f-997d-b019f7aa8cba","Type":"ContainerDied","Data":"790d9e2d2d793c062496e9319cada1c20cb83911fa2faa802932e2b821f53376"} Nov 25 16:23:30 crc kubenswrapper[4890]: I1125 16:23:30.778101 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5bgq9" event={"ID":"73aeebf4-f3e4-481f-997d-b019f7aa8cba","Type":"ContainerDied","Data":"705e99af3177a997ce7214817b7a5e0526fe0cf93734969f9d2a095951ba76c3"} Nov 25 16:23:30 crc kubenswrapper[4890]: I1125 16:23:30.778194 4890 scope.go:117] "RemoveContainer" containerID="790d9e2d2d793c062496e9319cada1c20cb83911fa2faa802932e2b821f53376" Nov 25 16:23:30 crc kubenswrapper[4890]: I1125 16:23:30.804607 4890 scope.go:117] "RemoveContainer" containerID="6b1abeeae137dab54175bb4bb7d5191f3de344dec47945ae357f42ee79b1ef3c" Nov 25 16:23:30 crc kubenswrapper[4890]: I1125 16:23:30.829065 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5bgq9"] Nov 25 16:23:30 crc kubenswrapper[4890]: I1125 16:23:30.845242 4890 scope.go:117] "RemoveContainer" containerID="ff2d817569c925414fd355a71b5cfa5f2e3a959c5539c787a020c6025d773855" Nov 25 16:23:30 crc kubenswrapper[4890]: I1125 16:23:30.846367 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5bgq9"] Nov 25 16:23:30 crc kubenswrapper[4890]: I1125 16:23:30.887534 4890 scope.go:117] "RemoveContainer" containerID="790d9e2d2d793c062496e9319cada1c20cb83911fa2faa802932e2b821f53376" Nov 25 16:23:30 crc kubenswrapper[4890]: E1125 16:23:30.888286 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"790d9e2d2d793c062496e9319cada1c20cb83911fa2faa802932e2b821f53376\": container with ID starting with 790d9e2d2d793c062496e9319cada1c20cb83911fa2faa802932e2b821f53376 not found: ID does not exist" containerID="790d9e2d2d793c062496e9319cada1c20cb83911fa2faa802932e2b821f53376" Nov 25 16:23:30 crc kubenswrapper[4890]: I1125 16:23:30.888444 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"790d9e2d2d793c062496e9319cada1c20cb83911fa2faa802932e2b821f53376"} err="failed to get container status \"790d9e2d2d793c062496e9319cada1c20cb83911fa2faa802932e2b821f53376\": rpc error: code = NotFound desc = could not find container \"790d9e2d2d793c062496e9319cada1c20cb83911fa2faa802932e2b821f53376\": container with ID starting with 790d9e2d2d793c062496e9319cada1c20cb83911fa2faa802932e2b821f53376 not found: ID does not exist" Nov 25 16:23:30 crc kubenswrapper[4890]: I1125 16:23:30.888577 4890 scope.go:117] "RemoveContainer" containerID="6b1abeeae137dab54175bb4bb7d5191f3de344dec47945ae357f42ee79b1ef3c" Nov 25 16:23:30 crc kubenswrapper[4890]: E1125 16:23:30.889094 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b1abeeae137dab54175bb4bb7d5191f3de344dec47945ae357f42ee79b1ef3c\": container with ID starting with 6b1abeeae137dab54175bb4bb7d5191f3de344dec47945ae357f42ee79b1ef3c not found: ID does not exist" containerID="6b1abeeae137dab54175bb4bb7d5191f3de344dec47945ae357f42ee79b1ef3c" Nov 25 16:23:30 crc kubenswrapper[4890]: I1125 16:23:30.889202 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b1abeeae137dab54175bb4bb7d5191f3de344dec47945ae357f42ee79b1ef3c"} err="failed to get container status \"6b1abeeae137dab54175bb4bb7d5191f3de344dec47945ae357f42ee79b1ef3c\": rpc error: code = NotFound desc = could not find container \"6b1abeeae137dab54175bb4bb7d5191f3de344dec47945ae357f42ee79b1ef3c\": container with ID starting with 6b1abeeae137dab54175bb4bb7d5191f3de344dec47945ae357f42ee79b1ef3c not found: ID does not exist" Nov 25 16:23:30 crc kubenswrapper[4890]: I1125 16:23:30.889283 4890 scope.go:117] "RemoveContainer" containerID="ff2d817569c925414fd355a71b5cfa5f2e3a959c5539c787a020c6025d773855" Nov 25 16:23:30 crc kubenswrapper[4890]: E1125 16:23:30.889723 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff2d817569c925414fd355a71b5cfa5f2e3a959c5539c787a020c6025d773855\": container with ID starting with ff2d817569c925414fd355a71b5cfa5f2e3a959c5539c787a020c6025d773855 not found: ID does not exist" containerID="ff2d817569c925414fd355a71b5cfa5f2e3a959c5539c787a020c6025d773855" Nov 25 16:23:30 crc kubenswrapper[4890]: I1125 16:23:30.889826 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff2d817569c925414fd355a71b5cfa5f2e3a959c5539c787a020c6025d773855"} err="failed to get container status \"ff2d817569c925414fd355a71b5cfa5f2e3a959c5539c787a020c6025d773855\": rpc error: code = NotFound desc = could not find container \"ff2d817569c925414fd355a71b5cfa5f2e3a959c5539c787a020c6025d773855\": container with ID starting with ff2d817569c925414fd355a71b5cfa5f2e3a959c5539c787a020c6025d773855 not found: ID does not exist" Nov 25 16:23:32 crc kubenswrapper[4890]: I1125 16:23:32.197705 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73aeebf4-f3e4-481f-997d-b019f7aa8cba" path="/var/lib/kubelet/pods/73aeebf4-f3e4-481f-997d-b019f7aa8cba/volumes" Nov 25 16:23:41 crc kubenswrapper[4890]: I1125 16:23:41.198268 4890 scope.go:117] "RemoveContainer" containerID="01a40daf86169b385ee69dcdf255d666eed7a7c14b963c23599beed322d2ac31" Nov 25 16:23:41 crc kubenswrapper[4890]: I1125 16:23:41.231209 4890 scope.go:117] "RemoveContainer" containerID="8d6c037730429a01e68fa7622e1b7c5fdbc0ffaa7e00575d34401c31641afed1" Nov 25 16:25:28 crc kubenswrapper[4890]: I1125 16:25:28.670662 4890 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zx8vw"] Nov 25 16:25:28 crc kubenswrapper[4890]: E1125 16:25:28.671476 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73aeebf4-f3e4-481f-997d-b019f7aa8cba" containerName="extract-content" Nov 25 16:25:28 crc kubenswrapper[4890]: I1125 16:25:28.671490 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="73aeebf4-f3e4-481f-997d-b019f7aa8cba" containerName="extract-content" Nov 25 16:25:28 crc kubenswrapper[4890]: E1125 16:25:28.671505 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73aeebf4-f3e4-481f-997d-b019f7aa8cba" containerName="registry-server" Nov 25 16:25:28 crc kubenswrapper[4890]: I1125 16:25:28.671511 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="73aeebf4-f3e4-481f-997d-b019f7aa8cba" containerName="registry-server" Nov 25 16:25:28 crc kubenswrapper[4890]: E1125 16:25:28.671527 4890 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73aeebf4-f3e4-481f-997d-b019f7aa8cba" containerName="extract-utilities" Nov 25 16:25:28 crc kubenswrapper[4890]: I1125 16:25:28.671534 4890 state_mem.go:107] "Deleted CPUSet assignment" podUID="73aeebf4-f3e4-481f-997d-b019f7aa8cba" containerName="extract-utilities" Nov 25 16:25:28 crc kubenswrapper[4890]: I1125 16:25:28.671734 4890 memory_manager.go:354] "RemoveStaleState removing state" podUID="73aeebf4-f3e4-481f-997d-b019f7aa8cba" containerName="registry-server" Nov 25 16:25:28 crc kubenswrapper[4890]: I1125 16:25:28.673686 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zx8vw" Nov 25 16:25:28 crc kubenswrapper[4890]: I1125 16:25:28.698397 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zx8vw"] Nov 25 16:25:28 crc kubenswrapper[4890]: I1125 16:25:28.785575 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fc1fa21-27ca-4804-b977-0efc708b3c66-utilities\") pod \"community-operators-zx8vw\" (UID: \"1fc1fa21-27ca-4804-b977-0efc708b3c66\") " pod="openshift-marketplace/community-operators-zx8vw" Nov 25 16:25:28 crc kubenswrapper[4890]: I1125 16:25:28.785814 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fc1fa21-27ca-4804-b977-0efc708b3c66-catalog-content\") pod \"community-operators-zx8vw\" (UID: \"1fc1fa21-27ca-4804-b977-0efc708b3c66\") " pod="openshift-marketplace/community-operators-zx8vw" Nov 25 16:25:28 crc kubenswrapper[4890]: I1125 16:25:28.785887 4890 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5v4g9\" (UniqueName: \"kubernetes.io/projected/1fc1fa21-27ca-4804-b977-0efc708b3c66-kube-api-access-5v4g9\") pod \"community-operators-zx8vw\" (UID: \"1fc1fa21-27ca-4804-b977-0efc708b3c66\") " pod="openshift-marketplace/community-operators-zx8vw" Nov 25 16:25:28 crc kubenswrapper[4890]: I1125 16:25:28.887642 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fc1fa21-27ca-4804-b977-0efc708b3c66-catalog-content\") pod \"community-operators-zx8vw\" (UID: \"1fc1fa21-27ca-4804-b977-0efc708b3c66\") " pod="openshift-marketplace/community-operators-zx8vw" Nov 25 16:25:28 crc kubenswrapper[4890]: I1125 16:25:28.887730 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5v4g9\" (UniqueName: \"kubernetes.io/projected/1fc1fa21-27ca-4804-b977-0efc708b3c66-kube-api-access-5v4g9\") pod \"community-operators-zx8vw\" (UID: \"1fc1fa21-27ca-4804-b977-0efc708b3c66\") " pod="openshift-marketplace/community-operators-zx8vw" Nov 25 16:25:28 crc kubenswrapper[4890]: I1125 16:25:28.887796 4890 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fc1fa21-27ca-4804-b977-0efc708b3c66-utilities\") pod \"community-operators-zx8vw\" (UID: \"1fc1fa21-27ca-4804-b977-0efc708b3c66\") " pod="openshift-marketplace/community-operators-zx8vw" Nov 25 16:25:28 crc kubenswrapper[4890]: I1125 16:25:28.888267 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fc1fa21-27ca-4804-b977-0efc708b3c66-catalog-content\") pod \"community-operators-zx8vw\" (UID: \"1fc1fa21-27ca-4804-b977-0efc708b3c66\") " pod="openshift-marketplace/community-operators-zx8vw" Nov 25 16:25:28 crc kubenswrapper[4890]: I1125 16:25:28.888322 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fc1fa21-27ca-4804-b977-0efc708b3c66-utilities\") pod \"community-operators-zx8vw\" (UID: \"1fc1fa21-27ca-4804-b977-0efc708b3c66\") " pod="openshift-marketplace/community-operators-zx8vw" Nov 25 16:25:28 crc kubenswrapper[4890]: I1125 16:25:28.909595 4890 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5v4g9\" (UniqueName: \"kubernetes.io/projected/1fc1fa21-27ca-4804-b977-0efc708b3c66-kube-api-access-5v4g9\") pod \"community-operators-zx8vw\" (UID: \"1fc1fa21-27ca-4804-b977-0efc708b3c66\") " pod="openshift-marketplace/community-operators-zx8vw" Nov 25 16:25:28 crc kubenswrapper[4890]: I1125 16:25:28.998392 4890 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zx8vw" Nov 25 16:25:29 crc kubenswrapper[4890]: I1125 16:25:29.531777 4890 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zx8vw"] Nov 25 16:25:29 crc kubenswrapper[4890]: I1125 16:25:29.913032 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zx8vw" event={"ID":"1fc1fa21-27ca-4804-b977-0efc708b3c66","Type":"ContainerStarted","Data":"fc1d7095b7ee3ec99808fb2d2738989063014f2bda9f432a63a62a0d97172038"} Nov 25 16:25:29 crc kubenswrapper[4890]: I1125 16:25:29.913086 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zx8vw" event={"ID":"1fc1fa21-27ca-4804-b977-0efc708b3c66","Type":"ContainerStarted","Data":"0f2905e0de6584c96339c04a3a472a7f62f475eec6d314cc250d740e834aab26"} Nov 25 16:25:30 crc kubenswrapper[4890]: I1125 16:25:30.932413 4890 generic.go:334] "Generic (PLEG): container finished" podID="1fc1fa21-27ca-4804-b977-0efc708b3c66" containerID="fc1d7095b7ee3ec99808fb2d2738989063014f2bda9f432a63a62a0d97172038" exitCode=0 Nov 25 16:25:30 crc kubenswrapper[4890]: I1125 16:25:30.932566 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zx8vw" event={"ID":"1fc1fa21-27ca-4804-b977-0efc708b3c66","Type":"ContainerDied","Data":"fc1d7095b7ee3ec99808fb2d2738989063014f2bda9f432a63a62a0d97172038"} Nov 25 16:25:34 crc kubenswrapper[4890]: I1125 16:25:34.981695 4890 generic.go:334] "Generic (PLEG): container finished" podID="1fc1fa21-27ca-4804-b977-0efc708b3c66" containerID="b87fb8e6db95d3d6b67856af4f4da578345105e784d7590360d6d5713eb144a3" exitCode=0 Nov 25 16:25:34 crc kubenswrapper[4890]: I1125 16:25:34.981779 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zx8vw" event={"ID":"1fc1fa21-27ca-4804-b977-0efc708b3c66","Type":"ContainerDied","Data":"b87fb8e6db95d3d6b67856af4f4da578345105e784d7590360d6d5713eb144a3"} Nov 25 16:25:39 crc kubenswrapper[4890]: I1125 16:25:39.023786 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zx8vw" event={"ID":"1fc1fa21-27ca-4804-b977-0efc708b3c66","Type":"ContainerStarted","Data":"ba8670c3b0c742203e078888af5b5e1f337b9aa199bb9b2e1cba4a4529663b60"} Nov 25 16:25:39 crc kubenswrapper[4890]: I1125 16:25:39.087921 4890 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zx8vw" podStartSLOduration=4.287284199 podStartE2EDuration="11.087886917s" podCreationTimestamp="2025-11-25 16:25:28 +0000 UTC" firstStartedPulling="2025-11-25 16:25:30.93561699 +0000 UTC m=+4989.378079610" lastFinishedPulling="2025-11-25 16:25:37.736219678 +0000 UTC m=+4996.178682328" observedRunningTime="2025-11-25 16:25:39.075013206 +0000 UTC m=+4997.517475836" watchObservedRunningTime="2025-11-25 16:25:39.087886917 +0000 UTC m=+4997.530349527" Nov 25 16:25:48 crc kubenswrapper[4890]: I1125 16:25:48.998597 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zx8vw" Nov 25 16:25:49 crc kubenswrapper[4890]: I1125 16:25:49.000419 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zx8vw" Nov 25 16:25:49 crc kubenswrapper[4890]: I1125 16:25:49.058707 4890 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zx8vw" Nov 25 16:25:49 crc kubenswrapper[4890]: I1125 16:25:49.181607 4890 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zx8vw" Nov 25 16:25:49 crc kubenswrapper[4890]: I1125 16:25:49.298754 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zx8vw"] Nov 25 16:25:51 crc kubenswrapper[4890]: I1125 16:25:51.155424 4890 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zx8vw" podUID="1fc1fa21-27ca-4804-b977-0efc708b3c66" containerName="registry-server" containerID="cri-o://ba8670c3b0c742203e078888af5b5e1f337b9aa199bb9b2e1cba4a4529663b60" gracePeriod=2 Nov 25 16:25:51 crc kubenswrapper[4890]: I1125 16:25:51.657523 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zx8vw" Nov 25 16:25:51 crc kubenswrapper[4890]: I1125 16:25:51.762485 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5v4g9\" (UniqueName: \"kubernetes.io/projected/1fc1fa21-27ca-4804-b977-0efc708b3c66-kube-api-access-5v4g9\") pod \"1fc1fa21-27ca-4804-b977-0efc708b3c66\" (UID: \"1fc1fa21-27ca-4804-b977-0efc708b3c66\") " Nov 25 16:25:51 crc kubenswrapper[4890]: I1125 16:25:51.762602 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fc1fa21-27ca-4804-b977-0efc708b3c66-catalog-content\") pod \"1fc1fa21-27ca-4804-b977-0efc708b3c66\" (UID: \"1fc1fa21-27ca-4804-b977-0efc708b3c66\") " Nov 25 16:25:51 crc kubenswrapper[4890]: I1125 16:25:51.765275 4890 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fc1fa21-27ca-4804-b977-0efc708b3c66-utilities\") pod \"1fc1fa21-27ca-4804-b977-0efc708b3c66\" (UID: \"1fc1fa21-27ca-4804-b977-0efc708b3c66\") " Nov 25 16:25:51 crc kubenswrapper[4890]: I1125 16:25:51.766188 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1fc1fa21-27ca-4804-b977-0efc708b3c66-utilities" (OuterVolumeSpecName: "utilities") pod "1fc1fa21-27ca-4804-b977-0efc708b3c66" (UID: "1fc1fa21-27ca-4804-b977-0efc708b3c66"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:25:51 crc kubenswrapper[4890]: I1125 16:25:51.768405 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1fc1fa21-27ca-4804-b977-0efc708b3c66-kube-api-access-5v4g9" (OuterVolumeSpecName: "kube-api-access-5v4g9") pod "1fc1fa21-27ca-4804-b977-0efc708b3c66" (UID: "1fc1fa21-27ca-4804-b977-0efc708b3c66"). InnerVolumeSpecName "kube-api-access-5v4g9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 16:25:51 crc kubenswrapper[4890]: I1125 16:25:51.818597 4890 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1fc1fa21-27ca-4804-b977-0efc708b3c66-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1fc1fa21-27ca-4804-b977-0efc708b3c66" (UID: "1fc1fa21-27ca-4804-b977-0efc708b3c66"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 16:25:51 crc kubenswrapper[4890]: I1125 16:25:51.867612 4890 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5v4g9\" (UniqueName: \"kubernetes.io/projected/1fc1fa21-27ca-4804-b977-0efc708b3c66-kube-api-access-5v4g9\") on node \"crc\" DevicePath \"\"" Nov 25 16:25:51 crc kubenswrapper[4890]: I1125 16:25:51.867689 4890 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fc1fa21-27ca-4804-b977-0efc708b3c66-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 16:25:51 crc kubenswrapper[4890]: I1125 16:25:51.867701 4890 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fc1fa21-27ca-4804-b977-0efc708b3c66-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 16:25:52 crc kubenswrapper[4890]: I1125 16:25:52.168079 4890 generic.go:334] "Generic (PLEG): container finished" podID="1fc1fa21-27ca-4804-b977-0efc708b3c66" containerID="ba8670c3b0c742203e078888af5b5e1f337b9aa199bb9b2e1cba4a4529663b60" exitCode=0 Nov 25 16:25:52 crc kubenswrapper[4890]: I1125 16:25:52.168126 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zx8vw" event={"ID":"1fc1fa21-27ca-4804-b977-0efc708b3c66","Type":"ContainerDied","Data":"ba8670c3b0c742203e078888af5b5e1f337b9aa199bb9b2e1cba4a4529663b60"} Nov 25 16:25:52 crc kubenswrapper[4890]: I1125 16:25:52.168133 4890 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zx8vw" Nov 25 16:25:52 crc kubenswrapper[4890]: I1125 16:25:52.168155 4890 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zx8vw" event={"ID":"1fc1fa21-27ca-4804-b977-0efc708b3c66","Type":"ContainerDied","Data":"0f2905e0de6584c96339c04a3a472a7f62f475eec6d314cc250d740e834aab26"} Nov 25 16:25:52 crc kubenswrapper[4890]: I1125 16:25:52.168192 4890 scope.go:117] "RemoveContainer" containerID="ba8670c3b0c742203e078888af5b5e1f337b9aa199bb9b2e1cba4a4529663b60" Nov 25 16:25:52 crc kubenswrapper[4890]: I1125 16:25:52.193981 4890 scope.go:117] "RemoveContainer" containerID="b87fb8e6db95d3d6b67856af4f4da578345105e784d7590360d6d5713eb144a3" Nov 25 16:25:52 crc kubenswrapper[4890]: I1125 16:25:52.220881 4890 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zx8vw"] Nov 25 16:25:52 crc kubenswrapper[4890]: I1125 16:25:52.237841 4890 scope.go:117] "RemoveContainer" containerID="fc1d7095b7ee3ec99808fb2d2738989063014f2bda9f432a63a62a0d97172038" Nov 25 16:25:52 crc kubenswrapper[4890]: I1125 16:25:52.245282 4890 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zx8vw"] Nov 25 16:25:52 crc kubenswrapper[4890]: I1125 16:25:52.270865 4890 scope.go:117] "RemoveContainer" containerID="ba8670c3b0c742203e078888af5b5e1f337b9aa199bb9b2e1cba4a4529663b60" Nov 25 16:25:52 crc kubenswrapper[4890]: E1125 16:25:52.271529 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba8670c3b0c742203e078888af5b5e1f337b9aa199bb9b2e1cba4a4529663b60\": container with ID starting with ba8670c3b0c742203e078888af5b5e1f337b9aa199bb9b2e1cba4a4529663b60 not found: ID does not exist" containerID="ba8670c3b0c742203e078888af5b5e1f337b9aa199bb9b2e1cba4a4529663b60" Nov 25 16:25:52 crc kubenswrapper[4890]: I1125 16:25:52.271566 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba8670c3b0c742203e078888af5b5e1f337b9aa199bb9b2e1cba4a4529663b60"} err="failed to get container status \"ba8670c3b0c742203e078888af5b5e1f337b9aa199bb9b2e1cba4a4529663b60\": rpc error: code = NotFound desc = could not find container \"ba8670c3b0c742203e078888af5b5e1f337b9aa199bb9b2e1cba4a4529663b60\": container with ID starting with ba8670c3b0c742203e078888af5b5e1f337b9aa199bb9b2e1cba4a4529663b60 not found: ID does not exist" Nov 25 16:25:52 crc kubenswrapper[4890]: I1125 16:25:52.271589 4890 scope.go:117] "RemoveContainer" containerID="b87fb8e6db95d3d6b67856af4f4da578345105e784d7590360d6d5713eb144a3" Nov 25 16:25:52 crc kubenswrapper[4890]: E1125 16:25:52.272357 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b87fb8e6db95d3d6b67856af4f4da578345105e784d7590360d6d5713eb144a3\": container with ID starting with b87fb8e6db95d3d6b67856af4f4da578345105e784d7590360d6d5713eb144a3 not found: ID does not exist" containerID="b87fb8e6db95d3d6b67856af4f4da578345105e784d7590360d6d5713eb144a3" Nov 25 16:25:52 crc kubenswrapper[4890]: I1125 16:25:52.272415 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b87fb8e6db95d3d6b67856af4f4da578345105e784d7590360d6d5713eb144a3"} err="failed to get container status \"b87fb8e6db95d3d6b67856af4f4da578345105e784d7590360d6d5713eb144a3\": rpc error: code = NotFound desc = could not find container \"b87fb8e6db95d3d6b67856af4f4da578345105e784d7590360d6d5713eb144a3\": container with ID starting with b87fb8e6db95d3d6b67856af4f4da578345105e784d7590360d6d5713eb144a3 not found: ID does not exist" Nov 25 16:25:52 crc kubenswrapper[4890]: I1125 16:25:52.272449 4890 scope.go:117] "RemoveContainer" containerID="fc1d7095b7ee3ec99808fb2d2738989063014f2bda9f432a63a62a0d97172038" Nov 25 16:25:52 crc kubenswrapper[4890]: E1125 16:25:52.272815 4890 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc1d7095b7ee3ec99808fb2d2738989063014f2bda9f432a63a62a0d97172038\": container with ID starting with fc1d7095b7ee3ec99808fb2d2738989063014f2bda9f432a63a62a0d97172038 not found: ID does not exist" containerID="fc1d7095b7ee3ec99808fb2d2738989063014f2bda9f432a63a62a0d97172038" Nov 25 16:25:52 crc kubenswrapper[4890]: I1125 16:25:52.272858 4890 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc1d7095b7ee3ec99808fb2d2738989063014f2bda9f432a63a62a0d97172038"} err="failed to get container status \"fc1d7095b7ee3ec99808fb2d2738989063014f2bda9f432a63a62a0d97172038\": rpc error: code = NotFound desc = could not find container \"fc1d7095b7ee3ec99808fb2d2738989063014f2bda9f432a63a62a0d97172038\": container with ID starting with fc1d7095b7ee3ec99808fb2d2738989063014f2bda9f432a63a62a0d97172038 not found: ID does not exist" Nov 25 16:25:52 crc kubenswrapper[4890]: E1125 16:25:52.390496 4890 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1fc1fa21_27ca_4804_b977_0efc708b3c66.slice/crio-0f2905e0de6584c96339c04a3a472a7f62f475eec6d314cc250d740e834aab26\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1fc1fa21_27ca_4804_b977_0efc708b3c66.slice\": RecentStats: unable to find data in memory cache]" Nov 25 16:25:54 crc kubenswrapper[4890]: I1125 16:25:54.183919 4890 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1fc1fa21-27ca-4804-b977-0efc708b3c66" path="/var/lib/kubelet/pods/1fc1fa21-27ca-4804-b977-0efc708b3c66/volumes" Nov 25 16:25:56 crc kubenswrapper[4890]: I1125 16:25:56.447194 4890 patch_prober.go:28] interesting pod/machine-config-daemon-xx2tz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 16:25:56 crc kubenswrapper[4890]: I1125 16:25:56.448110 4890 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xx2tz" podUID="4e4f849d-f239-4727-a73e-18327856929a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515111354247024450 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015111354250017357 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015111342070016477 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015111342070015447 5ustar corecore